Artefact2 / llm-evalLinks
A super simple web interface to perform blind tests on LLM outputs.
☆28Updated last year
Alternatives and similar repositories for llm-eval
Users that are interested in llm-eval are comparing it to the libraries listed below
Sorting:
- GGML implementation of BERT model with Python bindings and quantization.☆56Updated last year
- Python bindings for ggml☆146Updated last year
- ☆161Updated last month
- Scripts to create your own moe models using mlx☆90Updated last year
- Public reports detailing responses to sets of prompts by Large Language Models.☆31Updated 8 months ago
- inference code for mixtral-8x7b-32kseqlen☆101Updated last year
- LLaVA server (llama.cpp).☆182Updated last year
- an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated last year
- Inference of Mamba models in pure C☆191Updated last year
- A guidance compatibility layer for llama-cpp-python☆36Updated 2 years ago
- Embedding models from Jina AI☆64Updated last year
- llama.cpp to PyTorch Converter☆34Updated last year
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆170Updated last year
- Transformer GPU VRAM estimator☆66Updated last year
- Implementation of nougat that focuses on processing pdf locally.☆82Updated 8 months ago
- ☆36Updated last year
- ☆74Updated 2 years ago
- Benchmarks comparing PyTorch and MLX on Apple Silicon GPUs☆88Updated last year
- ☆116Updated 7 months ago
- ☆38Updated last year
- Low-Rank adapter extraction for fine-tuned transformers models☆177Updated last year
- GGUF implementation in C as a library and a tools CLI program☆290Updated 3 weeks ago
- An OpenAI API compatible LLM inference server based on ExLlamaV2.☆25Updated last year
- Client Code Examples, Use Cases and Benchmarks for Enterprise h2oGPTe RAG-Based GenAI Platform☆90Updated last week
- A minimalistic C++ Jinja templating engine for LLM chat templates☆180Updated last week
- Just a bunch of benchmark logs for different LLMs☆119Updated last year
- Merge Transformers language models by use of gradient parameters.☆208Updated last year
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes…☆146Updated last year
- For inferring and serving local LLMs using the MLX framework☆109Updated last year
- Tools for formatting large language model prompts.☆13Updated last year