LeonEricsson / llmcontext
Pressure testing the context window of open LLMs
☆23Updated 6 months ago
Alternatives and similar repositories for llmcontext:
Users that are interested in llmcontext are comparing it to the libraries listed below
- an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated last year
- Low-Rank adapter extraction for fine-tuned transformers models☆171Updated 10 months ago
- ☆65Updated 9 months ago
- An unsupervised model merging algorithm for Transformers-based language models.☆106Updated 10 months ago
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆137Updated last month
- [WIP] Transformer to embed Danbooru labelsets☆13Updated 11 months ago
- ☆111Updated 3 months ago
- ☆49Updated last year
- GPT-2 small trained on phi-like data☆65Updated last year
- ☆27Updated last year
- Easy to use, High Performant Knowledge Distillation for LLMs☆54Updated this week
- ☆73Updated last year
- Fast approximate inference on a single GPU with sparsity aware offloading☆38Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMs☆77Updated 11 months ago
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆224Updated 10 months ago
- Evaluating LLMs with CommonGen-Lite☆89Updated last year
- Testing LLM reasoning abilities with family relationship quizzes.☆62Updated last month
- Full finetuning of large language models without large memory requirements☆93Updated last year
- QLoRA with Enhanced Multi GPU Support☆36Updated last year
- ☆126Updated 7 months ago
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- Let's create synthetic textbooks together :)☆73Updated last year
- entropix style sampling + GUI☆25Updated 4 months ago
- A framework for evaluating function calls made by LLMs☆37Updated 8 months ago
- This is our own implementation of 'Layer Selective Rank Reduction'☆233Updated 9 months ago
- ☆113Updated 5 months ago
- The Benefits of a Concise Chain of Thought on Problem Solving in Large Language Models☆21Updated 3 months ago
- Gradio UI for a Cog API☆66Updated 11 months ago
- Preprint: Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning☆28Updated last year
- ☆152Updated 8 months ago