abacaj / fine-tune-mistral
Fine-tune mistral-7B on 3090s, a100s, h100s
☆702Updated last year
Related projects ⓘ
Alternatives and complementary repositories for fine-tune-mistral
- Customizable implementation of the self-instruct paper.☆1,024Updated 8 months ago
- Generate textbook-quality synthetic LLM pretraining data☆488Updated last year
- ☆411Updated last year
- Tune any FALCON in 4-bit☆468Updated last year
- ☆470Updated 2 months ago
- Extend existing LLMs way beyond the original training length with constant memory usage, without retraining☆675Updated 7 months ago
- A bagel, with everything.☆312Updated 7 months ago
- YaRN: Efficient Context Window Extension of Large Language Models☆1,353Updated 7 months ago
- Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verifi…☆1,634Updated this week
- Inference code for Persimmon-8B☆416Updated last year
- Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI☆1,336Updated 7 months ago
- batched loras☆336Updated last year
- ☆451Updated 3 weeks ago
- A benchmark to evaluate language models on questions I've previously asked them to solve.☆916Updated 2 weeks ago
- Automatically evaluate your LLMs in Google Colab☆559Updated 6 months ago
- Guide for fine-tuning Llama/Mistral/CodeLlama models and more☆534Updated 2 months ago
- ☆860Updated 11 months ago
- ☆718Updated 2 months ago
- ReFT: Representation Finetuning for Language Models☆1,159Updated 2 weeks ago
- Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends☆811Updated this week
- Code for fine-tuning Platypus fam LLMs using LoRA☆623Updated 9 months ago
- Inference code for Mistral and Mixtral hacked up into original Llama implementation☆373Updated 11 months ago
- Minimalistic large language model 3D-parallelism training☆1,260Updated this week
- Fast & Simple repository for pre-training and fine-tuning T5-style models☆970Updated 3 months ago
- Ungreedy subword tokenizer and vocabulary trainer for Python, Go & Javascript☆551Updated 4 months ago
- Finetuning Large Language Models on One Consumer GPU in 2 Bits☆707Updated 5 months ago
- [ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning☆613Updated 5 months ago
- The repository for the code of the UltraFastBERT paper☆514Updated 7 months ago