mistralai-sf24 / hackathon
☆448Updated 7 months ago
Related projects ⓘ
Alternatives and complementary repositories for hackathon
- Accelerate your Hugging Face Transformers 7.6-9x. Native to Hugging Face and PyTorch.☆687Updated 2 months ago
- Automatically evaluate your LLMs in Google Colab☆556Updated 6 months ago
- ☆634Updated 2 weeks ago
- Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends☆787Updated last week
- Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI☆1,333Updated 6 months ago
- ☆445Updated last week
- Train Models Contrastively in Pytorch☆543Updated 2 weeks ago
- Simple Python library/structure to ablate features in LLMs which are supported by TransformerLens☆324Updated 4 months ago
- Official implementation of "Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling"☆801Updated 2 months ago
- A bagel, with everything.☆312Updated 6 months ago
- ReFT: Representation Finetuning for Language Models☆1,145Updated this week
- ☆465Updated 2 months ago
- An Open Source Toolkit For LLM Distillation☆350Updated last month
- Evaluate your LLM's response with Prometheus and GPT4 💯☆794Updated 2 months ago
- Official implementation of Half-Quadratic Quantization (HQQ)☆698Updated last week
- A lightweight, low-dependency, unified API to use all common reranking and cross-encoder models.☆1,042Updated this week
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆219Updated last week
- Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verifi…☆1,612Updated this week
- ☆705Updated last month
- Minimalistic large language model 3D-parallelism training☆1,227Updated this week
- The official implementation of Self-Play Fine-Tuning (SPIN)☆1,034Updated 6 months ago
- Code for Quiet-STaR☆639Updated 2 months ago
- Generate textbook-quality synthetic LLM pretraining data☆488Updated last year
- This project showcases an LLMOps pipeline that fine-tunes a small-size LLM model to prepare for the outage of the service LLM.☆288Updated 2 months ago
- Fine-tune mistral-7B on 3090s, a100s, h100s☆702Updated last year
- Easily embed, cluster and semantically label text datasets☆459Updated 7 months ago
- A comprehensive repository of reasoning tasks for LLMs (and beyond)☆273Updated last month
- Banishing LLM Hallucinations Requires Rethinking Generalization☆260Updated 3 months ago