mistralai-sf24 / hackathon
☆445Updated 11 months ago
Alternatives and similar repositories for hackathon:
Users that are interested in hackathon are comparing it to the libraries listed below
- Accelerate your Hugging Face Transformers 7.6-9x. Native to Hugging Face and PyTorch.☆685Updated 7 months ago
- ☆937Updated last month
- Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI☆1,374Updated 11 months ago
- ☆694Updated this week
- ☆512Updated 7 months ago
- Automatically evaluate your LLMs in Google Colab☆603Updated 10 months ago
- Training LLMs with QLoRA + FSDP☆1,464Updated 4 months ago
- Fine-tune mistral-7B on 3090s, a100s, h100s☆709Updated last year
- A bagel, with everything.☆317Updated 11 months ago
- Train Models Contrastively in Pytorch☆666Updated last month
- Official implementation of "Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling"☆855Updated last month
- The official implementation of Self-Play Fine-Tuning (SPIN)☆1,135Updated 10 months ago
- Reaching LLaMA2 Performance with 0.1M Dollars☆980Updated 8 months ago
- Extend existing LLMs way beyond the original training length with constant memory usage, without retraining☆691Updated 11 months ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆230Updated 4 months ago
- Evaluation suite for LLMs☆338Updated 3 months ago
- A benchmark to evaluate language models on questions I've previously asked them to solve.☆990Updated last month
- Website for hosting the Open Foundation Models Cheat Sheet.☆264Updated 2 weeks ago
- ☆502Updated 4 months ago
- ☆864Updated last year
- A library for easily merging multiple LLM experts, and efficiently train the merged LLM.☆454Updated 6 months ago
- ☆412Updated last year
- Inference code for Persimmon-8B☆415Updated last year
- This project showcases an LLMOps pipeline that fine-tunes a small-size LLM model to prepare for the outage of the service LLM.☆302Updated last month
- [ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning☆646Updated 9 months ago
- Manage scalable open LLM inference endpoints in Slurm clusters☆253Updated 8 months ago
- Code for Quiet-STaR☆721Updated 7 months ago
- Banishing LLM Hallucinations Requires Rethinking Generalization☆272Updated 8 months ago
- Serving multiple LoRA finetuned LLM as one☆1,040Updated 10 months ago
- Official implementation of Half-Quadratic Quantization (HQQ)