huggingface / llm-swarm
Manage scalable open LLM inference endpoints in Slurm clusters
☆236Updated 4 months ago
Related projects ⓘ
Alternatives and complementary repositories for llm-swarm
- EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language M…☆180Updated 3 weeks ago
- awesome synthetic (text) datasets☆242Updated 3 weeks ago
- Multipack distributed sampler for fast padding-free training of LLMs☆178Updated 3 months ago
- ☆451Updated 3 weeks ago
- experiments with inference on llama☆105Updated 5 months ago
- An Open Source Toolkit For LLM Distillation☆356Updated 2 months ago
- The official evaluation suite and dynamic data release for MixEval.☆224Updated last week
- Positional Skip-wise Training for Efficient Context Window Extension of LLMs to Extremely Length (ICLR 2024)☆199Updated 6 months ago
- Scaling Data-Constrained Language Models☆321Updated last month
- ☆93Updated last month
- code for training & evaluating Contextual Document Embedding models☆117Updated this week
- Automated Identification of Redundant Layer Blocks for Pruning in Large Language Models☆196Updated 6 months ago
- BABILong is a benchmark for LLM evaluation using the needle-in-a-haystack approach.☆152Updated last week
- Benchmarking LLMs with Challenging Tasks from Real Users☆195Updated 2 weeks ago
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"☆293Updated 11 months ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆173Updated 4 months ago
- A bagel, with everything.☆312Updated 7 months ago
- Website for hosting the Open Foundation Models Cheat Sheet.☆257Updated 4 months ago
- ☆112Updated last month
- A pipeline to improve skills of large language models☆191Updated this week
- Repo for "LoLCATs: On Low-Rank Linearizing of Large Language Models"☆177Updated last month
- A pipeline for LLM knowledge distillation☆78Updated 3 months ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆221Updated 2 weeks ago
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆64Updated last month
- Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.☆307Updated 7 months ago
- Toolkit for attaching, training, saving and loading of new heads for transformer models☆246Updated 2 weeks ago
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024☆229Updated 3 weeks ago
- ☆108Updated this week
- This is our own implementation of 'Layer Selective Rank Reduction'☆232Updated 5 months ago
- ☆184Updated last month