cindysridykhan / instruct_storyteller_tinyllama2Links
Training and Fine-tuning an llm in Python and PyTorch.
☆42Updated last year
Alternatives and similar repositories for instruct_storyteller_tinyllama2
Users that are interested in instruct_storyteller_tinyllama2 are comparing it to the libraries listed below
Sorting:
- Micro Llama is a small Llama based model with 300M parameters trained from scratch with $500 budget☆153Updated 3 weeks ago
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆70Updated last year
- Small and Efficient Mathematical Reasoning LLMs☆71Updated last year
- ☆88Updated last year
- Data preparation code for Amber 7B LLM☆91Updated last year
- Pre-training code for Amber 7B LLM☆167Updated last year
- A pipeline for LLM knowledge distillation☆108Updated 4 months ago
- Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".☆277Updated last year
- minimal scripts for 24GB VRAM GPUs. training, inference, whatever☆41Updated last month
- A bagel, with everything.☆324Updated last year
- Experiments on speculative sampling with Llama models☆128Updated 2 years ago
- experiments with inference on llama☆104Updated last year
- ☆77Updated last year
- Simple GRPO scripts and configurations.☆59Updated 6 months ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆232Updated 9 months ago
- inference code for mixtral-8x7b-32kseqlen☆101Updated last year
- EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language M…☆232Updated 9 months ago
- Verifiers for LLM Reinforcement Learning☆69Updated 3 months ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆135Updated last year
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"☆306Updated last year
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆199Updated last year
- ☆95Updated 2 years ago
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆168Updated last year
- Just a bunch of benchmark logs for different LLMs☆119Updated last year
- ☆54Updated last year
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆88Updated this week
- Multipack distributed sampler for fast padding-free training of LLMs☆199Updated last year
- Positional Skip-wise Training for Efficient Context Window Extension of LLMs to Extremely Length (ICLR 2024)☆206Updated last year
- Code for NeurIPS LLM Efficiency Challenge☆59Updated last year
- Inference code for Mistral and Mixtral hacked up into original Llama implementation☆371Updated last year