sleepingcat4 / TinyStoriesLinks
code to train a gpt-2 model to train it on tiny stories dataset according to the TinyStories paper
☆39Updated last year
Alternatives and similar repositories for TinyStories
Users that are interested in TinyStories are comparing it to the libraries listed below
Sorting:
- ☆96Updated 2 years ago
- Code repository for the c-BTM paper☆107Updated last year
- ☆460Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers☆423Updated last year
- A minimum example of aligning language models with RLHF similar to ChatGPT☆221Updated last year
- Multipack distributed sampler for fast padding-free training of LLMs☆199Updated last year
- A bagel, with everything.☆324Updated last year
- TART: A plug-and-play Transformer module for task-agnostic reasoning☆200Updated 2 years ago
- Pre-training code for Amber 7B LLM☆167Updated last year
- batched loras☆345Updated last year
- Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".☆277Updated last year
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆136Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMs☆78Updated last year
- Inference code for Persimmon-8B☆415Updated last year
- Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA☆104Updated 3 months ago
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆103Updated 2 years ago
- ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward exp…☆223Updated last year
- Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks☆208Updated last year
- SAIL: Search Augmented Instruction Learning☆157Updated last month
- A crude RLHF layer on top of nanoGPT with Gumbel-Softmax trick☆290Updated last year
- Tune MPTs☆84Updated 2 years ago
- ☆180Updated 2 years ago
- Sakura-SOLAR-DPO: Merge, SFT, and DPO☆116Updated last year
- This repository contains code for extending the Stanford Alpaca synthetic instruction tuning to existing instruction-tuned models such as…☆353Updated 2 years ago
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes…☆145Updated last year
- A repository for transformer critique learning and generation☆90Updated last year
- RWKV in nanoGPT style☆193Updated last year
- JAX implementation of the Llama 2 model☆219Updated last year
- ☆273Updated 2 years ago
- Step by step explanation/tutorial of llama2.c☆223Updated last year