knotgrass / How-Transformers-Work
π§ A study guide to learn about Transformers
β10Updated 9 months ago
Related projects β
Alternatives and complementary repositories for How-Transformers-Work
- experiments with inference on llamaβ105Updated 5 months ago
- Prune transformer layersβ63Updated 5 months ago
- Tutorial for how to build BERT from scratchβ83Updated 5 months ago
- A set of scripts and notebooks on LLM finetunning and dataset creationβ92Updated last month
- LLM Workshop by Sourab Mangrulkarβ340Updated 4 months ago
- Positional Skip-wise Training for Efficient Context Window Extension of LLMs to Extremely Length (ICLR 2024)β200Updated 5 months ago
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Dayβ251Updated last year
- Explorations into some recent techniques surrounding speculative decodingβ209Updated last year
- Official PyTorch implementation of QA-LoRAβ116Updated 7 months ago
- EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language Mβ¦β169Updated last week
- This repository contains the code for dataset curation and finetuning of instruct variant of the Bilingual OpenHathi model. The resultinβ¦β23Updated 10 months ago
- Easy and Efficient Quantization for Transformersβ178Updated 3 months ago
- DSIR large-scale data selection framework for language model trainingβ227Updated 7 months ago
- Manage scalable open LLM inference endpoints in Slurm clustersβ237Updated 3 months ago
- β156Updated last month
- Cold Compress is a hackable, lightweight, and open-source toolkit for creating and benchmarking cache compression methods built on top ofβ¦β86Updated 3 months ago
- LLaMA 2 implemented from scratch in PyTorchβ250Updated last year
- The official evaluation suite and dynamic data release for MixEval.β222Updated last week
- β246Updated last year
- ring-attention experimentsβ95Updated 3 weeks ago
- Multipack distributed sampler for fast padding-free training of LLMsβ175Updated 3 months ago
- β13Updated last year
- A minimum example of aligning language models with RLHF similar to ChatGPTβ213Updated last year
- Implementation of CALM from the paper "LLM Augmented LLMs: Expanding Capabilities through Composition", out of Google Deepmindβ168Updated last month
- Simple implementation of Speculative Sampling in NumPy for GPT-2.β89Updated last year
- β96Updated last month
- Implementation of paper Data Engineering for Scaling Language Models to 128K Contextβ435Updated 7 months ago
- β63Updated last month
- Lightweight demos for finetuning LLMs. Powered by π€ transformers and open-source datasets.β64Updated 3 weeks ago
- BABILong is a benchmark for LLM evaluation using the needle-in-a-haystack approach.β150Updated 2 months ago
- ποΈ A unified multi-backend utility for benchmarking Transformers, Timm, PEFT, Diffusers and Sentence-Transformers with full support of Oβ¦β251Updated this week