knotgrass / How-Transformers-Work
π§ A study guide to learn about Transformers
β11Updated last year
Alternatives and similar repositories for How-Transformers-Work
Users that are interested in How-Transformers-Work are comparing it to the libraries listed below
Sorting:
- This repository contains an implementation of the LLaMA 2 (Large Language Model Meta AI) model, a Generative Pretrained Transformer (GPT)β¦β65Updated last year
- A set of scripts and notebooks on LLM finetunning and dataset creationβ109Updated 7 months ago
- This repository contains the code for dataset curation and finetuning of instruct variant of the Bilingual OpenHathi model. The resultinβ¦β23Updated last year
- Lightweight demos for finetuning LLMs. Powered by π€ transformers and open-source datasets.β76Updated 6 months ago
- Prune transformer layersβ69Updated 11 months ago
- LLM Workshop by Sourab Mangrulkarβ382Updated 11 months ago
- An extension of the nanoGPT repository for training small MOE models.β142Updated 2 months ago
- Tutorial for how to build BERT from scratchβ93Updated 11 months ago
- β163Updated 4 months ago
- LoRA and DoRA from Scratch Implementationsβ202Updated last year
- Easy and Efficient Quantization for Transformersβ197Updated 3 months ago
- EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language Mβ¦β217Updated 6 months ago
- Training and Fine-tuning an llm in Python and PyTorch.β41Updated last year
- Pre-training code for Amber 7B LLMβ166Updated last year
- Distributed training (multi-node) of a Transformer modelβ66Updated last year
- A collection of LogitsProcessors to customize and enhance LLM behavior for specific tasks.β282Updated last week
- β186Updated 3 months ago
- experiments with inference on llamaβ104Updated 11 months ago
- Fine-tune ModernBERT on a large Dataset with Custom Tokenizer Trainingβ66Updated 3 months ago
- ποΈ A unified multi-backend utility for benchmarking Transformers, Timm, PEFT, Diffusers and Sentence-Transformers with full support of Oβ¦β301Updated this week
- A repository to unravel the language of GPUs, making their kernel conversations easy to understandβ180Updated this week
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for freeβ231Updated 6 months ago
- LLaMA 2 implemented from scratch in PyTorchβ324Updated last year
- Simple implementation of Speculative Sampling in NumPy for GPT-2.β95Updated last year
- Fine-tuning Open-Source LLMs for Adaptive Machine Translationβ78Updated last month
- Notes about LLaMA 2 modelβ59Updated last year
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.β34Updated last week
- Well documented, unit tested, type checked and formatted implementation of a vanilla transformer - for educational purposes.β245Updated last year
- Code for NeurIPS LLM Efficiency Challengeβ58Updated last year
- Official PyTorch implementation of DistiLLM: Towards Streamlined Distillation for Large Language Models (ICML 2024)β215Updated 2 months ago