fredi-python / Fine-tune-RedPajama-Chat-3BLinks
Code for finetuning RedPajama-Chat-3B using LoRA
☆13Updated 2 years ago
Alternatives and similar repositories for Fine-tune-RedPajama-Chat-3B
Users that are interested in Fine-tune-RedPajama-Chat-3B are comparing it to the libraries listed below
Sorting:
- entropix style sampling + GUI☆26Updated 7 months ago
- ☆73Updated last year
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31Updated last year
- GPT-2 small trained on phi-like data☆66Updated last year
- Video+code lecture on building nanoGPT from scratch☆68Updated last year
- ☆20Updated last year
- Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hub☆162Updated last year
- Code for the paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot" with LLaMA implementation.☆71Updated 2 years ago
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆123Updated 2 years ago
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆69Updated last year
- ☆53Updated last year
- Falcon40B and 7B (Instruct) with streaming, top-k, and beam search☆40Updated 2 years ago
- Deploy your GGML models to HuggingFace Spaces with Docker and gradio☆37Updated 2 years ago
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated 2 years ago
- An OpenAI Completions API compatible server for NLP transformers models☆65Updated last year
- Experimental sampler to make LLMs more creative☆31Updated last year
- Plug n Play GBNF Compiler for llama.cpp☆25Updated last year
- ☆66Updated last year
- ☆40Updated 2 years ago
- ☆33Updated 2 years ago
- Mistral7B playing DOOM☆28Updated last year
- An OpenAI API compatible LLM inference server based on ExLlamaV2.☆25Updated last year
- QLoRA with Enhanced Multi GPU Support☆37Updated last year
- ☆22Updated last year
- ☆22Updated last year
- an implementation of Self-Extend, to expand the context window via grouped attention☆119Updated last year
- 5X faster 60% less memory QLoRA finetuning☆21Updated last year
- Model REVOLVER, a human in the loop model mixing system.☆33Updated last year
- The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆41Updated last year
- ☆16Updated 2 years ago