cedrickchee / llama
Inference code for LLaMA 2 models
☆30Updated 8 months ago
Alternatives and similar repositories for llama:
Users that are interested in llama are comparing it to the libraries listed below
- Inference code for LLaMA models☆46Updated 2 years ago
- ☆26Updated 2 years ago
- Modified Stanford-Alpaca Trainer for Training Replit's Code Model☆40Updated last year
- Inference script for Meta's LLaMA models using Hugging Face wrapper☆110Updated last year
- Instruct-tune LLaMA on consumer hardware☆73Updated last year
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆165Updated last year
- Train llama with lora on one 4090 and merge weight of lora to work as stanford alpaca.☆50Updated last year
- The GeoV model is a large langauge model designed by Georges Harik and uses Rotary Positional Embeddings with Relative distances (RoPER).…☆121Updated last year
- 4 bits quantization of SantaCoder using GPTQ☆51Updated last year
- ☆40Updated last year
- Code and documentation to train Stanford's Alpaca models, and generate the data.☆111Updated last year
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hub☆157Updated last year
- A converter and basic tester for rwkv onnx☆42Updated last year
- An unsupervised model merging algorithm for Transformers-based language models.☆106Updated 10 months ago
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆69Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMs☆77Updated 11 months ago
- NeurIPS 2023 - Cappy: Outperforming and Boosting Large Multi-Task LMs with a Small Scorer☆41Updated 11 months ago
- Instruct-tuning LLaMA on consumer hardware☆66Updated last year
- Extend the original llama.cpp repo to support redpajama model.☆117Updated 6 months ago
- ☆37Updated 2 years ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆65Updated last year
- RWKV (Receptance Weighted Key Value) is a RNN with Transformer-level performance☆39Updated 2 years ago
- LLaMA implementation for HuggingFace Transformers☆37Updated last year
- BlinkDL's RWKV-v4 running in the browser☆47Updated 2 years ago
- SparseGPT + GPTQ Compression of LLMs like LLaMa, OPT, Pythia☆41Updated last year
- Scripts to create your own moe models using mlx☆89Updated last year
- ☆81Updated 9 months ago
- A finetuning pipeline for instruct tuning Raven 14bn using QLORA 4bit and the Ditty finetuning library☆28Updated 9 months ago
- Inference code for LLaMA models with Gradio Interface and rolling generation like ChatGPT☆48Updated last year