cedrickchee / llama
Inference code for LLaMA 2 models
☆31Updated 7 months ago
Alternatives and similar repositories for llama:
Users that are interested in llama are comparing it to the libraries listed below
- Instruct-tuning LLaMA on consumer hardware☆66Updated last year
- ☆26Updated last year
- Train llama with lora on one 4090 and merge weight of lora to work as stanford alpaca.☆50Updated last year
- 4 bits quantization of SantaCoder using GPTQ☆51Updated last year
- Extend the original llama.cpp repo to support redpajama model.☆117Updated 5 months ago
- Modified Stanford-Alpaca Trainer for Training Replit's Code Model☆40Updated last year
- LoRA weights for Cerebras-GPT-2.7b finetuned on Alpaca dataset with shorter prompt☆63Updated last year
- Automated prompting and scoring framework to evaluate LLMs using updated human knowledge prompts☆111Updated last year
- ☆22Updated last year
- Embeddings focused small version of Llama NLP model☆103Updated last year
- Instruct-tune LLaMA on consumer hardware☆73Updated last year
- Tune MPTs☆84Updated last year
- Code and documentation to train Stanford's Alpaca models, and generate the data.☆111Updated last year
- Code and models for BERT on STILTs☆53Updated last year
- The Next Generation Multi-Modality Superintelligence☆71Updated 5 months ago
- This project aims to make RWKV Accessible to everyone using a Hugging Face like interface, while keeping it close to the R and D RWKV bra…☆64Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMs☆77Updated 10 months ago
- ☆36Updated last year
- ☆12Updated last year
- Command-line script for inferencing from models such as MPT-7B-Chat☆101Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆123Updated last year
- BlinkDL's RWKV-v4 running in the browser☆47Updated last year
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- Inference code for LLaMA models with Gradio Interface and rolling generation like ChatGPT☆48Updated last year
- Here is a Google Colab Notebook for fine-tuning Alpaca Lora (within 3 hours with a 40GB A100 GPU)☆38Updated last year
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆65Updated last year
- Adversarial Training and SFT for Bot Safety Models☆39Updated last year
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆164Updated last year
- Framework for finetunning the ToolFormer-based LM in a few shots manner☆24Updated last year
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆69Updated last year