Manuel030 / alpaca-optLinks
Yet another LLM
☆10Updated 2 years ago
Alternatives and similar repositories for alpaca-opt
Users that are interested in alpaca-opt are comparing it to the libraries listed below
Sorting:
- Experimental sampler to make LLMs more creative☆31Updated 2 years ago
- ☆73Updated 2 years ago
- The Next Generation Multi-Modality Superintelligence☆69Updated last year
- GPT-2 small trained on phi-like data☆67Updated last year
- ☆63Updated last year
- ☆40Updated 2 years ago
- A public implementation of the ReLoRA pretraining method, built on Lightning-AI's Pytorch Lightning suite.☆35Updated last year
- entropix style sampling + GUI☆27Updated last year
- Modified Beam Search with periodical restart☆12Updated last year
- A simple package for leveraging Falcon 180B and the HF ecosystem's tools, including training/inference scripts, safetensors, integrations…☆11Updated last year
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆62Updated 2 years ago
- ☆33Updated 2 years ago
- An open source replication of the stawberry method that leverages Monte Carlo Search with PPO and or DPO☆29Updated this week
- ☆49Updated last year
- Implementation of the Mamba SSM with hf_integration.☆56Updated last year
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31Updated last year
- Merge LLM that are split in to parts☆26Updated 3 months ago
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆102Updated 2 years ago
- The Benefits of a Concise Chain of Thought on Problem Solving in Large Language Models☆22Updated 11 months ago
- ☆39Updated 3 years ago
- SparseGPT + GPTQ Compression of LLMs like LLaMa, OPT, Pythia☆40Updated 2 years ago
- Finetune any model on HF in less than 30 seconds☆55Updated last week
- 4 bits quantization of SantaCoder using GPTQ☆50Updated 2 years ago
- QLoRA with Enhanced Multi GPU Support☆37Updated 2 years ago
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆69Updated 2 years ago
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆169Updated last year
- GoldFinch and other hybrid transformer components☆45Updated last year
- ☆17Updated last year
- Official code for ACL 2023 (short, findings) paper "Recursion of Thought: A Divide and Conquer Approach to Multi-Context Reasoning with L…☆45Updated 2 years ago
- ☆35Updated 2 years ago