salesforce / xgen
Salesforce open-source LLMs with 8k sequence length.
☆716Updated last year
Alternatives and similar repositories for xgen:
Users that are interested in xgen are comparing it to the libraries listed below
- [ICLR 2024] Lemur: Open Foundation Models for Language Agents☆541Updated last year
- Code for fine-tuning Platypus fam LLMs using LoRA☆625Updated 11 months ago
- LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale Instructions☆815Updated last year
- Inference code for Persimmon-8B☆416Updated last year
- Official repository for LongChat and LongEval☆518Updated 7 months ago
- ☆537Updated last month
- Official implementation of our NeurIPS 2023 paper "Augmenting Language Models with Long-Term Memory".☆775Updated 9 months ago
- [ACL2023] We introduce LLM-Blender, an innovative ensembling framework to attain consistently superior performance by leveraging the dive…☆902Updated 2 months ago
- ☆1,441Updated last year
- ☆456Updated last year
- OpenAlpaca: A Fully Open-Source Instruction-Following Model Based On OpenLLaMA☆301Updated last year
- Extend existing LLMs way beyond the original training length with constant memory usage, without retraining☆684Updated 9 months ago
- This repository contains code and tooling for the Abacus.AI LLM Context Expansion project. Also included are evaluation scripts and bench…☆583Updated last year
- LOMO: LOw-Memory Optimization☆979Updated 6 months ago
- CodeGen2 models for program synthesis☆275Updated last year
- Customizable implementation of the self-instruct paper.☆1,035Updated 10 months ago
- Dromedary: towards helpful, ethical and reliable LLMs.☆1,131Updated last year
- The complete training code of the open-source high-performance Llama model, including the full process from pre-training to RLHF.☆33Updated last year
- A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.☆791Updated 6 months ago
- OpenICL is an open-source framework to facilitate research, development, and prototyping of in-context learning.☆543Updated last year
- ☆348Updated last year
- Tune any FALCON in 4-bit☆466Updated last year
- CodeTF: One-stop Transformer Library for State-of-the-art Code LLM☆1,462Updated 7 months ago
- [NeurIPS 2023] MeZO: Fine-Tuning Language Models with Just Forward Passes. https://arxiv.org/abs/2305.17333☆1,075Updated last year
- Build, evaluate, understand, and fix LLM-based apps☆484Updated last year
- Alpaca dataset from Stanford, cleaned and curated☆1,531Updated last year
- Public repo for the NeurIPS 2023 paper "Unlimiformer: Long-Range Transformers with Unlimited Length Input"☆1,060Updated 10 months ago
- This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.☆538Updated 10 months ago
- ☆413Updated last year