Rallio67 / language-model-agents
Experiments with generating opensource language model assistants
☆97Updated last year
Alternatives and similar repositories for language-model-agents:
Users that are interested in language-model-agents are comparing it to the libraries listed below
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆115Updated 2 years ago
- QLoRA with Enhanced Multi GPU Support☆36Updated last year
- Anh - LAION's multilingual assistant datasets and models☆27Updated last year
- Code repository for the c-BTM paper☆106Updated last year
- ☆72Updated last year
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆93Updated 2 years ago
- Pipeline for pulling and processing online language model pretraining data from the web☆177Updated last year
- A library for squeakily cleaning and filtering language datasets.☆46Updated last year
- ☆44Updated 4 months ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆82Updated last year
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- Multipack distributed sampler for fast padding-free training of LLMs☆186Updated 7 months ago
- ☆97Updated 2 years ago
- ☆67Updated 2 years ago
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆73Updated 5 months ago
- An experimental implementation of the retrieval-enhanced language model☆74Updated 2 years ago
- Multi-Domain Expert Learning☆67Updated last year
- Adversarial Training and SFT for Bot Safety Models☆39Updated last year
- Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA☆102Updated 7 months ago
- Tutorial to pretrain & fine-tune a 🤗 Flax T5 model on a TPUv3-8 with GCP☆58Updated 2 years ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆117Updated last year
- some common Huggingface transformers in maximal update parametrization (µP)☆80Updated 3 years ago
- [ICLR 2023] Guess the Instruction! Flipped Learning Makes Language Models Stronger Zero-Shot Learners☆113Updated 6 months ago
- Code for the paper "The Impact of Positional Encoding on Length Generalization in Transformers", NeurIPS 2023☆131Updated 10 months ago
- ☆94Updated last year
- This repository contains code for cleaning your training data of benchmark data to help combat data snooping.☆25Updated last year
- ☆67Updated 2 years ago
- ☆37Updated last year
- A repository for transformer critique learning and generation☆89Updated last year
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated last year