CarperAI / OpenELM
Evolution Through Large Models
β717Updated last year
Alternatives and similar repositories for OpenELM:
Users that are interested in OpenELM are comparing it to the libraries listed below
- Minimal library to train LLMs on TPU in JAX with pjit().β284Updated last year
- Code for Parsel π - generate complex programs with language modelsβ430Updated last year
- β412Updated last year
- Implementation of Recurrent Memory Transformer, Neurips 2022 paper, in Pytorchβ407Updated 3 months ago
- Used for adaptive human in the loop evaluation of language and embedding models.β309Updated 2 years ago
- A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.β806Updated 9 months ago
- This is the official code for the paper CodeRL: Mastering Code Generation through Pretrained Models and Deep Reinforcement Learning (Neurβ¦β527Updated 3 months ago
- Convolutions for Sequence Modelingβ877Updated 10 months ago
- A repository for research on medium sized language models.β494Updated last week
- Ungreedy subword tokenizer and vocabulary trainer for Python, Go & Javascriptβ576Updated 9 months ago
- The hub for EleutherAI's work on interpretability and learning dynamicsβ2,463Updated last month
- Language Modeling with the H3 State Space Modelβ520Updated last year
- An open collection of methodologies to help with successful training of large language models.β486Updated last year
- Ongoing research training transformer models at scaleβ386Updated 8 months ago
- β1,004Updated last year
- Tools for understanding how transformer predictions are built layer-by-layerβ488Updated 10 months ago
- A framework for the evaluation of autoregressive code generation language models.β932Updated 5 months ago
- This repository contains code for extending the Stanford Alpaca synthetic instruction tuning to existing instruction-tuned models such asβ¦β351Updated last year
- MiniLLM is a minimal system for running modern LLMs on consumer-grade GPUsβ904Updated last year
- LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale Instructionsβ821Updated last year
- An open-source implementation of Google's PaLM modelsβ819Updated 10 months ago
- [NeurIPS '23 Spotlight] Thought Cloning: Learning to Think while Acting by Imitating Human Thinkingβ265Updated 9 months ago
- [NeurIPS 2023] MeZO: Fine-Tuning Language Models with Just Forward Passes. https://arxiv.org/abs/2305.17333β1,101Updated last year
- β270Updated 2 years ago
- Implementation of MEGABYTE, Predicting Million-byte Sequences with Multiscale Transformers, in Pytorchβ641Updated 4 months ago
- Inference code for Persimmon-8Bβ415Updated last year
- A prize for finding tasks that cause large language models to show inverse scalingβ612Updated last year
- Code for fine-tuning Platypus fam LLMs using LoRAβ629Updated last year
- Salesforce open-source LLMs with 8k sequence length.β717Updated 2 months ago
- Fine-tune mistral-7B on 3090s, a100s, h100sβ710Updated last year