chi2liu / mamba-gpt-3b
It is almost the best 3B model in the current open source industry, surpassing Dolly v2-3b, open lama-3b, and even outperforming the EleutherAI/pythia-12b model in terms of performance. Can refer to open_llm_leaderboard
☆13Updated last year
Alternatives and similar repositories for mamba-gpt-3b:
Users that are interested in mamba-gpt-3b are comparing it to the libraries listed below
- Very minimal (and stateless) agent framework☆41Updated last month
- Low-Rank Adaptation of Large Language Models clean implementation☆8Updated last year
- ☆38Updated last year
- Simple Implementation of a Transformer in the new framework MLX by Apple☆20Updated 3 months ago
- Eh, simple and works.☆27Updated last year
- A data-centric AI package for ML/AI. Get the best high-quality data for the best results. Discord: https://discord.gg/t6ADqBKrdZ☆64Updated last year
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆164Updated last year
- an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated last year
- ☆53Updated 8 months ago
- ☆49Updated 11 months ago
- ☆87Updated last year
- Luann allows you to create a LLM agent,which has complete memory module (long-term memory, short-term memory) and knowledge module(Variou…☆18Updated last week
- ☆52Updated 9 months ago
- Simple Model Similarities Analysis☆21Updated last year
- Generate High Quality textual or multi-modal datasets with Agents☆18Updated last year
- ☆60Updated last year
- ☆16Updated 8 months ago
- inference code for mixtral-8x7b-32kseqlen☆99Updated last year
- Here we collect trick questions and failed tasks for open source LLMs to improve them.☆32Updated last year
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆69Updated last year
- fine tuning mistral 7B using Huggingface, Weights and Biases, Choline, and Vast AI☆37Updated last year
- An open source replication of the stawberry method that leverages Monte Carlo Search with PPO and or DPO☆28Updated last week
- A repository re-creating the PromptBreeder Evolutionary Algorithm from the DeepMind Paper in Python using LMQL as the backend.☆27Updated last year
- Set of scripts to finetune LLMs☆36Updated 10 months ago
- LLM reads a paper and produce a working prototype☆48Updated 2 weeks ago
- ☆20Updated last year
- An OpenAI Completions API compatible server for NLP transformers models☆64Updated last year
- The Next Generation Multi-Modality Superintelligence☆71Updated 5 months ago
- ☆22Updated last year
- LLM based agents with proactive interactions, long-term memory, external tool integration, and local deployment capabilities.☆97Updated this week