character-ai / MuKoe
☆51Updated 6 months ago
Related projects ⓘ
Alternatives and complementary repositories for MuKoe
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆69Updated last year
- ☆22Updated last year
- Collection of autoregressive model implementation☆67Updated this week
- ☆57Updated 11 months ago
- Make triton easier☆41Updated 5 months ago
- LLM reads a paper and produce a working prototype☆36Updated last week
- An example implementation of RLHF (or, more accurately, RLAIF) built on MLX and HuggingFace.☆21Updated 5 months ago
- Jax like function transformation engine but micro, microjax☆26Updated 3 weeks ago
- ☆40Updated this week
- ☆36Updated 3 months ago
- A pipeline for LLM knowledge distillation☆78Updated 3 months ago
- A place to store reusable transformer components of my own creation or found on the interwebs☆44Updated 2 weeks ago
- Score LLM pretraining data with classifiers☆54Updated last year
- RWKV-7: Surpassing GPT☆45Updated this week
- ☆49Updated 8 months ago
- ☆156Updated last week
- inference code for mixtral-8x7b-32kseqlen☆98Updated 11 months ago
- Fast approximate inference on a single GPU with sparsity aware offloading☆38Updated 10 months ago
- look how they massacred my boy☆58Updated last month
- Evaluating LLMs with CommonGen-Lite☆85Updated 8 months ago
- ring-attention experiments☆97Updated last month
- ☆41Updated 2 weeks ago
- Implementation of nougat that focuses on processing pdf locally.☆73Updated 6 months ago
- ☆99Updated last month
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆84Updated this week
- Losslessly encode text natively with arithmetic coding and HuggingFace Transformers☆71Updated 3 months ago
- Using multiple LLMs for ensemble Forecasting☆16Updated 10 months ago
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆104Updated last month
- Routing on Random Forest (RoRF)☆84Updated last month