evintunador / FractalFormerLinks
A GPT with self-similar nested properties
☆20Updated last year
Alternatives and similar repositories for FractalFormer
Users that are interested in FractalFormer are comparing it to the libraries listed below
Sorting:
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31Updated last year
- The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆41Updated last year
- ☆68Updated last year
- ☆119Updated last year
- Video+code lecture on building nanoGPT from scratch☆68Updated last year
- ☆15Updated 2 years ago
- This is our own implementation of 'Layer Selective Rank Reduction'☆240Updated last year
- ☆112Updated 2 years ago
- Experimental LLM Inference UX to aid in creative writing☆127Updated last year
- MiniHF is an inference, human preference data collection, and fine-tuning tool for local language models. It is intended to help the user…☆183Updated 2 months ago
- look how they massacred my boy☆63Updated last year
- Cerule - A Tiny Mighty Vision Model☆68Updated 2 months ago
- ☆137Updated last year
- ☆27Updated 2 years ago
- 5X faster 60% less memory QLoRA finetuning☆21Updated last year
- Collection of autoregressive model implementation☆85Updated last week
- GPT-2 small trained on phi-like data☆67Updated last year
- Scripts to create your own moe models using mlx☆90Updated last year
- ☆63Updated last year
- All the world is a play, we are but actors in it.☆49Updated 5 months ago
- An unsupervised model merging algorithm for Transformers-based language models.☆108Updated last year
- Maybe the new state of the art vision model? we'll see 🤷♂️☆170Updated 2 years ago
- ☆125Updated last year
- entropix style sampling + GUI☆27Updated last year
- GRadient-INformed MoE☆264Updated last year
- Low-Rank adapter extraction for fine-tuned transformers models☆180Updated last year
- Modeling code for a BitNet b1.58 Llama-style model.☆25Updated last year
- ☆28Updated last year
- Code for ExploreTom☆89Updated 6 months ago
- Full finetuning of large language models without large memory requirements☆94Updated 3 months ago