catid / lllm
Latent Large Language Models
☆16Updated 3 months ago
Related projects ⓘ
Alternatives and complementary repositories for lllm
- [WIP] Transformer to embed Danbooru labelsets☆13Updated 7 months ago
- Training hybrid models for dummies.☆15Updated 3 weeks ago
- ☆36Updated 3 months ago
- A library for simplifying fine tuning with multi gpu setups in the Huggingface ecosystem.☆15Updated 3 weeks ago
- Using modal.com to process FineWeb-edu data☆19Updated 2 months ago
- Repository containing the SPIN experiments on the DIBT 10k ranked prompts☆23Updated 8 months ago
- An example implementation of RLHF (or, more accurately, RLAIF) built on MLX and HuggingFace.☆21Updated 5 months ago
- GoldFinch and other hybrid transformer components☆40Updated 4 months ago
- Jax like function transformation engine but micro, microjax☆26Updated 3 weeks ago
- Implementation of Spectral State Space Models☆17Updated 9 months ago
- ☆11Updated last month
- ☆20Updated 3 weeks ago
- alternative way to calculating self attention☆18Updated 5 months ago
- Unleash the full potential of exascale LLMs on consumer-class GPUs, proven by extensive benchmarks, with no long-term adjustments and min…☆23Updated last week
- QLoRA for Masked Language Modeling☆20Updated last year
- ☆41Updated 2 weeks ago
- ☆20Updated 9 months ago
- ☆22Updated last year
- implementation of https://arxiv.org/pdf/2312.09299☆19Updated 4 months ago
- utilities for loading and running text embeddings with onnx☆39Updated 3 months ago
- The Benefits of a Concise Chain of Thought on Problem Solving in Large Language Models☆20Updated 9 months ago
- Official repository for the paper "Approximating Two-Layer Feedforward Networks for Efficient Transformers"☆36Updated last year
- Implementation of SelfExtend from the paper "LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning" from Pytorch and Zeta☆13Updated last week
- ☆49Updated 8 months ago
- A clone of OpenAI's Tokenizer page for HuggingFace Models☆44Updated last year
- A public implementation of the ReLoRA pretraining method, built on Lightning-AI's Pytorch Lightning suite.☆33Updated 8 months ago
- look how they massacred my boy☆58Updated last month
- Efficient Dictionary Learning with Switch Sparse Autoencoders (SAEs)☆12Updated last month
- LLM reads a paper and produce a working prototype☆36Updated 2 weeks ago
- ☆57Updated 11 months ago