pyember / emberLinks
☆225Updated 3 months ago
Alternatives and similar repositories for ember
Users that are interested in ember are comparing it to the libraries listed below
Sorting:
- Storing long contexts in tiny caches with self-study☆192Updated 3 weeks ago
- Training-Ready RL Environments + Evals☆116Updated this week
- ☆103Updated 2 weeks ago
- Atropos is a Language Model Reinforcement Learning Environments framework for collecting and evaluating LLM trajectories through diverse …☆701Updated this week
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆184Updated 6 months ago
- ☆133Updated 6 months ago
- PyTorch Single Controller☆425Updated this week
- Post-training with Tinker☆550Updated this week
- Async RL Training at Scale☆669Updated this week
- ☆142Updated 3 weeks ago
- Library for text-to-text regression, applicable to any input string representation and allows pretraining and fine-tuning over multiple r…☆236Updated this week
- Long context evaluation for large language models☆222Updated 7 months ago
- Simple & Scalable Pretraining for Neural Architecture Research☆296Updated last month
- Inference-time scaling for LLMs-as-a-judge.☆299Updated last month
- ☆57Updated 8 months ago
- rl from zero pretrain, can it be done? yes.☆274Updated last week
- ☆773Updated 3 weeks ago
- PCCL (Prime Collective Communications Library) implements fault tolerant collective communications over IP☆123Updated 3 weeks ago
- ArcticTraining is a framework designed to simplify and accelerate the post-training process for large language models (LLMs)☆219Updated this week
- Accelerating your LLM training to full speed! Made with ❤️ by ServiceNow Research☆237Updated this week
- Matrix (Multi-Agent daTa geneRation Infra and eXperimentation framework) is a versatile engine for multi-agent conversational data genera…☆94Updated this week
- SIMD quantization kernels☆87Updated 3 weeks ago
- A framework for optimizing DSPy programs with RL☆185Updated last week
- Open source interpretability artefacts for R1.☆160Updated 5 months ago
- j1-micro (1.7B) & j1-nano (600M) are absurdly tiny but mighty reward models.☆98Updated 2 months ago
- The code for the paper ROUTERBENCH: A Benchmark for Multi-LLM Routing System☆142Updated last year
- Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.☆321Updated 11 months ago
- code for training & evaluating Contextual Document Embedding models☆197Updated 4 months ago
- Official repository for "Scaling Retrieval-Based Langauge Models with a Trillion-Token Datastore".☆216Updated 2 months ago
- ⚖️ Awesome LLM Judges ⚖️☆128Updated 5 months ago