pyember / emberLinks
☆234Updated 5 months ago
Alternatives and similar repositories for ember
Users that are interested in ember are comparing it to the libraries listed below
Sorting:
- Storing long contexts in tiny caches with self-study☆218Updated last month
- Training-Ready RL Environments + Evals☆182Updated last week
- ☆107Updated this week
- A scalable asynchronous reinforcement learning implementation with in-flight weight updates.☆316Updated last week
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆189Updated 8 months ago
- ArcticTraining is a framework designed to simplify and accelerate the post-training process for large language models (LLMs)☆257Updated this week
- rl from zero pretrain, can it be done? yes.☆281Updated 2 months ago
- Async RL Training at Scale☆867Updated this week
- ☆136Updated 8 months ago
- Library for text-to-text regression, applicable to any input string representation and allows pretraining and fine-tuning over multiple r…☆295Updated this week
- Long context evaluation for large language models☆224Updated 9 months ago
- ☆144Updated 2 months ago
- Simple & Scalable Pretraining for Neural Architecture Research☆304Updated last month
- Training API and CLI☆248Updated this week
- Super basic implementation (gist-like) of RLMs with REPL environments.☆273Updated last month
- SIMD quantization kernels☆92Updated 2 months ago
- Open source interpretability artefacts for R1.☆164Updated 7 months ago
- Atropos is a Language Model Reinforcement Learning Environments framework for collecting and evaluating LLM trajectories through diverse …☆760Updated this week
- PCCL (Prime Collective Communications Library) implements fault tolerant collective communications over IP☆140Updated 2 months ago
- code for training & evaluating Contextual Document Embedding models☆201Updated 6 months ago
- Inference-time scaling for LLMs-as-a-judge.☆314Updated last month
- ☆456Updated last week
- ☆59Updated 10 months ago
- PyTorch-native post-training at scale☆549Updated last week
- The code for the paper ROUTERBENCH: A Benchmark for Multi-LLM Routing System☆151Updated last year
- ☆917Updated last month
- A zero-to-one guide on scaling modern transformers with n-dimensional parallelism.☆105Updated 2 months ago
- Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.☆329Updated last year
- Accelerating your LLM training to full speed! Made with ❤️ by ServiceNow Research☆262Updated last week
- Open-source framework for the research and development of foundation models.☆640Updated this week