Amplify-Partners / annotation-reading-listLinks
A reading list of relevant papers and projects on foundation model annotation
☆27Updated 3 months ago
Alternatives and similar repositories for annotation-reading-list
Users that are interested in annotation-reading-list are comparing it to the libraries listed below
Sorting:
- ☆38Updated 10 months ago
- Simple repository for training small reasoning models☆31Updated 4 months ago
- A framework for optimizing DSPy programs with RL☆58Updated this week
- j1-micro (1.7B) & j1-nano (600M) are absurdly tiny but mighty reward models.☆74Updated last week
- Small, simple agent task environments for training and evaluation☆18Updated 7 months ago
- Compiling useful links, papers, benchmarks, ideas, etc.☆46Updated 2 months ago
- gzip Predicts Data-dependent Scaling Laws☆35Updated last year
- Collection of LLM completions for reasoning-gym task datasets☆22Updated last week
- Official Repo for InSTA: Towards Internet-Scale Training For Agents☆42Updated last week
- Synthetic data generation and benchmark implementation for "Episodic Memories Generation and Evaluation Benchmark for Large Language Mode…☆45Updated last month
- Sphynx Hallucination Induction☆54Updated 4 months ago
- Official repo for Learning to Reason for Long-Form Story Generation☆60Updated last month
- ☆126Updated 2 months ago
- an open source reproduction of NVIDIA's nGPT (Normalized Transformer with Representation Learning on the Hypersphere)☆100Updated 2 months ago
- ☆29Updated 6 months ago
- look how they massacred my boy☆63Updated 7 months ago
- A framework for pitting LLMs against each other in an evolving library of games ⚔☆32Updated last month
- ☆19Updated last week
- Simple GRPO scripts and configurations.☆58Updated 4 months ago
- ☆131Updated 2 months ago
- ☆22Updated last year
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆53Updated 4 months ago
- Open source interpretability artefacts for R1.☆140Updated last month
- ☆20Updated last year
- [ACL 2024] Do Large Language Models Latently Perform Multi-Hop Reasoning?☆67Updated 2 months ago
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆173Updated 2 months ago
- ☆83Updated 5 months ago
- ☆76Updated last month
- ⚖️ Awesome LLM Judges ⚖️☆103Updated last month
- LLM training in simple, raw C/CUDA☆14Updated 6 months ago