google-deepmind / calmLinks
☆55Updated 4 months ago
Alternatives and similar repositories for calm
Users that are interested in calm are comparing it to the libraries listed below
Sorting:
- Maya: An Instruction Finetuned Multilingual Multimodal Model using Aya☆123Updated 4 months ago
- ☆55Updated last year
- Using open source LLMs to build synthetic datasets for direct preference optimization☆71Updated last year
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆60Updated last year
- ☆17Updated 8 months ago
- Easy to use, High Performant Knowledge Distillation for LLMs☆97Updated 7 months ago
- The Benefits of a Concise Chain of Thought on Problem Solving in Large Language Models☆22Updated last year
- ☆31Updated last year
- ☆120Updated last year
- A public implementation of the ReLoRA pretraining method, built on Lightning-AI's Pytorch Lightning suite.☆35Updated last year
- An unofficial pytorch implementation of 'Efficient Infinite Context Transformers with Infini-attention'☆54Updated last year
- Repository for "I am a Strange Dataset: Metalinguistic Tests for Language Models"☆45Updated last year
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆51Updated last year
- Jina VDR is a multilingual, multi-domain benchmark for visual document retrieval☆37Updated 4 months ago
- Tokun to can tokens☆18Updated 6 months ago
- Improving Text Embedding of Language Models Using Contrastive Fine-tuning☆66Updated last year
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆59Updated 2 months ago
- Modified Beam Search with periodical restart☆12Updated last year
- MEXMA: Token-level objectives improve sentence representations☆42Updated 11 months ago
- 🚢 Data Toolkit for Sailor Language Models☆94Updated 9 months ago
- ☆53Updated 10 months ago
- ☆39Updated last year
- Verifiers for LLM Reinforcement Learning☆80Updated 8 months ago
- [ACL 2025] How Do LLMs Acquire New Knowledge? A Knowledge Circuits Perspective on Continual Pre-Training☆45Updated 5 months ago
- Exploring limitations of LLM-as-a-judge☆19Updated last year
- Luth is a state-of-the-art series of fine-tuned LLMs for French☆40Updated 2 months ago
- LLM reads a paper and produce a working prototype☆60Updated 8 months ago
- Streamline on-policy/off-policy distillation workflows in a few lines of code☆79Updated last week
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆78Updated last year
- ☆20Updated 8 months ago