bjoernpl / GermanBenchmarkLinks
A repository containing the code for translating popular LLM benchmarks to German.
☆30Updated 2 years ago
Alternatives and similar repositories for GermanBenchmark
Users that are interested in GermanBenchmark are comparing it to the libraries listed below
Sorting:
- Code for Zero-Shot Tokenizer Transfer☆138Updated 8 months ago
- A framework for few-shot evaluation of autoregressive language models.☆13Updated last year
- Minimum Bayes Risk Decoding for Hugging Face Transformers☆60Updated last year
- Official implementation of "GPT or BERT: why not both?"☆60Updated 2 months ago
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆107Updated 6 months ago
- Manage scalable open LLM inference endpoints in Slurm clusters☆273Updated last year
- ☆76Updated last year
- Official Code for M-RᴇᴡᴀʀᴅBᴇɴᴄʜ: Evaluating Reward Models in Multilingual Settings (ACL 2025 Main)☆35Updated 4 months ago
- Prune transformer layers☆69Updated last year
- ☆72Updated 2 years ago
- Erasing concepts from neural representations with provable guarantees☆236Updated 8 months ago
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆256Updated last year
- Let's build better datasets, together!☆262Updated 9 months ago
- BABILong is a benchmark for LLM evaluation using the needle-in-a-haystack approach.☆213Updated last month
- Okapi: Instruction-tuned Large Language Models in Multiple Languages with Reinforcement Learning from Human Feedback☆97Updated 2 years ago
- Datasets collection and preprocessings framework for NLP extreme multitask learning☆187Updated 3 months ago
- Implementation of CALM from the paper "LLM Augmented LLMs: Expanding Capabilities through Composition", out of Google Deepmind☆177Updated last year
- [Data + code] ExpertQA : Expert-Curated Questions and Attributed Answers☆133Updated last year
- code for training & evaluating Contextual Document Embedding models☆197Updated 4 months ago
- ☆52Updated 8 months ago
- ☆39Updated last year
- nanoGPT-like codebase for LLM training☆108Updated 4 months ago
- Evaluation pipeline for the BabyLM Challenge 2023.☆76Updated last year
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆107Updated 2 years ago
- State-of-the-art paired encoder and decoder models (17M-1B params)☆50Updated 2 months ago
- RuLES: a benchmark for evaluating rule-following in language models☆234Updated 7 months ago
- ☆65Updated 2 years ago
- Code repository for the c-BTM paper☆107Updated 2 years ago
- ☆54Updated 2 years ago
- Scaling Data-Constrained Language Models☆342Updated 3 months ago