abacaj / code-eval
Run evaluation on LLMs using human-eval benchmark
β395Updated last year
Alternatives and similar repositories for code-eval:
Users that are interested in code-eval are comparing it to the libraries listed below
- β267Updated last year
- π OctoPack: Instruction Tuning Code Large Language Modelsβ450Updated last week
- Open Source WizardCoder Datasetβ156Updated last year
- [ICLR 2024] Lemur: Open Foundation Models for Language Agentsβ540Updated last year
- β305Updated 8 months ago
- Implementation of paper Data Engineering for Scaling Language Models to 128K Contextβ451Updated 10 months ago
- Code and data for "Lost in the Middle: How Language Models Use Long Contexts"β332Updated last year
- NexusRaven-13B, a new SOTA Open-Source LLM for function calling. This repo contains everything for reproducing our evaluation on NexusRavβ¦β312Updated last year
- An Analytical Evaluation Board of Multi-turn LLM Agentsβ279Updated 8 months ago
- [NeurIPS'24] SelfCodeAlign: Self-Alignment for Code Generationβ294Updated 3 months ago
- This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.β541Updated 11 months ago
- [ACL'24 Outstanding] Data and code for L-Eval, a comprehensive long context language models evaluation benchmarkβ369Updated 7 months ago
- [ICML 2023] Data and code release for the paper "DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation".β233Updated 3 months ago
- RewardBench: the first evaluation tool for reward models.β503Updated this week
- Generative Judge for Evaluating Alignmentβ225Updated last year
- [ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuningβ638Updated 8 months ago
- FireAct: Toward Language Agent Fine-tuningβ265Updated last year
- [NeurIPS 2023 D&B] Code repository for InterCode benchmark https://arxiv.org/abs/2306.14898β205Updated 9 months ago
- β84Updated last year
- Official repository for LongChat and LongEvalβ519Updated 8 months ago
- Generate textbook-quality synthetic LLM pretraining dataβ494Updated last year
- Learning to Compress Prompts with Gist Tokens - https://arxiv.org/abs/2304.08467β274Updated this week
- β¨ RepoBench: Benchmarking Repository-Level Code Auto-Completion Systems - ICLR 2024β144Updated 6 months ago
- β496Updated 2 months ago
- A framework for the evaluation of autoregressive code generation language models.β884Updated 3 months ago
- Build Hierarchical Autonomous Agents through Config. Collaborative Growth of Specialized Agents.β310Updated last year
- Code and data for "MAmmoTH: Building Math Generalist Models through Hybrid Instruction Tuning" (ICLR 2024)β360Updated 5 months ago
- β352Updated last year
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"β296Updated last year
- [COLM 2024] LoraHub: Efficient Cross-Task Generalization via Dynamic LoRA Compositionβ614Updated 6 months ago