hgaurav2k / JEEBenchLinks
Repository for the code and dataset for the paper: "Have LLMs Advanced enough? Towards Harder Problem Solving Benchmarks For Large Language Models"
☆39Updated last year
Alternatives and similar repositories for JEEBench
Users that are interested in JEEBench are comparing it to the libraries listed below
Sorting:
- Google Research☆46Updated 3 years ago
- ☆52Updated last year
- Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Fl…☆75Updated last year
- [NeurIPS 2023] Learning Transformer Programs☆162Updated last year
- ☆80Updated 7 months ago
- A repository for transformer critique learning and generation☆88Updated last year
- A library to create and manage configuration files, especially for machine learning projects.☆80Updated 3 years ago
- ☆23Updated this week
- ☆76Updated last year
- Code Release for "Broken Neural Scaling Laws" (BNSL) paper☆59Updated 2 years ago
- The GitHub repo for Goal Driven Discovery of Distributional Differences via Language Descriptions☆71Updated 2 years ago
- code for "Natural Language to Code Translation with Execution"☆41Updated 2 years ago
- Official code release for the paper Coder Reviewer Reranking for Code Generation.☆45Updated 2 years ago
- ☆27Updated 2 years ago
- A dataset of LLM-generated chain-of-thought steps annotated with mistake location.☆82Updated last year
- CodeUltraFeedback: aligning large language models to coding preferences (TOSEM 2025)☆72Updated last year
- ☆44Updated 11 months ago
- Embroid: Unsupervised Prediction Smoothing Can Improve Few-Shot Classification☆11Updated 2 years ago
- Understanding the correlation between different LLM benchmarks☆29Updated last year
- ☆85Updated last year
- Minimum Description Length probing for neural network representations☆20Updated 9 months ago
- Reference implementation for Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model☆43Updated last month
- This repository contains some of the code used in the paper "Training Language Models with Langauge Feedback at Scale"☆27Updated 2 years ago
- Codes and files for the paper Are Emergent Abilities in Large Language Models just In-Context Learning☆33Updated 9 months ago
- Code of ICLR paper: https://openreview.net/forum?id=-cqvvvb-NkI☆94Updated 2 years ago
- ☆55Updated 2 years ago
- Official Repository of Pretraining Without Attention (BiGS), BiGS is the first model to achieve BERT-level transfer learning on the GLUE …☆114Updated last year
- Utilities for Training Very Large Models☆58Updated last year
- Code for the paper "The Impact of Positional Encoding on Length Generalization in Transformers", NeurIPS 2023☆136Updated last year
- We view Large Language Models as stochastic language layers in a network, where the learnable parameters are the natural language prompts…☆94Updated last year