facebookresearch / BigOBench
BigOBench assesses the capacity of Large Language Models (LLMs) to comprehend time-space computational complexity of input or generated code.
☆32Updated 3 weeks ago
Alternatives and similar repositories for BigOBench
Users that are interested in BigOBench are comparing it to the libraries listed below
Sorting:
- ☆17Updated 3 months ago
- ☆24Updated 3 months ago
- ☆31Updated 4 months ago
- Improving Your Model Ranking on Chatbot Arena by Vote Rigging (ICML 2025)☆20Updated 2 months ago
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆57Updated 8 months ago
- ☆27Updated 3 weeks ago
- Official repository of "LiNeS: Post-training Layer Scaling Prevents Forgetting and Enhances Model Merging"☆26Updated 6 months ago
- [NeurIPS 2024] Goldfish Loss: Mitigating Memorization in Generative LLMs☆84Updated 5 months ago
- ☆24Updated 3 weeks ago
- Exploration of automated dataset selection approaches at large scales.☆40Updated 2 months ago
- An official implementation of "Catastrophic Failure of LLM Unlearning via Quantization" (ICLR 2025)☆26Updated 2 months ago
- PaCE: Parsimonious Concept Engineering for Large Language Models (NeurIPS 2024)☆35Updated 6 months ago
- Stanford NLP Python library for benchmarking the utility of LLM interpretability methods☆77Updated last month
- This is the oficial repository for "Safer-Instruct: Aligning Language Models with Automated Preference Data"☆17Updated last year
- AIR-Bench 2024 is a safety benchmark that aligns with emerging government regulations and company policies☆19Updated 8 months ago
- Code for reproducing our paper "Not All Language Model Features Are Linear"☆73Updated 5 months ago
- Privacy backdoors☆51Updated last year
- Adversarial Attacks on GPT-4 via Simple Random Search [Dec 2023]☆43Updated last year
- ☆17Updated last week
- ☆29Updated last year
- ☆78Updated 8 months ago
- Code Implementation, Evaluations, Documentation, Links and Resources for Min P paper☆33Updated last month
- Mixture-of-Transformers A Sparse and Scalable Architecture for Multi-Modal Foundation Models. TMLR 2025. 🔗 https//arxiv.org/abs/2411.049…☆31Updated this week
- Code for reproducing our paper "Low Rank Adapting Models for Sparse Autoencoder Features"☆10Updated last month
- ☆15Updated this week
- Knowledge Unlearning for Large Language Models☆25Updated last week
- One Initialization to Rule them All: Fine-tuning via Explained Variance Adaptation☆39Updated 6 months ago
- ☆13Updated 5 months ago
- ☆19Updated 10 months ago
- [NeurIPS 2024 Spotlight] Code and data for the paper "Finding Transformer Circuits with Edge Pruning".☆55Updated 2 months ago