ntunlp / OpenSource-LLMs-better-than-OpenAI
Listing all reported open-source LLMs achieving a higher score than proprietary, paying OpenAI models (ChatGPT, GPT-4).
☆69Updated last year
Alternatives and similar repositories for OpenSource-LLMs-better-than-OpenAI:
Users that are interested in OpenSource-LLMs-better-than-OpenAI are comparing it to the libraries listed below
- ☆120Updated 8 months ago
- ☆66Updated last year
- Code for ICLR 2024 paper "CRAFT: Customizing LLMs by Creating and Retrieving from Specialized Toolsets"☆51Updated 8 months ago
- We aim to provide the best references to search, select, and synthesize high-quality and large-quantity data for post-training your LLMs.☆49Updated 4 months ago
- Reformatted Alignment☆114Updated 4 months ago
- List of papers on Self-Correction of LLMs.☆71Updated last month
- Open Implementations of LLM Analyses☆98Updated 4 months ago
- augmented LLM with self reflection☆112Updated last year
- ☆98Updated 2 months ago
- A dataset of LLM-generated chain-of-thought steps annotated with mistake location.☆77Updated 6 months ago
- ☆42Updated 2 months ago
- FuseAI Project☆83Updated 3 weeks ago
- Source code of "Reasons to Reject? Aligning Language Models with Judgments"☆58Updated 11 months ago
- ⏳ ChatLog: Recording and Analysing ChatGPT Across Time☆96Updated 8 months ago
- [NeurIPS2024] OlympicArena: Benchmarking Multi-discipline Cognitive Reasoning for Superintelligent AI☆92Updated 2 months ago
- LongEmbed: Extending Embedding Models for Long Context Retrieval (EMNLP 2024)☆128Updated 3 months ago
- Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]☆135Updated 3 months ago
- [ICLR 2025] InstructRAG: Instructing Retrieval-Augmented Generation via Self-Synthesized Rationales☆72Updated 2 weeks ago
- Scalable Meta-Evaluation of LLMs as Evaluators☆43Updated last year
- Source code for our paper: "Put Your Money Where Your Mouth Is: Evaluating Strategic Planning and Execution of LLM Agents in an Auction A…☆43Updated last year
- Codebase accompanying the Summary of a Haystack paper.☆74Updated 5 months ago
- [EMNLP 2023 Industry Track] A simple prompting approach that enables the LLMs to run inference in batches.☆72Updated 11 months ago
- Official implementation for 'Extending LLMs’ Context Window with 100 Samples'☆76Updated last year
- A simple GPT-based evaluation tool for multi-aspect, interpretable assessment of LLMs.☆83Updated last year
- Official repo of Respond-and-Respond: data, code, and evaluation☆104Updated 6 months ago
- Critique-out-Loud Reward Models☆52Updated 4 months ago
- Code for paper titled "Towards the Law of Capacity Gap in Distilling Language Models"☆99Updated 7 months ago
- ☆34Updated 2 months ago
- a curated list of the role of small models in the LLM era☆91Updated 4 months ago
- 🚢 Data Toolkit for Sailor Language Models☆85Updated 2 months ago