CSHaitao / LexEval
LexEval: A Comprehensive Benchmark for Evaluating Large Language Models in Legal Domain
☆61Updated 4 months ago
Alternatives and similar repositories for LexEval:
Users that are interested in LexEval are comparing it to the libraries listed below
- This is the code repo for our paper "Autonomously Knowledge Assimilation and Accommodation through Retrieval-Augmented Agents".☆104Updated 5 months ago
- ☆47Updated last month
- Code for Search-in-the-Chain: Towards Accurate, Credible and Traceable Large Language Models for Knowledge-intensive Tasks☆55Updated 11 months ago
- [EMNLP 2024 (Oral)] Leave No Document Behind: Benchmarking Long-Context LLMs with Extended Multi-Doc QA☆118Updated 4 months ago
- Open source code of the paper: "OmniEval: An Omnidirectional and Automatic RAG Evaluation Benchmark in Financial Domain"☆53Updated 3 months ago
- ☆53Updated 5 months ago
- The demo, code and data of FollowRAG☆70Updated 3 months ago
- Code and data for CoachLM, an automatic instruction revision approach LLM instruction tuning.☆61Updated last year
- [Preprint] Learning to Filter Context for Retrieval-Augmented Generaton☆190Updated 11 months ago
- Github repository for "RAGTruth: A Hallucination Corpus for Developing Trustworthy Retrieval-Augmented Language Models"☆161Updated 3 months ago
- [IJCAI 2024] FactCHD: Benchmarking Fact-Conflicting Hallucination Detection☆86Updated 11 months ago
- Small Models, Big Insights: Leveraging Slim Proxy Models To Decide When and What to Retrieve for LLMs (ACL 2024)☆59Updated 5 months ago
- The GitHub repository for the paper "Self-prompted Chain-of-Thought on Large Language Models for Open-domain Multi-hop Reasoning" accepte…☆18Updated last year
- [ACL'24] Superfiltering: Weak-to-Strong Data Filtering for Fast Instruction-Tuning☆147Updated 6 months ago
- AutoCoA (Automatic generation of Chain-of-Action) is an agent model framework that enhances the multi-turn tool usage capability of reaso…☆68Updated last week
- The code and data of DPA-RAG☆58Updated 2 months ago
- MTU-Bench: A Multi-granularity Tool-Use Benchmark for Large Language Models☆40Updated last month
- Test-time compute in information retrieval☆20Updated 2 weeks ago
- RECOMP: Improving Retrieval-Augmented LMs with Compression and Selective Augmentation.☆123Updated 8 months ago
- Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]☆137Updated 5 months ago
- [ICLR 2025] InstructRAG: Instructing Retrieval-Augmented Generation via Self-Synthesized Rationales☆78Updated last month
- PGRAG☆47Updated 8 months ago
- EMNLP'23 survey: a curation of awesome papers and resources on refreshing large language models (LLMs) without expensive retraining.☆132Updated last year
- The official repo for our paper: LegalAgentBench: Evaluating LLM Agents in Legal Domainl☆17Updated 2 months ago
- Code implementation of synthetic continued pretraining☆95Updated 2 months ago
- ☆28Updated 4 months ago
- ☆40Updated 2 months ago
- The code of arxiv paper: "CoT-based Synthesizer: Enhancing LLM Performance through Answer Synthesis"☆23Updated 2 months ago
- BRIGHT: A Realistic and Challenging Benchmark for Reasoning-Intensive Retrieval☆91Updated last month
- Source code for our paper: "Put Your Money Where Your Mouth Is: Evaluating Strategic Planning and Execution of LLM Agents in an Auction A…☆44Updated last year