spapicchio / QATCHLinks
Official implementation of QATCH: Benchmarking SQL-centric tasks with Table Representation Learning Models on Your Data
β32Updated 5 months ago
Alternatives and similar repositories for QATCH
Users that are interested in QATCH are comparing it to the libraries listed below
Sorting:
- Interpretability for sequence generation models π πβ450Updated 2 weeks ago
- SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Modelsβ586Updated last year
- β189Updated 5 months ago
- β371Updated 2 years ago
- Powerful unsupervised domain adaptation method for dense retrieval. Requires only unlabeled corpus and yields massive improvement: "GPL: β¦β338Updated 2 years ago
- Train and Infer Powerful Sentence Embeddings with AnglE | π₯ SOTA on STS and MTEB Leaderboardβ568Updated 2 months ago
- Repository for EMNLP 2022 Paper: Towards a Unified Multi-Dimensional Evaluator for Text Generationβ214Updated last year
- Code for paper "G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment"β398Updated last year
- Train Llama 2 & 3 on the SQuAD v2 task as an example of how to specialize a generalized (foundation) model.β53Updated last year
- Multilingual Large Language Models Evaluation Benchmarkβ133Updated last year
- Calculate perplexity on a text with pre-trained language models. Support MLM (eg. DeBERTa), recurrent LM (eg. GPT3), and encoder-decoder β¦β164Updated 6 months ago
- ACL2023 - AlignScore, a metric for factual consistency evaluation.β147Updated last year
- SemEval2024-task8: Multidomain, Multimodel and Multilingual Machine-Generated Text Detectionβ78Updated last year
- [EMNLP 2023] Enabling Large Language Models to Generate Text with Citations. Paper: https://arxiv.org/abs/2305.14627β502Updated last year
- The implementation of "RQUGE: Reference-Free Metric for Evaluating Question Generation by Answering the Question" [ACL 2023]β16Updated last year
- Code and model release for the paper "Task-aware Retrieval with Instructions" by Asai et al.β165Updated 2 years ago
- A package to evaluate factuality of long-form generation. Original implementation of our EMNLP 2023 paper "FActScore: Fine-grained Atomicβ¦β411Updated 8 months ago
- Source Code of Paper "GPTScore: Evaluate as You Desire"β257Updated 2 years ago
- Multilingual/multidomain question generation datasets, models, and python library for question generation.β370Updated last year
- Code repository for supporting the paper "Atlas Few-shot Learning with Retrieval Augmented Language Models",(https//arxiv.org/abs/2208.03β¦β551Updated 2 years ago
- Code for Multilingual Eval of Generative AI paper published at EMNLP 2023β71Updated last year
- All-in-one text de-duplicationβ736Updated 3 months ago
- 𦫠BEAVER: An Enterprise Benchmark for Text-to-SQLβ24Updated 6 months ago
- Github repository for "RAGTruth: A Hallucination Corpus for Developing Trustworthy Retrieval-Augmented Language Models"β215Updated last year
- The prime repository for state-of-the-art Multilingual Question Answering research and development.β740Updated 3 months ago
- β401Updated last week
- This is the repository of HaluEval, a large-scale hallucination evaluation benchmark for Large Language Models.β536Updated last year
- Code for T-Few from "Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning"β456Updated 2 years ago
- GEMBA β GPT Estimation Metric Based Assessmentβ135Updated this week
- Efficient Attention for Long Sequence Processingβ98Updated 2 years ago