withmartian / leaderboard-backend
Open sourced backend for Martian's LLM Inference Provider Leaderboard
☆18Updated 9 months ago
Alternatives and similar repositories for leaderboard-backend
Users that are interested in leaderboard-backend are comparing it to the libraries listed below
Sorting:
- [EMNLP 2023 Industry Track] A simple prompting approach that enables the LLMs to run inference in batches.☆72Updated last year
- ☆64Updated last year
- Tools for content datamining and NLP at scale☆43Updated 10 months ago
- This is a new metric that can be used to evaluate faithfulness of text generated by LLMs. The work behind this repository can be found he…☆31Updated last year
- NeurIPS 2023 - Cappy: Outperforming and Boosting Large Multi-Task LMs with a Small Scorer☆43Updated last year
- Index of URLs to pdf files all over the internet and scripts☆23Updated 2 years ago
- A collection of reproducible inference engine benchmarks☆30Updated 3 weeks ago
- ☆27Updated 2 weeks ago
- ☆27Updated last month
- Advanced Reasoning Benchmark Dataset for LLMs☆46Updated last year
- Adversarial Training and SFT for Bot Safety Models☆39Updated 2 years ago
- Astraios: Parameter-Efficient Instruction Tuning Code Language Models☆57Updated last year
- Benchmark suite for LLMs from Fireworks.ai☆73Updated this week
- ☆20Updated 11 months ago
- Code for NeurIPS LLM Efficiency Challenge☆58Updated last year
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated last year
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.☆34Updated last week
- Data preparation code for CrystalCoder 7B LLM☆44Updated last year
- ☆45Updated 2 months ago
- Contextual Position Encoding but with some custom CUDA Kernels https://arxiv.org/abs/2405.18719☆22Updated 11 months ago
- Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundry☆41Updated last year
- Repository for CPU Kernel Generation for LLM Inference☆26Updated last year
- ☆33Updated 10 months ago
- In-Context Alignment: Chat with Vanilla Language Models Before Fine-Tuning☆34Updated last year
- Implementation of the paper: "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" from Google in pyTO…☆55Updated 3 weeks ago
- Benchmark for machine learning model online serving (LLM, embedding, Stable-Diffusion, Whisper)☆28Updated last year
- Reasoning by Communicating with Agents☆28Updated 2 weeks ago
- ☆38Updated last year
- ☆50Updated last year
- ☆47Updated 8 months ago