Efficient, Low-Resource, Distributed transformer implementation based on BMTrain
☆267Nov 27, 2023Updated 2 years ago
Alternatives and similar repositories for ModelCenter
Users that are interested in ModelCenter are comparing it to the libraries listed below
Sorting:
- Model Compression for Big Models☆168Jun 30, 2023Updated 2 years ago
- Efficient Training (including pre-training and fine-tuning) for Big Models☆621Oct 27, 2025Updated 4 months ago
- Live Training for Open-source Big Models☆505May 30, 2023Updated 2 years ago
- Efficient Inference for Big Models☆587Jan 24, 2023Updated 3 years ago
- A plug-and-play library for parameter-efficient-tuning (Delta Tuning)☆1,039Sep 19, 2024Updated last year
- An Open-Source Package for Information Retrieval☆168Updated this week
- Tool Learning for Big Models, Open-Source Solutions of ChatGPT-Plugins☆2,786Dec 5, 2023Updated 2 years ago
- [ICLR 2023] Code for our paper "Selective Annotation Makes Language Models Better Few-Shot Learners"☆109Jul 15, 2023Updated 2 years ago
- reStructured Pre-training☆99Dec 22, 2022Updated 3 years ago
- This repo contains the code for Late Prompt Tuning.☆12Dec 22, 2025Updated 2 months ago
- EMNLP 2022: Finding Dataset Shortcuts with Grammar Induction https://arxiv.org/abs/2210.11560☆58Feb 28, 2025Updated last year
- EVA: Large-scale Pre-trained Chit-Chat Models☆306Mar 11, 2023Updated 2 years ago
- Repo for paper "Tell Me More! Towards Implicit User Intention Understanding of Language Model Driven Agents"☆61Feb 20, 2024Updated 2 years ago
- Gaokao Benchmark for AI☆107Jul 8, 2022Updated 3 years ago
- A novel method to tune language models. Codes and datasets for paper ``GPT understands, too''.☆938Oct 6, 2022Updated 3 years ago
- This repository contains the code for the paper: SirLLM: Streaming Infinite Retentive LLM☆60May 28, 2024Updated last year
- Large-scale, Informative, and Diverse Multi-round Chat Data (and Models)☆2,789Mar 13, 2024Updated last year
- Princeton NLP's pre-training library based on fairseq with DeepSpeed kernel integration 🚃☆116Oct 27, 2022Updated 3 years ago
- ☆99Jul 25, 2023Updated 2 years ago
- [EMNLP 2022] Training Language Models with Memory Augmentation https://arxiv.org/abs/2205.12674☆195Jun 14, 2023Updated 2 years ago
- Code for CPM-2 Pre-Train☆158Mar 18, 2023Updated 2 years ago
- Apply Iprompt on GLM with innovative new methods. Currently support Chinese QA, English QA and Chinese poem generation.☆20Jun 16, 2022Updated 3 years ago
- An Open-Source Framework for Prompt-Learning.☆4,828Jul 16, 2024Updated last year
- Vision-Language Models Toolbox: Your all-in-one solution for multimodal research and experimentation☆12Feb 16, 2025Updated last year
- The code of our paper "InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Mem…☆396Apr 20, 2024Updated last year
- Benchmarking large language models' complex reasoning ability with chain-of-thought prompting☆2,766Aug 4, 2024Updated last year
- [ACL 2021] LM-BFF: Better Few-shot Fine-tuning of Language Models https://arxiv.org/abs/2012.15723☆730Aug 29, 2022Updated 3 years ago
- Expanding natural instructions☆1,035Dec 11, 2023Updated 2 years ago
- Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]☆588Dec 9, 2024Updated last year
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆1,435Mar 20, 2024Updated last year
- ☆43Aug 17, 2024Updated last year
- 百亿参数的中英文双语基座大模型☆2,413Jul 28, 2023Updated 2 years ago
- ☆84Apr 18, 2024Updated last year
- A large-scale, fine-grained, diverse preference dataset (and models).☆363Dec 29, 2023Updated 2 years ago
- This is the official implementation of "Progressive-Hint Prompting Improves Reasoning in Large Language Models"☆209Oct 11, 2023Updated 2 years ago
- ☆22Dec 11, 2025Updated 2 months ago
- KuaiSearch PERKS☆12Nov 16, 2021Updated 4 years ago
- Official codes for ACL 2023 paper "WebCPM: Interactive Web Search for Chinese Long-form Question Answering"☆912Nov 25, 2023Updated 2 years ago
- Mengzi Pretrained Models☆540Nov 29, 2022Updated 3 years ago