OpenBMB / ModelCenterLinks
Efficient, Low-Resource, Distributed transformer implementation based on BMTrain
☆258Updated last year
Alternatives and similar repositories for ModelCenter
Users that are interested in ModelCenter are comparing it to the libraries listed below
Sorting:
- Finetuning LLaMA with RLHF (Reinforcement Learning with Human Feedback) based on DeepSpeed Chat☆114Updated 2 years ago
- Naive Bayes-based Context Extension☆325Updated 7 months ago
- Implementation of Chinese ChatGPT☆286Updated last year
- Collaborative Training of Large Language Models in an Efficient Way☆417Updated 11 months ago
- ☆280Updated last year
- 中文 Instruction tuning datasets☆133Updated last year
- A Massive Multi-Level Multi-Subject Knowledge Evaluation benchmark☆102Updated 2 years ago
- 中文图书语料MD5链接☆216Updated last year
- A collection of phenomenons observed during the scaling of big foundation models, which may be developed into consensus, principles, or l…☆284Updated last year
- Model Compression for Big Models☆164Updated 2 years ago
- [NAACL'24] Self-data filtering of LLM instruction-tuning data using a novel perplexity-based difficulty score, without using any other mo…☆381Updated last month
- ☆308Updated 2 years ago
- ☆162Updated 2 years ago
- ☆173Updated last year
- ☆459Updated last year
- 大模型多维度中文对齐评测基准 (ACL 2024)☆402Updated 11 months ago
- ☆83Updated last year
- 语言模型中文认知能力分析☆236Updated last year
- 怎么训练一个LLM分词器☆151Updated 2 years ago
- 对ChatGLM直接使用RLHF提升或降低目标输出概率|Modify ChatGLM output with only RLHF☆195Updated 2 years ago
- Efficient Training (including pre-training and fine-tuning) for Big Models☆604Updated 2 months ago
- ☆144Updated last year
- ☆172Updated 2 years ago
- OpenLLMWiki: Docs of OpenLLMAI. Survey, reproduction and domain/task adaptation of open source chatgpt alternatives/implementations. PiXi…☆260Updated 7 months ago
- ☆128Updated 2 years ago
- ☆96Updated last year
- MEASURING MASSIVE MULTITASK CHINESE UNDERSTANDING☆88Updated last year
- LLM Zoo collects information of various open- and close-sourced LLMs☆271Updated last year
- train llama on a single A100 80G node using 🤗 transformers and 🚀 Deepspeed Pipeline Parallelism☆224Updated last year
- A collection for math word problem (MWP) works, including datasets, algorithms and so on.☆44Updated last year