OpenBMB / ModelCenterLinks
Efficient, Low-Resource, Distributed transformer implementation based on BMTrain
☆265Updated 2 years ago
Alternatives and similar repositories for ModelCenter
Users that are interested in ModelCenter are comparing it to the libraries listed below
Sorting:
- Finetuning LLaMA with RLHF (Reinforcement Learning with Human Feedback) based on DeepSpeed Chat☆117Updated 2 years ago
- Model Compression for Big Models☆167Updated 2 years ago
- Collaborative Training of Large Language Models in an Efficient Way☆417Updated last year
- ☆282Updated last year
- Naive Bayes-based Context Extension☆326Updated last year
- Implementation of Chinese ChatGPT☆288Updated 2 years ago
- 中文 Instruction tuning datasets☆141Updated last year
- 中文图书语料MD5链接☆218Updated 2 years ago
- ☆459Updated last year
- 语言模型中文认知能力分析☆236Updated 2 years ago
- ☆313Updated 2 years ago
- ☆184Updated 2 years ago
- A collection of phenomenons observed during the scaling of big foundation models, which may be developed into consensus, principles, or l…☆284Updated 2 years ago
- 怎么训练一个LLM分词器☆153Updated 2 years ago
- OpenLLMWiki: Docs of OpenLLMAI. Survey, reproduction and domain/task adaptation of open source chatgpt alternatives/implementations. PiXi…☆262Updated last year
- Efficient Training (including pre-training and fine-tuning) for Big Models☆618Updated 3 months ago
- ☆84Updated 2 years ago
- [NAACL'24] Self-data filtering of LLM instruction-tuning data using a novel perplexity-based difficulty score, without using any other mo…☆415Updated 7 months ago
- 大语言模型指令调优工具(支持 FlashAttention)☆177Updated 2 years ago
- ☆129Updated 2 years ago
- ☆173Updated 2 years ago
- 大模型多维度中文对齐评测基准 (ACL 2024)☆421Updated 3 months ago
- 对ChatGLM直接使用RLHF提升或降低目标输出概率|Modify ChatGLM output with only RLHF☆197Updated 2 years ago
- [EMNLP 2023] Lion: Adversarial Distillation of Proprietary Large Language Models☆212Updated last year
- MEASURING MASSIVE MULTITASK CHINESE UNDERSTANDING☆89Updated last year
- LLM Zoo collects information of various open- and close-sourced LLMs☆271Updated 2 years ago
- train llama on a single A100 80G node using 🤗 transformers and 🚀 Deepspeed Pipeline Parallelism☆224Updated 2 years ago
- A framework for cleaning Chinese dialog data☆273Updated 4 years ago
- The complete training code of the open-source high-performance Llama model, including the full process from pre-training to RLHF.☆67Updated 2 years ago
- ☆164Updated 2 years ago