schinger / FullLLM
Full stack LLM (Pre-training/finetuning, PPO(RLHF), Inference, Quant, etc.)
☆19Updated 2 months ago
Alternatives and similar repositories for FullLLM:
Users that are interested in FullLLM are comparing it to the libraries listed below
- 基于DPO算法微调语言大模型,简单好上手。☆37Updated 9 months ago
- pytorch分布式训练☆65Updated last year
- Code for a New Loss for Mitigating the Bias of Learning Difficulties in Generative Language Models☆62Updated 2 months ago
- ☆40Updated 2 months ago
- ☆16Updated last year
- llama,chatglm 等模型 的微调☆88Updated 9 months ago
- 擂台赛3-大规模预训练调优比赛的示例代码与baseline实现☆38Updated 2 years ago
- text classification compitioin top 10 strategy☆17Updated 3 years ago
- ☆67Updated last year
- NTK scaled version of ALiBi position encoding in Transformer.☆67Updated last year
- ☆40Updated 8 months ago
- BLOOM 模型的指令微调☆24Updated last year
- Reinforcement Learning in LLM and NLP.☆34Updated last week
- 记录NLP、CV、搜索、推荐等AI岗位最新情况。☆29Updated 2 years ago
- ☆143Updated 9 months ago
- ChatGLM-6B添加了RLHF的实现,以及部分核心代码的逐行讲解 ,实例部分是做了个新闻短标题的生成,以及指定context推荐的RLHF的实现☆82Updated last year
- [ICML'2024] Can AI Assistants Know What They Don't Know?☆80Updated last year
- ☆140Updated last year
- [EMNLP2024] Aligning Large Language Models on Information Extraction☆46Updated 5 months ago
- ☆18Updated 3 years ago
- ☆22Updated last year
- ChatGPT相关资源汇总☆55Updated 2 years ago
- This is a repo for showcasing using MCTS with LLMs to solve gsm8k problems☆74Updated last month
- 2023全球智能汽车AI挑战赛——赛道一:AI大模型检索问答, 75+ baseline☆57Updated last year
- deepspeed+trainer简单高效实现多卡微调大模型☆124Updated last year
- Clustering and Ranking: Diversity-preserved Instruction Selection through Expert-aligned Quality Estimation☆78Updated 5 months ago
- 使用单个24G显卡,从0开始训练LLM☆53Updated 6 months ago
- 基于Bart语言模型的指针生成网络,用于中文语法纠错任务☆15Updated 2 years ago
- Finetuning LLaMA with RLHF (Reinforcement Learning with Human Feedback) based on DeepSpeed Chat☆115Updated last year
- ☆10Updated last year