InternLM / InternLM-WQX
☆19Updated 9 months ago
Alternatives and similar repositories for InternLM-WQX:
Users that are interested in InternLM-WQX are comparing it to the libraries listed below
- code for Scaling Laws of RoPE-based Extrapolation☆73Updated last year
- [ACL2024] T-Eval: Evaluating Tool Utilization Capability of Large Language Models Step by Step☆267Updated last year
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆173Updated last month
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆132Updated 10 months ago
- Repository of LV-Eval Benchmark☆63Updated 7 months ago
- [ACL2024 Findings] Agent-FLAN: Designing Data and Methods of Effective Agent Tuning for Large Language Models☆346Updated last year
- ☆226Updated 11 months ago
- [EMNLP 2024] LongAlign: A Recipe for Long Context Alignment of LLMs☆249Updated 4 months ago
- InternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencie…☆382Updated this week
- ☆168Updated last year
- An automated pipeline for evaluating LLMs for role-playing.☆168Updated 7 months ago
- SuperCLUE-Math6:新一代中文原生多轮多步数学推理数据集的探索之旅☆54Updated last year
- CLongEval: A Chinese Benchmark for Evaluating Long-Context Large Language Models☆41Updated last year
- This is a personal reimplementation of Google's Infini-transformer, utilizing a small 2b model. The project includes both model and train…☆56Updated last year
- A visuailzation tool to make deep understaning and easier debugging for RLHF training.☆187Updated 2 months ago
- LongQLoRA: Extent Context Length of LLMs Efficiently☆164Updated last year
- [ICML'24] The official implementation of “Rethinking Optimization and Architecture for Tiny Language Models”☆121Updated 3 months ago
- Codes for the paper "∞Bench: Extending Long Context Evaluation Beyond 100K Tokens": https://arxiv.org/abs/2402.13718☆321Updated 7 months ago
- ☆48Updated this week
- Data and code for paper "M3Exam: A Multilingual, Multimodal, Multilevel Benchmark for Examining Large Language Models"☆99Updated last year
- Mixture-of-Experts (MoE) Language Model☆186Updated 7 months ago
- [ACL 2024 Demo] Official GitHub repo for UltraEval: An open source framework for evaluating foundation models.☆240Updated 5 months ago
- A repository sharing the literatures about long-context large language models, including the methodologies and the evaluation benchmarks☆260Updated 8 months ago
- ☆267Updated 9 months ago
- Baichuan-Omni: Towards Capable Open-source Omni-modal LLM 🌊☆267Updated 3 months ago
- The RedStone repository includes code for preparing extensive datasets used in training large language models.☆131Updated 2 months ago
- MPB (Miner-PDF-Benchmark) is an end-to-end PDF document comprehension evaluation suite designed for large-scale model data scenarios.☆22Updated 4 months ago
- Evaluating LLMs' multi-round chatting capability via assessing conversations generated by two LLM instances.☆150Updated last year
- ☆46Updated 10 months ago
- [ACL 2024 Findings] MathBench: A Comprehensive Multi-Level Difficulty Mathematics Evaluation Dataset☆98Updated 9 months ago