rbiswasfc / llm-science-exam
6th Position Solution Code for Kaggle - LLM Science Exam Competition
☆21Updated 8 months ago
Alternatives and similar repositories for llm-science-exam:
Users that are interested in llm-science-exam are comparing it to the libraries listed below
- ☆11Updated 7 months ago
- ☆65Updated 7 months ago
- 3rd Place solution for Feedback Prize - Predicting Effective Arguments Kaggle competition☆15Updated 2 years ago
- ☆166Updated 8 months ago
- kaggle 2024 Eedi 第10名 金牌方案☆31Updated 3 months ago
- Solution of Kaggle competition: LMSYS - Chatbot Arena Human Preference Predictions☆42Updated 7 months ago
- 9th solution☆11Updated 2 years ago
- Llama-3-SynE: A Significantly Enhanced Version of Llama-3 with Advanced Scientific Reasoning and Chinese Language Capabilities | 继续预训练提升 …☆33Updated 3 months ago
- ☆28Updated last month
- KDD 2024 AQA competition 2nd place solution☆11Updated 8 months ago
- Fantastic Data Engineering for Large Language Models☆84Updated 2 months ago
- Clustering and Ranking: Diversity-preserved Instruction Selection through Expert-aligned Quality Estimation☆76Updated 4 months ago
- ☆46Updated 3 months ago
- [ACL'24] Superfiltering: Weak-to-Strong Data Filtering for Fast Instruction-Tuning☆147Updated 6 months ago
- ☆40Updated 2 years ago
- ☆66Updated last year
- The code and data of DPA-RAG☆58Updated 2 months ago
- Official github repo for AutoDetect, an automated weakness detection framework for LLMs.☆42Updated 9 months ago
- 天池算法比赛《BetterMixture - 大模型数据混合挑战赛》的第一名top1解决方案☆27Updated 8 months ago
- Text to text with reinforcement learning☆30Updated last year
- 1st Place Solution for Eedi - Mining Misconceptions in Mathematics Kaggle Competition☆32Updated 3 months ago
- ☆33Updated last month
- 第四届厦门国际银行数创金融杯建模大赛冠军方案(含OCR预训练权重)☆13Updated 9 months ago
- ☆80Updated last year
- 使用单个24G显卡,从0开始训练LLM☆50Updated 5 months ago
- ☆12Updated 7 months ago
- 🐋 An unofficial implementation of Self-Alignment with Instruction Backtranslation.☆137Updated 9 months ago
- ☆13Updated 2 years ago
- This is a detailed code demo on how to conduct Full-Param Supervised Fine-tuning (SFT) and DPO (Direct Preference Optimization)☆12Updated 2 months ago
- ☆40Updated 7 months ago