hongzhouyu / FineMedLinks
The codebase and some introductions of FineMed.
☆24Updated last week
Alternatives and similar repositories for FineMed
Users that are interested in FineMed are comparing it to the libraries listed below
Sorting:
- ☆118Updated 5 months ago
- m1: Unleash the Potential of Test-Time Scaling for Medical Reasoning in Large Language Models☆40Updated 3 months ago
- [NeurIPS 2024] A Novel Rank-Based Metric for Evaluating Large Language Models☆51Updated 2 months ago
- Reproduction of the complete process of DeepSeek-R1 on small-scale models, including Pre-training, SFT, and RL.☆27Updated 5 months ago
- ☆18Updated last week
- X-Reasoner: Towards Generalizable Reasoning Across Modalities and Domains☆47Updated 3 months ago
- An Efficient LLM Fine-Tuning Factory Optimized for MoE PEFT☆108Updated 5 months ago
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆82Updated 5 months ago
- [ACL'25] We propose a novel fine-tuning method, Separate Memory and Reasoning, which combines prompt tuning with LoRA.☆70Updated 3 weeks ago
- [NeurIPS'24] Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Models☆62Updated 8 months ago
- MedGen: Unlocking Medical Video Generation by Scaling Granularly-annotated Medical Videos.☆21Updated last month
- A curated list of awesome LLM Inference-Time Self-Improvement (ITSI, pronounced "itsy") papers from our recent survey: A Survey on Large …☆88Updated 7 months ago
- Official Implementation for EMNLP 2024 (main) "AgentReview: Exploring Academic Peer Review with LLM Agent."☆83Updated 9 months ago
- ☆53Updated 6 months ago
- Official code implementation for the ACL 2025 paper: 'CoT-based Synthesizer: Enhancing LLM Performance through Answer Synthesis'☆26Updated 2 months ago
- [ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".☆126Updated 9 months ago
- [ICLR 2025] SuperCorrect: Advancing Small LLM Reasoning with Thought Template Distillation and Self-Correction☆76Updated 4 months ago
- ☆67Updated last month
- [arxiv: 2505.02156] Adaptive Thinking via Mode Policy Optimization for Social Language Agents☆39Updated last month
- A light-weight tool for evaluating LLMs in rule-based ways.☆68Updated last month
- Code for "CREAM: Consistency Regularized Self-Rewarding Language Models", ICLR 2025.☆25Updated 5 months ago
- State-of-the-art Parameter-Efficient MoE Fine-tuning Method☆176Updated 11 months ago
- Analyzing and Reducing Catastrophic Forgetting in Parameter Efficient Tuning☆34Updated 8 months ago
- Parameter-Efficient Fine-Tuning for Foundation Models☆81Updated 4 months ago
- ☆112Updated last year
- ☆141Updated 2 months ago
- [ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models☆124Updated last month
- [SIGIR'24] The official implementation code of MOELoRA.☆175Updated last year
- 🚀 LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training☆86Updated 8 months ago
- [Preprint] On the Generalization of SFT: A Reinforcement Learning Perspective with Reward Rectification.☆59Updated this week