NJUDeepEngine / CAEF
Code for paper: "Executing Arithmetic: Fine-Tuning Large Language Models as Turing Machines"
☆11Updated 4 months ago
Alternatives and similar repositories for CAEF:
Users that are interested in CAEF are comparing it to the libraries listed below
- [NAACL 2025] Source code for MMEvalPro, a more trustworthy and efficient benchmark for evaluating LMMs☆23Updated 5 months ago
- [NAACL 2025] Representing Rule-based Chatbots with Transformers☆19Updated 3 weeks ago
- The code and data for the paper JiuZhang3.0☆40Updated 9 months ago
- PreAct: Prediction Enhances Agent's Planning Ability (Coling2025)☆26Updated 2 months ago
- Official repository for Decentralized Arena via Collective LLM Intelligence☆8Updated 4 months ago
- ☆23Updated 8 months ago
- ☆33Updated this week
- On The Planning Abilities of OpenAI's o1 Models: Feasibility, Optimality, and Generalizability☆37Updated last month
- In-Context Alignment: Chat with Vanilla Language Models Before Fine-Tuning☆33Updated last year
- SLED: Self Logits Evolution Decoding for Improving Factuality in Large Language Model https://arxiv.org/pdf/2411.02433☆22Updated 2 months ago
- ☆13Updated 2 months ago
- Codebase for Instruction Following without Instruction Tuning☆33Updated 5 months ago
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆46Updated last year
- ☆15Updated 7 months ago
- ☆20Updated 4 months ago
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆44Updated 2 months ago
- Fast LLM Training CodeBase With dynamic strategy choosing [Deepspeed+Megatron+FlashAttention+CudaFusionKernel+Compiler];☆36Updated last year
- ☆64Updated 10 months ago
- SELF-GUIDE: Better Task-Specific Instruction Following via Self-Synthetic Finetuning. COLM 2024 Accepted Paper☆28Updated 9 months ago
- Code for preprint "Metadata Conditioning Accelerates Language Model Pre-training (MeCo)"☆36Updated last month
- ☆35Updated 5 months ago
- [NAACL 2025] A Closer Look into Mixture-of-Experts in Large Language Models☆44Updated 3 weeks ago
- A Framework for Decoupling and Assessing the Capabilities of VLMs☆40Updated 8 months ago
- ☆20Updated 7 months ago