huggingface / latex2sympy2_extended
Parse LaTeX math expressions
☆21Updated 3 weeks ago
Alternatives and similar repositories for latex2sympy2_extended:
Users that are interested in latex2sympy2_extended are comparing it to the libraries listed below
- ☆69Updated 2 months ago
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models" [AISTATS …☆59Updated 6 months ago
- ☆77Updated 3 months ago
- [ACL 2024] RelayAttention for Efficient Large Language Model Serving with Long System Prompts☆39Updated last year
- qwen-nsa☆57Updated 2 weeks ago
- Efficient triton implementation of Native Sparse Attention.☆139Updated 2 weeks ago
- Code for paper: [ICLR2025 Oral] FlexPrefill: A Context-Aware Sparse Attention Mechanism for Efficient Long-Sequence Inference☆94Updated last week
- Here we will test various linear attention designs.☆60Updated last year
- Pruner-Zero: Evolving Symbolic Pruning Metric from scratch for LLMs☆81Updated 5 months ago
- Low-bit optimizers for PyTorch☆128Updated last year
- ☆43Updated 2 months ago
- ☆30Updated 11 months ago
- Awesome Triton Resources☆26Updated 3 weeks ago
- ☆76Updated last week
- Repo for "Z1: Efficient Test-time Scaling with Code"☆55Updated 2 weeks ago
- Odysseus: Playground of LLM Sequence Parallelism☆68Updated 10 months ago
- ZO2 (Zeroth-Order Offloading): Full Parameter Fine-Tuning 175B LLMs with 18GB GPU Memory☆91Updated 3 weeks ago
- FastCuRL: Curriculum Reinforcement Learning with Progressive Context Extension for Efficient Training R1-like Reasoning Models☆43Updated last week
- XAttention: Block Sparse Attention with Antidiagonal Scoring☆140Updated 3 weeks ago
- [ICLR 2025] Official PyTorch implementation of "Forgetting Transformer: Softmax Attention with a Forget Gate"☆95Updated 2 weeks ago
- Efficient Mixture of Experts for LLM Paper List☆61Updated 4 months ago
- ☆34Updated 9 months ago
- ☆17Updated 3 months ago
- Transformers components but in Triton☆32Updated last month
- 🔥 A minimal training framework for scaling FLA models☆111Updated this week
- An unofficial implementation of "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"☆35Updated 10 months ago
- Rethinking RL Scaling for Vision Language Models: A Transparent, From-Scratch Framework and Comprehensive Evaluation Scheme☆115Updated 2 weeks ago
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆173Updated last month
- A repository for research on medium sized language models.☆76Updated 11 months ago
- [ICML'24 Oral] The official code of "DiJiang: Efficient Large Language Models through Compact Kernelization", a novel DCT-based linear at…☆100Updated 10 months ago