seongminp / hypersegLinks
Code for HyperSeg and HyperSum
☆15Updated 3 months ago
Alternatives and similar repositories for hyperseg
Users that are interested in hyperseg are comparing it to the libraries listed below
Sorting:
- ☆119Updated last year
- First-principle implementations of groundbreaking AI algorithms using a wide range of deep learning frameworks, accompanied by supporting…☆178Updated 3 months ago
- Code for ExploreTom☆86Updated 4 months ago
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024☆344Updated 6 months ago
- Notebook and Scripts that showcase running quantized diffusion models on consumer GPUs☆38Updated last year
- Block Transformer: Global-to-Local Language Modeling for Fast Inference (NeurIPS 2024)☆162Updated 6 months ago
- RetroLLM: Empowering LLMs to Retrieve Fine-grained Evidence within Generation [ACL 2025]☆119Updated 9 months ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆202Updated last year
- A single repo with all scripts and utils to train / fine-tune the Mamba model with or without FIM☆59Updated last year
- code for training & evaluating Contextual Document Embedding models☆199Updated 5 months ago
- ☆96Updated 7 months ago
- ☆146Updated last year
- Banishing LLM Hallucinations Requires Rethinking Generalization☆275Updated last year
- Survey: A collection of AWESOME papers and resources on the latest research in Mixture of Experts.☆137Updated last year
- Repo hosting codes and materials related to speeding LLMs' inference using token merging.☆37Updated 3 weeks ago
- MatFormer repo☆64Updated 10 months ago
- The code repository of the paper: Competition and Attraction Improve Model Fusion☆163Updated 2 months ago
- An open source implementation of LFMs from Liquid AI: Liquid Foundation Models☆113Updated last year
- Getting started with TensorRT-LLM using BLOOM as a case study☆23Updated last year
- Prune transformer layers☆69Updated last year
- An unofficial pytorch implementation of 'Efficient Infinite Context Transformers with Infini-attention'☆53Updated last year
- Landing repository for the paper "Softpick: No Attention Sink, No Massive Activations with Rectified Softmax"☆85Updated last month
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆154Updated last year
- Enhancing Translation with RAG-Powered Large Language Models☆84Updated last month
- Fine-tuning large language models (LLMs) is crucial for enhancing performance across domain-specific task applications. This comprehensiv…☆12Updated last year
- Mixture-of-Recursions: Learning Dynamic Recursive Depths for Adaptive Token-Level Computation (NeurIPS 2025)☆493Updated last month
- ☆136Updated last year
- Source code of the paper: RetrievalQA: Assessing Adaptive Retrieval-Augmented Generation for Short-form Open-Domain Question Answering [F…☆68Updated last year
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆231Updated last year
- ☆266Updated 4 months ago