magicproduct / hash-hop
Long context evaluation for large language models
☆198Updated this week
Alternatives and similar repositories for hash-hop:
Users that are interested in hash-hop are comparing it to the libraries listed below
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆158Updated 2 weeks ago
- Experiments on speculative sampling with Llama models☆123Updated last year
- A simple unified framework for evaluating LLMs☆172Updated this week
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆113Updated last month
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆288Updated last month
- Code for Paper: Training Software Engineering Agents and Verifiers with SWE-Gym☆230Updated 2 weeks ago
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆156Updated 3 months ago
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆118Updated 3 months ago
- Repo for "LoLCATs: On Low-Rank Linearizing of Large Language Models"☆204Updated last month
- PyTorch implementation of models from the Zamba2 series.☆173Updated this week
- Multipack distributed sampler for fast padding-free training of LLMs☆184Updated 5 months ago
- Code for NeurIPS'24 paper 'Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization'☆176Updated last month
- smolLM with Entropix sampler on pytorch☆148Updated 2 months ago
- A comprehensive repository of reasoning tasks for LLMs (and beyond)☆394Updated 4 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆91Updated 2 months ago
- an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated last year
- Manage scalable open LLM inference endpoints in Slurm clusters☆249Updated 6 months ago
- A puzzle to learn about prompting☆123Updated last year
- A framework to study AI models in Reasoning, Alignment, and use of Memory (RAM).☆157Updated 2 weeks ago
- ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward exp…☆215Updated 9 months ago
- MiniHF is an inference, human preference data collection, and fine-tuning tool for local language models. It is intended to help the user…☆163Updated this week
- Public repository for "The Surprising Effectiveness of Test-Time Training for Abstract Reasoning"☆285Updated 2 months ago
- Extract full next-token probabilities via language model APIs☆229Updated 11 months ago
- Normalized Transformer (nGPT)☆146Updated 2 months ago
- Code for exploring Based models from "Simple linear attention language models balance the recall-throughput tradeoff"☆219Updated last month
- BABILong is a benchmark for LLM evaluation using the needle-in-a-haystack approach.☆184Updated 2 months ago
- DeMo: Decoupled Momentum Optimization☆171Updated last month
- ☆142Updated last week