QuixiAI / extract-expertLinks
Extract a single expert from a Mixture Of Experts model using slerp interpolation.
☆19Updated last year
Alternatives and similar repositories for extract-expert
Users that are interested in extract-expert are comparing it to the libraries listed below
Sorting:
- ☆68Updated last year
- ☆165Updated 5 months ago
- entropix style sampling + GUI☆27Updated last year
- Official homepage for "Self-Harmonized Chain of Thought" (NAACL 2025)☆92Updated last year
- Low-Rank adapter extraction for fine-tuned transformers models☆180Updated last year
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆150Updated 3 weeks ago
- This is our own implementation of 'Layer Selective Rank Reduction'☆240Updated last year
- autologic is a Python package that implements the SELF-DISCOVER framework proposed in the paper SELF-DISCOVER: Large Language Models Self…☆60Updated last year
- ☆119Updated last year
- Just a bunch of benchmark logs for different LLMs☆119Updated last year
- Merge Transformers language models by use of gradient parameters.☆213Updated last year
- An unsupervised model merging algorithm for Transformers-based language models.☆108Updated last year
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆222Updated last year
- ☆74Updated 2 years ago
- ☆21Updated 2 years ago
- ☆137Updated last year
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31Updated last year
- ☆140Updated 5 months ago
- an implementation of Self-Extend, to expand the context window via grouped attention☆119Updated 2 years ago
- Let's create synthetic textbooks together :)☆76Updated 2 years ago
- Simple examples using Argilla tools to build AI☆57Updated last year
- ☆55Updated last year
- A library for benchmarking the Long Term Memory and Continual learning capabilities of LLM based agents. With all the tests and code you…☆82Updated last year
- ☆137Updated 10 months ago
- A framework for evaluating function calls made by LLMs☆40Updated last year
- EvaByte: Efficient Byte-level Language Models at Scale☆115Updated 9 months ago
- GPT-2 small trained on phi-like data☆68Updated last year
- Small, simple agent task environments for training and evaluation☆19Updated last year
- ☆45Updated last year
- ☆125Updated last year