valine / NeuralFlow
Visualize the intermediate output of Mistral 7B
☆341Updated last month
Alternatives and similar repositories for NeuralFlow:
Users that are interested in NeuralFlow are comparing it to the libraries listed below
- Stop messing around with finicky sampling parameters and just use DRµGS!☆345Updated 9 months ago
- a small code base for training large models☆287Updated 2 months ago
- A library for making RepE control vectors☆552Updated last month
- A complete end-to-end pipeline for LLM interpretability with sparse autoencoders (SAEs) using Llama 3.2, written in pure PyTorch and full…☆603Updated 3 months ago
- LLM Analytics☆643Updated 4 months ago
- This is our own implementation of 'Layer Selective Rank Reduction'☆233Updated 9 months ago
- Visualizing the internal board state of a GPT trained on chess PGN strings, and performing interventions on its internal board state and …☆200Updated 3 months ago
- Bayesian Optimization as a Coverage Tool for Evaluating LLMs. Accurate evaluation (benchmarking) that's 10 times faster with just a few l…☆279Updated last week
- A bagel, with everything.☆316Updated 10 months ago
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆223Updated 10 months ago
- ☆412Updated last year
- a curated list of data for reasoning ai☆130Updated 6 months ago
- Inference code for Persimmon-8B☆416Updated last year
- The repository for the code of the UltraFastBERT paper☆517Updated 11 months ago
- batched loras☆338Updated last year
- An implementation of bucketMul LLM inference☆215Updated 8 months ago
- Low-Rank adapter extraction for fine-tuned transformers models☆170Updated 10 months ago
- Steer LLM outputs towards a certain topic/subject and enhance response capabilities using activation engineering by adding steering vecto…☆224Updated 2 weeks ago
- Mistral7B playing DOOM☆129Updated 7 months ago
- Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".☆268Updated last year
- Finetune llama2-70b and codellama on MacBook Air without quantization☆448Updated 11 months ago
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆154Updated 4 months ago
- ☆506Updated 6 months ago
- Toolkit for attaching, training, saving and loading of new heads for transformer models☆264Updated last month
- Extend existing LLMs way beyond the original training length with constant memory usage, without retraining☆689Updated 10 months ago
- Absolute minimalistic implementation of a GPT-like transformer using only numpy (<650 lines).☆250Updated last year
- Alice in Wonderland code base for experiments and raw experiments data☆127Updated 2 weeks ago
- ☆152Updated 7 months ago
- ☆111Updated last month
- Fine-tune mistral-7B on 3090s, a100s, h100s☆706Updated last year