AMD-AGI / AMD-LLMLinks
☆189Updated last year
Alternatives and similar repositories for AMD-LLM
Users that are interested in AMD-LLM are comparing it to the libraries listed below
Sorting:
- ☆196Updated 4 months ago
- Docker-based inference engine for AMD GPUs☆230Updated 11 months ago
- Code sample showing how to run and benchmark models on Qualcomm's Window PCs☆101Updated 11 months ago
- Run and explore Llama models locally with minimal dependencies on CPU☆189Updated 11 months ago
- Absolute minimalistic implementation of a GPT-like transformer using only numpy (<650 lines).☆253Updated last year
- Algebraic enhancements for GEMM & AI accelerators☆280Updated 7 months ago
- Richard is gaining power☆194Updated 3 months ago
- Bayesian Optimization as a Coverage Tool for Evaluating LLMs. Accurate evaluation (benchmarking) that's 10 times faster with just a few l…☆285Updated 2 weeks ago
- ☆248Updated last year
- An implementation of bucketMul LLM inference☆223Updated last year
- Neurox control helm chart details☆30Updated 5 months ago
- This repo contains a new way to use bloom filters to do lossless video compression☆251Updated 3 months ago
- Run larger LLMs with longer contexts on Apple Silicon by using differentiated precision for KV cache quantization. KVSplit enables 8-bit …☆360Updated 4 months ago
- Felafax is building AI infra for non-NVIDIA GPUs☆567Updated 8 months ago
- ☆163Updated last year
- Dead Simple LLM Abliteration☆232Updated 7 months ago
- ☆125Updated 4 months ago
- A copy of ONNX models, datasets, and code all in one GitHub repository. Follow the README to learn more.☆104Updated last year
- Pytorch script hot swap: Change code without unloading your LLM from VRAM☆125Updated 5 months ago
- OpenTSLM: Time-Series Language Models for Reasoning over Multivariate Medical Text- and Time-Series Data☆486Updated this week
- ☆231Updated 6 months ago
- Online compiler for HIP and NVIDIA® CUDA® code to WebGPU☆195Updated 8 months ago
- Visualize the intermediate output of Mistral 7B☆371Updated 8 months ago
- A CLI to manage install and configure llama inference implemenation in multiple languages☆67Updated last year
- A GPU Accelerated Binary Vector Store☆47Updated 7 months ago
- Grow virtual creatures in static and physics simulated environments.☆53Updated last year
- minimal yet working VPN daemon for Linux☆106Updated last month
- High-Performance Implementation of OpenAI's TikToken.☆455Updated 3 months ago
- Live-bending a foundation model’s output at neural network level.☆265Updated 5 months ago
- This project collects GPU benchmarks from various cloud providers and compares them to fixed per token costs. Use our tool for efficient …☆219Updated 9 months ago