AMD-AIG-AIMA / AMD-LLMLinks
☆188Updated 11 months ago
Alternatives and similar repositories for AMD-LLM
Users that are interested in AMD-LLM are comparing it to the libraries listed below
Sorting:
- Docker-based inference engine for AMD GPUs☆231Updated 9 months ago
- ☆196Updated 3 months ago
- Code sample showing how to run and benchmark models on Qualcomm's Window PCs☆100Updated 10 months ago
- Run and explore Llama models locally with minimal dependencies on CPU☆191Updated 9 months ago
- Algebraic enhancements for GEMM & AI accelerators☆278Updated 5 months ago
- Richard is gaining power☆196Updated last month
- This repo contains a new way to use bloom filters to do lossless video compression☆247Updated 2 months ago
- Run larger LLMs with longer contexts on Apple Silicon by using differentiated precision for KV cache quantization. KVSplit enables 8-bit …☆356Updated 2 months ago
- A GPU Accelerated Binary Vector Store☆47Updated 5 months ago
- ☆249Updated last year
- Online compiler for HIP and NVIDIA® CUDA® code to WebGPU☆190Updated 6 months ago
- A copy of ONNX models, datasets, and code all in one GitHub repository. Follow the README to learn more.☆105Updated last year
- Absolute minimalistic implementation of a GPT-like transformer using only numpy (<650 lines).☆253Updated last year
- throwaway GPT inference☆140Updated last year
- ☆124Updated 2 months ago
- An implementation of bucketMul LLM inference☆221Updated last year
- GPU-targeted vendor-agnostic AI library for Windows, and Mistral model implementation.☆58Updated last year
- Mistral7B playing DOOM☆133Updated last year
- ☆1,045Updated 2 months ago
- Neurox control helm chart details☆30Updated 3 months ago
- Pytorch script hot swap: Change code without unloading your LLM from VRAM☆126Updated 3 months ago
- ☆164Updated 4 months ago
- Lamport's Bakery Algorithm Demonstrated in Python☆96Updated last year
- A CLI to manage install and configure llama inference implemenation in multiple languages☆67Updated last year
- High-Performance Implementation of OpenAI's TikToken.☆445Updated last month
- Dead Simple LLM Abliteration☆230Updated 5 months ago
- ☆163Updated last year
- Tensor library & inference framework for machine learning☆106Updated 3 weeks ago
- WIP: Next-gen network protocol for reliable data transfer in lossy environments. Outperforms TCP/UDP in high packet loss scenarios.☆310Updated 10 months ago
- Sequential Logic☆111Updated this week