Linaro / tinyBLAS
A fork of OpenBLAS with Armv8-A SVE (Scalable Vector Extension) support
☆17Updated 5 years ago
Alternatives and similar repositories for tinyBLAS
Users that are interested in tinyBLAS are comparing it to the libraries listed below
Sorting:
- 33B Chinese LLM, DPO QLORA, 100K context, AirLLM 70B inference with single 4GB GPU☆13Updated last year
- Editor with LLM generation tree exploration☆66Updated 3 months ago
- Lightweight Llama 3 8B Inference Engine in CUDA C☆47Updated last month
- ☆54Updated 8 months ago
- A faithful clone of Karpathy's llama2.c (one file inference, zero dependency) but fully functional with LLaMA 3 8B base and instruct mode…☆127Updated 9 months ago
- A library for incremental loading of large PyTorch checkpoints☆56Updated 2 years ago
- The Finite Field Assembly Programming Language☆36Updated last month
- ☆19Updated last month
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31Updated 11 months ago
- General purpose GPU compute framework built on Vulkan to support 1000s of cross vendor graphics cards (AMD, Qualcomm, NVIDIA & friends). …☆46Updated 2 months ago
- Neurox control helm chart details☆31Updated 2 weeks ago
- Inference RWKV v7 in pure C.☆33Updated last month
- GPU-targeted vendor-agnostic AI library for Windows, and Mistral model implementation.☆57Updated last year
- Simple LLM inference server☆20Updated 11 months ago
- ☆58Updated 10 months ago
- ☆18Updated 10 months ago
- Prepare for DeekSeek R1 inference: Benchmark CPU, DRAM, SSD, iGPU, GPU, ... with efficient code.☆72Updated 3 months ago
- Lego for GRPO☆28Updated last month
- Aana SDK is a powerful framework for building AI enabled multimodal applications.☆47Updated this week
- GRDN.AI app for garden optimization☆70Updated last year
- a simplified version of Google's Gemma model to be used for learning☆24Updated last year
- new optimizer☆20Updated 9 months ago
- Transformer GPU VRAM estimator☆61Updated last year
- Because it's there.☆16Updated 7 months ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆61Updated this week
- Blindspots in LLMs I've noticed while AI coding. Sonnet family emphasis.☆14Updated last month
- The DPAB-α Benchmark☆21Updated 4 months ago
- ☆66Updated 11 months ago
- look how they massacred my boy☆63Updated 7 months ago
- llm plugin for Cerebras fast inference API☆25Updated 2 months ago