Const-me / Cgml
GPU-targeted vendor-agnostic AI library for Windows, and Mistral model implementation.
☆56Updated last year
Alternatives and similar repositories for Cgml:
Users that are interested in Cgml are comparing it to the libraries listed below
- A library for incremental loading of large PyTorch checkpoints☆56Updated 2 years ago
- A JPEG Image Compression Service using Part Homomorphic Encryption.☆30Updated last month
- A playground to make it easy to try crazy things☆33Updated this week
- A fork of llama3.c used to do some R&D on inferencing☆20Updated 3 months ago
- Mistral7B playing DOOM☆130Updated 9 months ago
- Richard is gaining power☆184Updated 4 months ago
- A Javascript library (with Typescript types) to parse metadata of GGML based GGUF files.☆47Updated 8 months ago
- Port of Suno AI's Bark in C/C++ for fast inference☆52Updated last year
- Experiments with BitNet inference on CPU☆53Updated last year
- Web browser version of StarCoder.cpp☆44Updated last year
- Tiny Dream - An embedded, Header Only, Stable Diffusion C++ implementation☆260Updated last year
- Editor with LLM generation tree exploration☆66Updated 2 months ago
- A copy of ONNX models, datasets, and code all in one GitHub repository. Follow the README to learn more.☆105Updated last year
- Local LLM inference & management server with built-in OpenAI API☆31Updated 11 months ago
- A CLI to manage install and configure llama inference implemenation in multiple languages☆65Updated last year
- Heirarchical Navigable Small Worlds☆94Updated last week
- Tiny inference-only implementation of LLaMA☆92Updated last year
- throwaway GPT inference☆138Updated 10 months ago
- A web-app to explore topics using LLM (less typing and more clicks)☆66Updated last year
- Algebraic enhancements for GEMM & AI accelerators☆275Updated last month
- ☆34Updated 3 months ago
- A GPU Accelerated Binary Vector Store☆47Updated 2 months ago
- An implementation of bucketMul LLM inference☆216Updated 9 months ago
- A live multiplayer trivia game where users can bid for the subject of the next question☆28Updated 2 weeks ago
- Official implementation of "WhisperNER: Unified Open Named Entity and Speech Recognition"☆186Updated last month
- Agent Based Model on GPU using CUDA 12.2.1 and OpenGL 4.5 (CUDA OpenGL interop) on Windows/Linux☆69Updated last month
- ☆163Updated 10 months ago
- A super simple web interface to perform blind tests on LLM outputs.☆28Updated last year
- Docker-based inference engine for AMD GPUs☆230Updated 6 months ago
- Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization, with PyTorch/CUDA☆36Updated last year