Const-me / Cgml
GPU-targeted vendor-agnostic AI library for Windows, and Mistral model implementation.
☆46Updated 9 months ago
Related projects ⓘ
Alternatives and complementary repositories for Cgml
- A library for incremental loading of large PyTorch checkpoints☆56Updated last year
- A live multiplayer trivia game where users can bid for the subject of the next question☆22Updated last week
- Web browser version of StarCoder.cpp☆43Updated last year
- Port of Suno AI's Bark in C/C++ for fast inference☆54Updated 6 months ago
- A copy of ONNX models, datasets, and code all in one GitHub repository. Follow the README to learn more.☆106Updated 11 months ago
- Tiny Dream - An embedded, Header Only, Stable Diffusion C++ implementation☆252Updated last year
- Image Generation API Server - Similar to https://text-generator.io but for images☆46Updated 2 months ago
- C++ raytracer that supports custom models. Supports running the calculations on the CPU using C++11 threads or in the GPU via CUDA.☆74Updated last year
- GGML implementation of BERT model with Python bindings and quantization.☆51Updated 8 months ago
- Mistral7B playing DOOM☆122Updated 3 months ago
- WebGPU LLM inference tuned by hand☆146Updated last year
- An implementation of bucketMul LLM inference☆214Updated 4 months ago
- Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization, with PyTorch/CUDA☆35Updated 8 months ago
- LLaVA server (llama.cpp).☆177Updated last year
- Richard is gaining power☆174Updated 2 months ago
- Local LLM inference & management server with built-in OpenAI API☆31Updated 6 months ago
- Wang Yi's GPT solution☆142Updated 10 months ago
- Experiments with BitNet inference on CPU☆50Updated 7 months ago
- The procedure and the code to run shap-e sample code locally.☆116Updated last year
- 100k real ( +100k random ) galaxies from a sector. Visualized with Raylib.☆87Updated last month
- ☆39Updated 8 months ago
- Command-line script for inferencing from models such as falcon-7b-instruct☆75Updated last year
- llama.cpp fork with additional SOTA quants and improved performance☆89Updated this week
- ☆137Updated last month
- The creative suite for character-driven AI experiences.☆181Updated 2 months ago
- Port of Facebook's LLaMA model in C/C++☆20Updated last year
- General purpose GPU compute framework built on Vulkan to support 1000s of cross vendor graphics cards (AMD, Qualcomm, NVIDIA & friends). …☆41Updated last month
- DiscoGrad - automatically differentiate across conditional branches in C++ programs☆204Updated 2 months ago
- Generate ideal question-answers for testing RAG☆123Updated 4 months ago
- A CLI to manage install and configure llama inference implemenation in multiple languages☆66Updated 10 months ago