PABannier / biogpt.cppLinks
Port of Microsoft's BioGPT in C/C++ using ggml
☆85Updated last year
Alternatives and similar repositories for biogpt.cpp
Users that are interested in biogpt.cpp are comparing it to the libraries listed below
Sorting:
- GGML implementation of BERT model with Python bindings and quantization.☆58Updated last year
- LLaVA server (llama.cpp).☆183Updated 2 years ago
- Python bindings for ggml☆147Updated last year
- Inference of Large Multimodal Models in C/C++. LLaVA and others☆48Updated 2 years ago
- Web browser version of StarCoder.cpp☆46Updated 2 years ago
- Embeddings focused small version of Llama NLP model☆107Updated 2 years ago
- tinygrad port of the RWKV large language model.☆45Updated 11 months ago
- Python bindings for llama.cpp☆198Updated 2 years ago
- Port of Meta's Encodec in C/C++☆227Updated last year
- SoTA Transformers with C-backend for fast inference on your CPU.☆311Updated 2 years ago
- Inference of Mamba and Mamba2 models in pure C☆196Updated 2 weeks ago
- LLM-based code completion engine☆190Updated last year
- Port of Suno AI's Bark in C/C++ for fast inference☆54Updated last year
- Demo python script app to interact with llama.cpp server using whisper API, microphone and webcam devices.☆46Updated 2 years ago
- ggml implementation of BERT☆498Updated last year
- ggml implementation of embedding models including SentenceTransformer and BGE☆63Updated 2 years ago
- ☆26Updated 2 years ago
- Stable Diffusion in pure C/C++☆62Updated 2 years ago
- A ggml (C++) re-implementation of tortoise-tts☆193Updated last year
- WebGPU LLM inference tuned by hand☆151Updated 2 years ago
- Experiments with BitNet inference on CPU☆55Updated last year
- Command-line script for inferencing from models such as MPT-7B-Chat☆100Updated 2 years ago
- Local ML voice chat using high-end models.☆182Updated last month
- Unofficial python bindings for the rust llm library. 🐍❤️🦀☆76Updated 2 years ago
- A torchless, c++ rwkv implementation using 8bit quantization, written in cuda/hip/vulkan for maximum compatibility and minimum dependenci…☆313Updated 2 years ago
- ☆40Updated 2 years ago
- iterate quickly with llama.cpp hot reloading. use the llama.cpp bindings with bun.sh☆50Updated 2 years ago
- Trying to deconstruct RWKV in understandable terms☆14Updated 2 years ago
- Framework agnostic python runtime for RWKV models☆147Updated 2 years ago
- Fast inference of Instruct tuned LLaMa on your personal devices.☆23Updated 2 years ago