cztomsik / ggml-jsLinks
JavaScript bindings for the ggml-js library
☆44Updated 5 months ago
Alternatives and similar repositories for ggml-js
Users that are interested in ggml-js are comparing it to the libraries listed below
Sorting:
- A Javascript library (with Typescript types) to parse metadata of GGML based GGUF files.☆51Updated last year
- TypeScript generator for llama.cpp Grammar directly from TypeScript interfaces☆139Updated last year
- Add local LLMs to your Web or Electron apps! Powered by Rust + WebGPU☆103Updated 2 years ago
- WebGPU LLM inference tuned by hand☆151Updated 2 years ago
- BlinkDL's RWKV-v4 running in the browser☆47Updated 2 years ago
- A JavaScript and TypeScript port of PyTorch C++ library (libtorch) - Node.js N-API bindings for libtorch.☆16Updated 2 years ago
- ☆40Updated 2 years ago
- Inference Llama 2 in one file of pure JavaScript(HTML)☆33Updated 3 months ago
- iterate quickly with llama.cpp hot reloading. use the llama.cpp bindings with bun.sh☆51Updated last year
- Generates grammer files from typescript for LLM generation☆38Updated last year
- tinygrad port of the RWKV large language model.☆45Updated 6 months ago
- WebAssembly (Wasm) Build and Bindings for llama.cpp☆280Updated last year
- Embeddings focused small version of Llama NLP model☆104Updated 2 years ago
- GPU accelerated client-side embeddings for vector search, RAG etc.☆65Updated last year
- Unofficial python bindings for the rust llm library. 🐍❤️🦀☆75Updated 2 years ago
- JS tokenizer for LLaMA 3 and LLaMA 3.1☆115Updated last month
- Inference of Mamba models in pure C☆191Updated last year
- LLM-based code completion engine☆193Updated 7 months ago
- Browser-compatible JS library for running language models☆231Updated 3 years ago
- A ggml (C++) re-implementation of tortoise-tts☆188Updated last year
- SoTA Transformers with C-backend for fast inference on your CPU.☆309Updated last year
- Python bindings for ggml☆146Updated last year
- TTS support with GGML☆176Updated 3 weeks ago
- Llama2 inference in one TypeScript file☆18Updated 3 months ago
- Full finetuning of large language models without large memory requirements☆94Updated last year
- Implementation of the RWKV language model in pure WebGPU/Rust.☆315Updated 3 weeks ago
- Run Large-Language Models (LLMs) 🚀 directly in your browser!☆215Updated last year
- Vercel and web-llm template to run wasm models directly in the browser.☆160Updated last year
- trying to make WebGPU a bit easier to use☆17Updated last year
- Run ONNX RWKV-v4 models with GPU acceleration using DirectML [Windows], or just on CPU [Windows AND Linux]; Limited to 430M model at this…☆21Updated 2 years ago