lxe / wasm-gpt
Tensor library for machine learning
☆278Updated 2 years ago
Alternatives and similar repositories for wasm-gpt:
Users that are interested in wasm-gpt are comparing it to the libraries listed below
- OpenAI-compatible Python client that can call any LLM☆371Updated last year
- WebGPU LLM inference tuned by hand☆149Updated last year
- Augment GPT-4 Environment Access☆287Updated 2 years ago
- Instruct-tune LLaMA on consumer hardware☆362Updated 2 years ago
- JS tokenizer for LLaMA 1 and 2☆351Updated 10 months ago
- Add local LLMs to your Web or Electron apps! Powered by Rust + WebGPU☆102Updated last year
- Run GGML models with Kubernetes.☆173Updated last year
- An implementation of bucketMul LLM inference☆217Updated 10 months ago
- ☆144Updated last year
- TypeScript generator for llama.cpp Grammar directly from TypeScript interfaces☆135Updated 9 months ago
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆567Updated last year
- Enforce structured output from LLMs 100% of the time☆249Updated 9 months ago
- A tiny implementation of an autonomous agent powered by LLMs (OpenAI GPT-4)☆442Updated 2 years ago
- ☆252Updated last year
- Mistral7B playing DOOM☆131Updated 9 months ago
- Large language model evaluation and workflow framework from Phase AI.☆457Updated 3 months ago
- Implement recursion using English as the programming language and an LLM as the runtime.☆232Updated 2 years ago
- Easy-to-use headless React Hooks to run LLMs in the browser with WebGPU. Just useLLM().☆692Updated last year
- LLaMa retrieval plugin script using OpenAI's retrieval plugin☆324Updated 2 years ago
- SoTA Transformers with C-backend for fast inference on your CPU.☆310Updated last year
- A program synthesis agent that autonomously fixes its output by running tests!☆452Updated 7 months ago
- Revealing example of self-attention, the building block of transformer AI models☆131Updated 2 years ago
- Constrained Decoding for LLMs against JSON Schema☆326Updated last year
- LLM-based tool for parsing information and chatting with it☆214Updated last year
- Marsha is a functional, higher-level, English-based programming language that gets compiled into tested Python software by an LLM☆470Updated last year
- Redteaming LLMs using other LLMs☆253Updated 2 years ago
- ☆126Updated 2 years ago
- Command-line script for inferencing from models such as MPT-7B-Chat☆101Updated last year
- ☆163Updated 11 months ago
- LLaMA Cog template☆308Updated last year