lxe / wasm-gpt
Tensor library for machine learning
☆278Updated last year
Alternatives and similar repositories for wasm-gpt:
Users that are interested in wasm-gpt are comparing it to the libraries listed below
- WebGPU LLM inference tuned by hand☆149Updated last year
- Add local LLMs to your Web or Electron apps! Powered by Rust + WebGPU☆102Updated last year
- JS tokenizer for LLaMA 1 and 2☆351Updated 9 months ago
- Web-optimized vector database (written in Rust).☆224Updated last month
- OpenAI-compatible Python client that can call any LLM☆371Updated last year
- TypeScript generator for llama.cpp Grammar directly from TypeScript interfaces☆135Updated 9 months ago
- An implementation of bucketMul LLM inference☆216Updated 9 months ago
- Run GGML models with Kubernetes.☆174Updated last year
- ☆143Updated last year
- Command-line script for inferencing from models such as MPT-7B-Chat☆101Updated last year
- Extend the original llama.cpp repo to support redpajama model.☆117Updated 7 months ago
- Revealing example of self-attention, the building block of transformer AI models☆131Updated last year
- Enforce structured output from LLMs 100% of the time☆249Updated 8 months ago
- Tensor computation with WebGPU acceleration☆614Updated 8 months ago
- Augment GPT-4 Environment Access☆287Updated 2 years ago
- Easy-to-use headless React Hooks to run LLMs in the browser with WebGPU. Just useLLM().☆690Updated last year
- An AI-driven tool to analyze your profile and gain insights into how ChatGPT interprets your personality.☆180Updated last year
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆565Updated last year
- 🦜️🔗 This is a very simple re-implementation of LangChain, in ~100 lines of code☆253Updated last year
- Vercel and web-llm template to run wasm models directly in the browser.☆146Updated last year
- Browser-compatible JS library for running language models☆228Updated 2 years ago
- Instruct-tune LLaMA on consumer hardware☆362Updated 2 years ago
- Run inference on replit-3B code instruct model using CPU☆154Updated last year
- Layered, depth-first reading—start with summaries, tap to explore details, and gain clarity on complex topics.☆271Updated last year
- LLM-based code completion engine☆182Updated 2 months ago
- C++ implementation for 💫StarCoder☆453Updated last year
- C++ implementation for BLOOM☆809Updated last year
- LLaMa retrieval plugin script using OpenAI's retrieval plugin☆324Updated 2 years ago
- Implement recursion using English as the programming language and an LLM as the runtime.☆137Updated 2 years ago
- ☆155Updated 2 years ago