IntrinsicLabsAI / gbnfgenLinks
TypeScript generator for llama.cpp Grammar directly from TypeScript interfaces
☆141Updated last year
Alternatives and similar repositories for gbnfgen
Users that are interested in gbnfgen are comparing it to the libraries listed below
Sorting:
- Generates grammer files from typescript for LLM generation☆38Updated last year
- JS tokenizer for LLaMA 1 and 2☆362Updated last year
- Converts JSON-Schema to GBNF grammar to use with llama.cpp☆55Updated 2 years ago
- WebGPU LLM inference tuned by hand☆151Updated 2 years ago
- A Javascript library (with Typescript types) to parse metadata of GGML based GGUF files.☆51Updated last year
- GPU accelerated client-side embeddings for vector search, RAG etc.☆65Updated 2 years ago
- Enforce structured output from LLMs 100% of the time☆249Updated last year
- Extend the original llama.cpp repo to support redpajama model.☆118Updated last year
- LLaMa retrieval plugin script using OpenAI's retrieval plugin☆323Updated 2 years ago
- Constrained Decoding for LLMs against JSON Schema☆327Updated 2 years ago
- Add local LLMs to your Web or Electron apps! Powered by Rust + WebGPU☆107Updated 2 years ago
- SemanticFinder - frontend-only live semantic search with transformers.js☆314Updated 8 months ago
- ☆114Updated last year
- An HTTP serving framework by Banana☆101Updated 2 years ago
- ☆135Updated 2 years ago
- ☆164Updated 4 months ago
- ☆32Updated last year
- LLM-based code completion engine☆190Updated 11 months ago
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆222Updated last year
- Full finetuning of large language models without large memory requirements☆94Updated 3 months ago
- Plug n Play GBNF Compiler for llama.cpp☆28Updated 2 years ago
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆124Updated 2 years ago
- A guidance compatibility layer for llama-cpp-python☆36Updated 2 years ago
- iterate quickly with llama.cpp hot reloading. use the llama.cpp bindings with bun.sh☆50Updated 2 years ago
- an implementation of Self-Extend, to expand the context window via grouped attention☆119Updated last year
- Command-line script for inferencing from models such as MPT-7B-Chat☆100Updated 2 years ago
- Run GGML models with Kubernetes.☆175Updated 2 years ago
- WebAssembly (Wasm) Build and Bindings for llama.cpp☆285Updated last year
- An implementation of bucketMul LLM inference☆222Updated last year
- Unofficial python bindings for the rust llm library. 🐍❤️🦀☆76Updated 2 years ago