adrienbrault / json-schema-to-gbnfLinks
Converts JSON-Schema to GBNF grammar to use with llama.cpp
☆55Updated last year
Alternatives and similar repositories for json-schema-to-gbnf
Users that are interested in json-schema-to-gbnf are comparing it to the libraries listed below
Sorting:
- TypeScript generator for llama.cpp Grammar directly from TypeScript interfaces☆140Updated last year
- Generates grammer files from typescript for LLM generation☆38Updated last year
- Embedding models from Jina AI☆65Updated last year
- GPU accelerated client-side embeddings for vector search, RAG etc.☆65Updated last year
- Plug n Play GBNF Compiler for llama.cpp☆27Updated last year
- Extend the original llama.cpp repo to support redpajama model.☆118Updated last year
- A library for working with GBNF files☆25Updated last month
- utilities for loading and running text embeddings with onnx☆44Updated 2 months ago
- ☆31Updated 2 years ago
- Using modal.com to process FineWeb-edu data☆20Updated 6 months ago
- Use context-free grammars with an LLM☆175Updated last year
- Chat Markup Language conversation library☆55Updated last year
- A clone of OpenAI's Tokenizer page for HuggingFace Models☆45Updated last year
- A library for incremental loading of large PyTorch checkpoints☆56Updated 2 years ago
- iterate quickly with llama.cpp hot reloading. use the llama.cpp bindings with bun.sh☆50Updated last year
- Replace expensive LLM calls with finetunes automatically☆64Updated last year
- ☆112Updated last year
- Unofficial python bindings for the rust llm library. 🐍❤️🦀☆76Updated 2 years ago
- Command-line script for inferencing from models such as falcon-7b-instruct☆74Updated 2 years ago
- llama.cpp gguf file parser for javascript☆49Updated 10 months ago
- ☆88Updated this week
- JS tokenizer for LLaMA 1 and 2☆360Updated last year
- WebGPU LLM inference tuned by hand☆150Updated 2 years ago
- LLM plugin for clustering embeddings☆82Updated last year
- ☆35Updated 2 years ago
- Run GGML models with Kubernetes.☆173Updated last year
- ⚡️ A fast and flexible PyTorch inference server that runs locally, on any cloud or AI HW.☆145Updated last year
- Using Large Language Models for Repo-wide Type Prediction☆112Updated last year
- A web-app to explore topics using LLM (less typing and more clicks)☆67Updated last year
- Add local LLMs to your Web or Electron apps! Powered by Rust + WebGPU☆104Updated 2 years ago