belladoreai / llama-tokenizer-js
JS tokenizer for LLaMA 1 and 2
☆351Updated 10 months ago
Alternatives and similar repositories for llama-tokenizer-js
Users that are interested in llama-tokenizer-js are comparing it to the libraries listed below
Sorting:
- LLaMa retrieval plugin script using OpenAI's retrieval plugin☆324Updated 2 years ago
- OpenAI-compatible Python client that can call any LLM☆372Updated last year
- TypeScript generator for llama.cpp Grammar directly from TypeScript interfaces☆136Updated 10 months ago
- 🦜️🔗 This is a very simple re-implementation of LangChain, in ~100 lines of code☆253Updated last year
- Finetune llama2-70b and codellama on MacBook Air without quantization☆448Updated last year
- Enforce structured output from LLMs 100% of the time☆249Updated 9 months ago
- Extend the original llama.cpp repo to support redpajama model.☆117Updated 8 months ago
- Complex LLM Workflows from Simple JSON.☆299Updated last year
- Constrained Decoding for LLMs against JSON Schema☆327Updated 2 years ago
- Simple repo that compiles and runs llama2.c on the Web☆55Updated last year
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆223Updated last year
- ☆458Updated last year
- Code Indexer Loop is a Python library for indexing and retrieving source code files through an integrated vector database that's continuo…☆175Updated last year
- ☆113Updated 3 months ago
- ☆156Updated 10 months ago
- ☆135Updated last year
- JS tokenizer for LLaMA 3 and LLaMA 3.1☆109Updated 2 months ago
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes…☆147Updated last year
- LLaMA Cog template☆308Updated last year
- An Autonomous LLM Agent that runs on Wizcoder-15B☆335Updated 6 months ago
- This project is an attempt to create a common metric to test LLM's for progress in eliminating hallucinations which is the most serious c…☆222Updated 2 years ago
- WebGPU LLM inference tuned by hand☆149Updated last year
- This repository contains code for extending the Stanford Alpaca synthetic instruction tuning to existing instruction-tuned models such as…☆352Updated last year
- ☆198Updated last year
- Add local LLMs to your Web or Electron apps! Powered by Rust + WebGPU☆102Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆123Updated last year
- ☆412Updated last year
- Command-line script for inferencing from models such as falcon-7b-instruct☆76Updated last year
- Self-extracting GPT prompts for ~70% token savings☆218Updated 2 years ago
- A simple "Be My Eyes" web app with a llama.cpp/llava backend☆489Updated last year