recmo / cria
Tiny inference-only implementation of LLaMA
☆92Updated 10 months ago
Alternatives and similar repositories for cria:
Users that are interested in cria are comparing it to the libraries listed below
- A library for incremental loading of large PyTorch checkpoints☆56Updated last year
- ☆153Updated last year
- Command-line script for inferencing from models such as MPT-7B-Chat☆101Updated last year
- Chat Markup Language conversation library☆55Updated last year
- Simple embedding -> text model trained on a small subset of Wikipedia sentences.☆153Updated last year
- ☆143Updated last year
- Simplex Random Feature attention, in PyTorch☆74Updated last year
- ☆107Updated last year
- An implementation of bucketMul LLM inference☆215Updated 7 months ago
- an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated last year
- Embedding models from Jina AI☆58Updated last year
- Run GGML models with Kubernetes.☆174Updated last year
- ☆163Updated 8 months ago
- Use context-free grammars with an LLM☆168Updated 10 months ago
- Full finetuning of large language models without large memory requirements☆93Updated last year
- utilities for loading and running text embeddings with onnx☆44Updated 6 months ago
- [Added T5 support to TRLX] A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)☆47Updated 2 years ago
- Proxy server for triton gRPC server that inferences embedding model in Rust☆20Updated 6 months ago
- Turing machines, Rule 110, and A::B reversal using Claude 3 Opus.☆59Updated 9 months ago
- Private inference over your sensitive data with off-the-shelf models☆34Updated last year
- Modified Stanford-Alpaca Trainer for Training Replit's Code Model☆40Updated last year
- Binary vector search example using Unum's USearch engine and pre-computed Wikipedia embeddings from Co:here and MixedBread☆18Updated 10 months ago
- A miniature version of Modal☆19Updated 8 months ago
- Implement recursion using English as the programming language and an LLM as the runtime.☆136Updated last year
- Array-Inspired Pipeline Language☆119Updated last year
- WebGPU LLM inference tuned by hand☆148Updated last year
- The Fast Vector Similarity Library is designed to provide efficient computation of various similarity measures between vectors.☆379Updated 5 months ago
- An HTTP serving framework by Banana☆98Updated last year
- Command-line script for inferencing from models such as falcon-7b-instruct☆76Updated last year
- Mistral7B playing DOOM☆127Updated 7 months ago