abacaj / replit-3B-inferenceLinks
Run inference on replit-3B code instruct model using CPU
β155Updated last year
Alternatives and similar repositories for replit-3B-inference
Users that are interested in replit-3B-inference are comparing it to the libraries listed below
Sorting:
- β134Updated last year
- π The open-source autonomous agent LLM initiative πβ91Updated last year
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAIβ221Updated last year
- Command-line script for inferencing from models such as MPT-7B-Chatβ101Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRAβ122Updated last year
- llama.cpp with BakLLaVA model describes what does it seeβ383Updated last year
- Chat Bot with LLM and Fact Reference. RAG(Retrieval Augmented Generation) and LangChain backedβ129Updated last year
- Full finetuning of large language models without large memory requirementsβ93Updated last year
- A Personalised AI Assistant Inspired by 'Diamond Age, Powered by SMSβ92Updated 2 years ago
- β113Updated 5 months ago
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytesβ¦β147Updated last year
- Your automated SWE fleet to get your tickets from the Backlog to Prod!β96Updated last year
- Small finetuned LLMs for a diverse set of useful tasksβ125Updated last year
- An mlx project to train a base model on your whatsapp chats using (Q)Lora finetuningβ166Updated last year
- Extend the original llama.cpp repo to support redpajama model.β117Updated 9 months ago
- automatically generate @openai plugins by specifying your API in markdown in smol-developer styleβ120Updated 2 years ago
- A Simple Discord Bot for the Alpaca LLMβ100Updated last year
- β120Updated last year
- An Autonomous LLM Agent that runs on Wizcoder-15Bβ335Updated 7 months ago
- Open Source Embeddings Optimisation and Eval Framework for RAG/LLM Applications. Documentations at https://docs.vectorboard.ai/introductiβ¦β49Updated last year
- inference code for mixtral-8x7b-32kseqlenβ99Updated last year
- β79Updated last year
- A collection of LLM services you can self host via docker or modal labs to support your applications developmentβ187Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMsβ78Updated last year
- an implementation of Self-Extend, to expand the context window via grouped attentionβ119Updated last year
- Modified Stanford-Alpaca Trainer for Training Replit's Code Modelβ40Updated 2 years ago
- The code we currently use to fine-tune models.β113Updated last year
- The one who calls upon functions - Function-Calling Language Modelβ36Updated last year
- β48Updated last year
- LangChain chat model abstractions for dynamic failover, load balancing, chaos engineering, and more!β81Updated last year