CVxTz / llm-serve-tutorialLinks
☆20Updated last year
Alternatives and similar repositories for llm-serve-tutorial
Users that are interested in llm-serve-tutorial are comparing it to the libraries listed below
Sorting:
- Code Repository for Blog - How to Productionize Large Language Models (LLMs)☆12Updated last year
- Data extraction with LLM on CPU☆112Updated 2 years ago
- 🤖 AI Assistant fine-tuned to provide support for coding and design questions based on the latest trends in the industry.☆17Updated 2 years ago
- This is a repository for the course "From Beginner to LLM Developer" by Towards AI.☆12Updated last year
- A collection of hand on notebook for LLMs practitioner☆51Updated last year
- Agent Watch is an AgentOps monitoring library designed for Crew AI applications.☆21Updated last year
- GenAI Experimentation☆59Updated 4 months ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆51Updated last year
- Repository of the code base for KT Generation process that we worked at Google Cloud and Searce GenAI Hackathon.☆77Updated 2 years ago
- Large Language Model (LLM) Inference API and Chatbot☆127Updated last year
- ☆55Updated 4 months ago
- Mistral + Haystack: build RAG pipelines that rock 🤘☆106Updated last year
- Table detection with Florence.☆15Updated last year
- ☆44Updated last year
- Data extraction with LLM on CPU☆270Updated last year
- ☆14Updated last year
- A semantic research engine to get relevant papers based on a user query. Application frontend with Chainlit Copilot. Observability with L…☆81Updated last year
- Machine Learning Serving focused on GenAI with simplicity as the top priority.☆59Updated last week
- ☆80Updated last year
- Fine-tune an LLM to perform batch inference and online serving.☆115Updated 7 months ago
- ☆45Updated last year
- High level library for batched embeddings generation, blazingly-fast web-based RAG and quantized indexes processing ⚡☆69Updated last month
- Data extraction with LLM on CPU☆86Updated 2 years ago
- ☆15Updated 2 years ago
- Dynamic Metadata based RAG Framework☆78Updated last month
- Data extraction with LLM on CPU☆68Updated 2 years ago
- LlamaWorksDB is a Retrieval Augmented Generation (RAG) product designed to interact with the documentation of various products such as Ll…☆17Updated last year
- Fun project: LLM powered RAG Discord Bot that works seamlessly on CPU☆33Updated 2 years ago
- Test LLMs automatically with Giskard and CI/CD☆31Updated last year
- Multi-Agent LLM System for Digital Scam Protection☆12Updated last year