chenhunghan / ialacolLinks
πͺΆ Lightweight OpenAI drop-in replacement for Kubernetes
β147Updated last year
Alternatives and similar repositories for ialacol
Users that are interested in ialacol are comparing it to the libraries listed below
Sorting:
- Open Weight, tool-calling LLMsβ155Updated last year
- Finetune LLMs on K8s by using Runbooksβ170Updated last year
- Kuberentes LangChain Agent - Interact with Kubernetes Clusters using LLMsβ27Updated 2 years ago
- β‘Instant Stable Diffusion on k8s(Kubernetes) with Helmβ92Updated 2 years ago
- An open-source cloud-native of large multi-modal models (LMMs) serving framework.β164Updated 2 years ago
- Chart for deploying ChromaDB in Kubernetesβ53Updated 7 months ago
- Knowledge for GPTScriptβ29Updated last year
- OpenAI compatible API for LLMs and embeddings (LLaMA, Vicuna, ChatGLM and many others)β275Updated 2 years ago
- A fully in-browser privacy solution to make Conversational AI privacy-friendlyβ233Updated last year
- Helm charts to deploy Weaviate to k8sβ65Updated 2 months ago
- go-skynet helm chart repositoryβ72Updated last month
- β67Updated 9 months ago
- Extensible generative AI platform on Kubernetes with OpenAI-compatible APIs.β91Updated 2 months ago
- ποΈ Fine-tune, build, and deploy open-source LLMs easily!β502Updated last week
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.β118Updated last year
- An OpenAI-like LLaMA inference APIβ113Updated 2 years ago
- One Repo To Quickly Build One Docker File for HuggingChat Front and BackEndβ26Updated 2 years ago
- Host the GPTQ model using AutoGPTQ as an API that is compatible with text generation UI API.β90Updated 2 years ago
- Python client library for improving your LLM app accuracyβ97Updated 10 months ago
- A OpenAI API compatible REST server for llama.β208Updated 10 months ago
- An endpoint server for efficiently serving quantized open-source LLMs for code.β58Updated 2 years ago
- Visual Studio Code extension for WizardCoderβ148Updated 2 years ago
- starcoder server for huggingface-vscdoe custom endpointβ179Updated 2 years ago
- π’ Yet another operator for running large language models on Kubernetes with ease. Powered by Ollama! π«β227Updated last week
- a curated collection of models ready-to-use with LocalAIβ270Updated last year
- β173Updated last year
- A high performance batching router optimises max throughput for text inference workloadβ16Updated 2 years ago
- β39Updated 2 years ago
- Self-host LLMs with vLLM and BentoMLβ163Updated last month
- π‘ Deploy AI models and apps to Kubernetes without developing a herniaβ33Updated last year