runpod / runpod-python
🐍 | Python library for RunPod API and serverless worker SDK.
☆228Updated last month
Alternatives and similar repositories for runpod-python
Users that are interested in runpod-python are comparing it to the libraries listed below
Sorting:
- Starting point to build your own custom serverless endpoint☆104Updated last week
- 🧰 | RunPod CLI for pod management☆304Updated 4 months ago
- A curated list of amazing RunPod projects, libraries, and resources☆112Updated 8 months ago
- 🐳 | Dockerfiles for the RunPod container images used for our official templates.☆181Updated 2 weeks ago
- The RunPod worker template for serving our large language model endpoints. Powered by vLLM.☆313Updated last week
- Examples of models deployable with Truss☆170Updated this week
- ☆52Updated last year
- ☆84Updated last year
- Running Ollama with Runpod☆59Updated 9 months ago
- A fast batching API to serve LLM models☆182Updated last year
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆115Updated 11 months ago
- ⚙️ | REPLACED BY https://github.com/runpod-workers | Official set of serverless worker provided by RunPod as endpoints.☆58Updated last year
- RunPod Serverless Worker for Oobabooga Text Generation API for LLMs☆2Updated 11 months ago
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆123Updated last year
- The code we currently use to fine-tune models.☆114Updated last year
- ☆39Updated this week
- TheBloke's Dockerfiles☆303Updated last year
- Create embeddings with infinity as serverless endpoint☆28Updated last week
- LoRA inference model packaged with Cog☆74Updated last year
- 🖼️ | Quickly deploy a custom RunPod Endpoint API using your own model ckpt.☆27Updated 2 years ago
- ☆88Updated last year
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆64Updated last year
- Gradio UI for a Cog API☆67Updated last year
- LoRA Explorer model to test with LoRAs using Flux.1[Dev] as the base model☆47Updated 7 months ago
- An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.☆37Updated last year
- An endpoint server for efficiently serving quantized open-source LLMs for code.☆55Updated last year
- Falcon LLM ggml framework with CPU and GPU support☆246Updated last year
- Postgres/pgvector Python Client☆254Updated last month
- A GitHub Action for installing and using Cog so you can run, test, and push machine learning models☆70Updated 3 months ago
- ☆199Updated last year