vast-ai / vast-cliLinks
Vast.ai python and cli api client
☆174Updated last week
Alternatives and similar repositories for vast-cli
Users that are interested in vast-cli are comparing it to the libraries listed below
Sorting:
- 🐍 | Python library for RunPod API and serverless worker SDK.☆263Updated this week
- ☆122Updated last year
- 🧰 | Runpod CLI for pod management☆353Updated 3 weeks ago
- DiffusionWithAutoscaler☆29Updated last year
- ☆142Updated 2 years ago
- ☆64Updated last year
- ☆172Updated 10 months ago
- ☆54Updated 2 months ago
- ☆87Updated 2 years ago
- Examples of models deployable with Truss☆212Updated 2 weeks ago
- My swiftsknife for vast.ai service☆153Updated 2 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆53Updated 2 years ago
- 🏥 Health monitor for a Petals swarm☆40Updated last year
- ☆51Updated 2 years ago
- 🐳 | Dockerfiles for the RunPod container images used for our official templates.☆216Updated 3 weeks ago
- Drop in replacement for OpenAI, but with Open models.☆153Updated 2 years ago
- Deploy your GGML models to HuggingFace Spaces with Docker and gradio☆38Updated 2 years ago
- Making the world's first and smartest opensource any-to-any AGI system☆44Updated last month
- Using langchain, deeplake and openai to create a Q&A on the Mojo lang programming manual☆22Updated 2 years ago
- ☆119Updated last year
- A community list of common phrases generated by GPT and Claude models☆79Updated 2 years ago
- GPU accelerated client-side embeddings for vector search, RAG etc.☆65Updated 2 years ago
- Pipeline is an open source python SDK for building AI/ML workflows☆138Updated last year
- inference code for mixtral-8x7b-32kseqlen☆105Updated 2 years ago
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆167Updated 8 months ago
- A template to run LLaMA in Cog☆66Updated 2 years ago
- Modified Stanford-Alpaca Trainer for Training Replit's Code Model☆42Updated 2 years ago
- faster-whisper as serverless endpoint☆127Updated last month
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆64Updated 2 years ago
- Replace expensive LLM calls with finetunes automatically☆66Updated last year