vast-ai / vast-cliLinks
Vast.ai python and cli api client
☆152Updated this week
Alternatives and similar repositories for vast-cli
Users that are interested in vast-cli are comparing it to the libraries listed below
Sorting:
- 🐍 | Python library for RunPod API and serverless worker SDK.☆239Updated this week
- 🧰 | RunPod CLI for pod management☆320Updated last month
- 🐳 | Dockerfiles for the RunPod container images used for our official templates.☆196Updated this week
- Starting point to build your own custom serverless endpoint☆111Updated 2 months ago
- ☆122Updated last year
- ☆63Updated 6 months ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆64Updated last year
- Pipeline is an open source python SDK for building AI/ML workflows☆134Updated 9 months ago
- ☆141Updated last year
- ☆171Updated 5 months ago
- DiffusionWithAutoscaler☆29Updated last year
- ☆55Updated 2 years ago
- ☆85Updated 2 years ago
- Examples of models deployable with Truss☆189Updated last week
- inference code for mixtral-8x7b-32kseqlen☆100Updated last year
- RunPod Serverless Worker for Oobabooga Text Generation API for LLMs☆2Updated last year
- My swiftsknife for vast.ai service☆138Updated 5 months ago
- A template to run LLaMA in Cog☆66Updated 2 years ago
- ☆52Updated last week
- A curated list of amazing RunPod projects, libraries, and resources☆117Updated 10 months ago
- Deploy your HPC Cluster on AWS in 20min. with just 1-Click.☆55Updated 3 months ago
- ☆50Updated last year
- Utility library to work with character cards and roleplay AI in general☆33Updated last year
- Chat to Compose Video☆189Updated last year
- ☆199Updated last year
- An OpenAI API compatible LLM inference server based on ExLlamaV2.☆25Updated last year
- Simple setup to self-host LLaMA3-70B model with an OpenAI API☆19Updated last year
- ⚙️ | REPLACED BY https://github.com/runpod-workers | Official set of serverless worker provided by RunPod as endpoints.☆60Updated last month
- ☆40Updated 2 months ago
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆123Updated 2 years ago