l4rz / building-a-poor-mans-supercomputer
I've built a 4x V100 box for less than $5,500.
☆137Updated 3 years ago
Alternatives and similar repositories for building-a-poor-mans-supercomputer:
Users that are interested in building-a-poor-mans-supercomputer are comparing it to the libraries listed below
- Running SXM2/SXM3/SXM4 NVidia data center GPUs in consumer PCs☆102Updated last year
- A library for incremental loading of large PyTorch checkpoints☆56Updated 2 years ago
- ☆56Updated 2 years ago
- ☆130Updated 2 years ago
- Hashed Lookup Table based Matrix Multiplication (halutmatmul) - Stella Nera accelerator☆209Updated last year
- Unified API to facilitate usage of pre-trained "perceptor" models, a la CLIP☆39Updated 2 years ago
- Unsupported GPUs in Dell C4130 get throttled, here's how to prevent this from happening.☆123Updated 3 years ago
- Babysit your preemptible TPUs☆85Updated 2 years ago
- Latent Diffusion Language Models☆68Updated last year
- Embeddings focused small version of Llama NLP model☆103Updated last year
- Anim·E, Anime Enhanced dalle mini☆42Updated 2 years ago
- Inference code for LLaMA models☆42Updated 2 years ago
- ☆32Updated 3 months ago
- The GeoV model is a large langauge model designed by Georges Harik and uses Rotary Positional Embeddings with Relative distances (RoPER).…☆122Updated last year
- Simple setup to self-host LLaMA3-70B model with an OpenAI API☆19Updated last year
- tinygrad port of the RWKV large language model.☆44Updated last month
- Tune MPTs☆84Updated last year
- RWKV-v2-RNN trained on the Pile. See https://github.com/BlinkDL/RWKV-LM for details.☆67Updated 2 years ago
- Thispersondoesnotexist went down, so this time, while building it back up, I am going to open source all of it.☆90Updated last year
- ☆88Updated 2 years ago
- Command-line script for inferencing from models such as MPT-7B-Chat☆101Updated last year
- WebGPU LLM inference tuned by hand☆149Updated last year
- Quantized inference code for LLaMA models☆13Updated 2 years ago
- Extend the original llama.cpp repo to support redpajama model.☆117Updated 7 months ago
- ☆74Updated last year
- Swarm training framework using Haiku + JAX + Ray for layer parallel transformer language models on unreliable, heterogeneous nodes☆238Updated last year
- ☆130Updated 2 years ago
- Various handy scripts to quickly setup new Linux and Windows sandboxes, containers and WSL.☆40Updated this week
- Backend for the diffusion-ui frontend☆25Updated last year
- Code for the paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot" with LLaMA implementation.☆71Updated 2 years ago