broncotc / bitsandbytes-rocmLinks
☆37Updated 2 years ago
Alternatives and similar repositories for bitsandbytes-rocm
Users that are interested in bitsandbytes-rocm are comparing it to the libraries listed below
Sorting:
- ☆158Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆124Updated 2 years ago
- ☆535Updated last year
- A torchless, c++ rwkv implementation using 8bit quantization, written in cuda/hip/vulkan for maximum compatibility and minimum dependenci…☆314Updated last year
- C/C++ implementation of PygmalionAI/pygmalion-6b☆56Updated 2 years ago
- A gradio web UI for running Large Language Models like GPT-J 6B, OPT, GALACTICA, LLaMA, and Pygmalion.☆309Updated 2 years ago
- ☆404Updated 2 years ago
- ChatGPT-like Web UI for RWKVstic☆100Updated 2 years ago
- 4 bits quantization of SantaCoder using GPTQ☆51Updated 2 years ago
- 4 bits quantization of LLMs using GPTQ☆49Updated 2 years ago
- 8-bit CUDA functions for PyTorch Rocm compatible☆41Updated last year
- Automated prompting and scoring framework to evaluate LLMs using updated human knowledge prompts☆110Updated 2 years ago
- Framework agnostic python runtime for RWKV models☆146Updated 2 years ago
- This repo turns your PC into a AI Horde worker node☆275Updated 8 months ago
- Linux based GDDR6/GDDR6X VRAM temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆105Updated 5 months ago
- Oobabooga extension for Bark TTS☆118Updated last year
- SoTA Transformers with C-backend for fast inference on your CPU.☆309Updated last year
- ☆347Updated last year
- Inference code for LLaMA models☆189Updated 2 years ago
- Efficient 3bit/4bit quantization of LLaMA models☆19Updated 2 years ago
- Prototype UI for chatting with the Pygmalion models.☆235Updated 2 years ago
- rwkv_chatbot☆62Updated 2 years ago
- Generate Large Language Model text in a container.☆20Updated 2 years ago
- A repository to run gpt-j-6b on low vram machines (4.2 gb minimum vram for 2000 token context, 3.5 gb for 1000 token context). Model load…☆114Updated 3 years ago
- Simple, hackable and fast implementation for training/finetuning medium-sized LLaMA-based models☆180Updated 3 weeks ago
- ☆42Updated 2 years ago
- 8-bit CUDA functions for PyTorch, ported to HIP for use in AMD GPUs☆51Updated 2 years ago
- Falcon LLM ggml framework with CPU and GPU support☆247Updated last year
- Code for the paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot" with LLaMA implementation.☆71Updated 2 years ago
- A prompt/context management system☆170Updated 2 years ago