agrocylo / bitsandbytes-rocm
8-bit CUDA functions for PyTorch, ported to HIP for use in AMD GPUs
☆44Updated last year
Related projects ⓘ
Alternatives and complementary repositories for bitsandbytes-rocm
- 8-bit CUDA functions for PyTorch Rocm compatible☆39Updated 7 months ago
- 8-bit CUDA functions for PyTorch☆38Updated last week
- Fast and memory-efficient exact attention☆139Updated this week
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆89Updated this week
- 4 bits quantization of LLaMA using GPTQ, ported to HIP for use in AMD GPUs.☆32Updated last year
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆66Updated last year
- An unsupervised model merging algorithm for Transformers-based language models.☆100Updated 6 months ago
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆173Updated last month
- ☆37Updated last year
- Wheels for llama-cpp-python compiled with cuBLAS support☆94Updated 9 months ago
- Implements harmful/harmless refusal removal using pure HF Transformers☆25Updated 5 months ago
- DEPRECATED!☆53Updated 5 months ago
- Automated prompting and scoring framework to evaluate LLMs using updated human knowledge prompts☆111Updated last year
- Efficient 3bit/4bit quantization of LLaMA models☆19Updated last year
- 4 bits quantization of LLMs using GPTQ☆47Updated last year
- 5X faster 60% less memory QLoRA finetuning☆21Updated 5 months ago
- Science-driven chatbot development☆55Updated 6 months ago
- A KoboldAI-like memory extension for oobabooga's text-generation-webui☆107Updated 3 weeks ago
- ☆150Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆124Updated last year
- A prompt/context management system☆165Updated last year
- Merge Transformers language models by use of gradient parameters.☆201Updated 3 months ago
- Text WebUI extension to add clever Notebooks to Chat mode☆133Updated 10 months ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆126Updated 6 months ago
- Low-Rank adapter extraction for fine-tuned transformers model☆162Updated 6 months ago
- Model REVOLVER, a human in the loop model mixing system.☆33Updated last year
- ☆52Updated 5 months ago
- Code for the paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot" with LLaMA implementation.☆70Updated last year
- Python bindings for ggml☆132Updated 2 months ago