fpgasystems / Chameleon-RAG-AccelerationLinks
☆19Updated 4 months ago
Alternatives and similar repositories for Chameleon-RAG-Acceleration
Users that are interested in Chameleon-RAG-Acceleration are comparing it to the libraries listed below
Sorting:
- [SIGMOD 2025] PQCache: Product Quantization-based KVCache for Long Context LLM Inference☆71Updated last week
- ☆22Updated 2 weeks ago
- Artifact for OSDI'23: MGG: Accelerating Graph Neural Networks with Fine-grained intra-kernel Communication-Computation Pipelining on Mult…☆40Updated last year
- ArkVale: Efficient Generative LLM Inference with Recallable Key-Value Eviction (NIPS'24)☆43Updated 10 months ago
- [HPCA'24] Smart-Infinity: Fast Large Language Model Training using Near-Storage Processing on a Real System☆49Updated 3 months ago
- ☆24Updated last month
- A High-Throughput Multi-GPU System for Graph-Based Approximate Nearest Neighbor Search☆18Updated 3 months ago
- Vector search with bounded performance.☆36Updated last year
- PerFlow-AI is a programmable performance analysis, modeling, prediction tool for AI system.☆24Updated 2 weeks ago
- TiledLower is a Dataflow Analysis and Codegen Framework written in Rust.☆14Updated 11 months ago
- Supplemental materials for The ASPLOS 2025 / EuroSys 2025 Contest on Intra-Operator Parallelism for Distributed Deep Learning☆23Updated 5 months ago
- Artifact for "Apparate: Rethinking Early Exits to Tame Latency-Throughput Tensions in ML Serving" [SOSP '24]☆25Updated 11 months ago
- MAGIS: Memory Optimization via Coordinated Graph Transformation and Scheduling for DNN (ASPLOS'24)☆55Updated last year
- Scalable long-context LLM decoding that leverages sparsity—by treating the KV cache as a vector storage system.☆88Updated last month
- NEO is a LLM inference engine built to save the GPU memory crisis by CPU offloading☆67Updated 4 months ago
- ☆25Updated 2 years ago
- ☆39Updated 4 months ago
- Ginex: SSD-enabled Billion-scale Graph Neural Network Training on a Single Machine via Provably Optimal In-memory Caching☆40Updated last year
- Artifact for USENIX ATC'23: TC-GNN: Bridging Sparse GNN Computation and Dense Tensor Cores on GPUs.☆50Updated 2 years ago
- Open-source implementation for "Helix: Serving Large Language Models over Heterogeneous GPUs and Network via Max-Flow"☆68Updated last week
- Tacker: Tensor-CUDA Core Kernel Fusion for Improving the GPU Utilization while Ensuring QoS☆31Updated 8 months ago
- A Skew-Resistant Index for Processing-in-Memory☆26Updated last year
- ☆15Updated 3 years ago
- SHADE: Enable Fundamental Cacheability for Distributed Deep Learning Training☆35Updated 2 years ago
- APEX+ is an LLM Serving Simulator☆36Updated 4 months ago
- ☆55Updated 4 months ago
- FGNN's artifact evaluation (EuroSys 2022)☆17Updated 3 years ago
- Medusa: Accelerating Serverless LLM Inference with Materialization [ASPLOS'25]☆33Updated 5 months ago
- Compiler for Dynamic Neural Networks☆46Updated last year
- DISB is a new DNN inference serving benchmark with diverse workloads and models, as well as real-world traces.☆54Updated last year