brilliantrough / not-powerful-translatorLinks
想打造一个全能的翻译,但并不全能,甚至全不能。。
☆10Updated 4 months ago
Alternatives and similar repositories for not-powerful-translator
Users that are interested in not-powerful-translator are comparing it to the libraries listed below
Sorting:
- Code release for AdapMoE accepted by ICCAD 2024☆26Updated 2 months ago
- ☆101Updated last year
- ☆110Updated 3 weeks ago
- 中国科学院大学-C语言编程-五子棋☆13Updated last year
- Lab for Parallel computing (USTC COMP6201P)☆22Updated last year
- [CVPR 2025] APHQ-ViT: Post-Training Quantization with Average Perturbation Hessian Based Reconstruction for Vision Transformers☆23Updated 2 months ago
- ☆21Updated last year
- A co-design architecture on sparse attention☆52Updated 3 years ago
- ☆44Updated 6 months ago
- ArkVale: Efficient Generative LLM Inference with Recallable Key-Value Eviction (NIPS'24)☆40Updated 6 months ago
- [ICLR2025]: OSTQuant: Refining Large Language Model Quantization with Orthogonal and Scaling Transformations for Better Distribution Fitt…☆61Updated 2 months ago
- Open source RTL implementation of Tensor Core, Sparse Tensor Core, BitWave and SparSynergy in the article: "SparSynergy: Unlocking Flexib…☆18Updated 3 months ago
- ☆160Updated 11 months ago
- An efficient spatial accelerator enabling hybrid sparse attention mechanisms for long sequences☆28Updated last year
- List of papers related to Vision Transformers quantization and hardware acceleration in recent AI conferences and journals.☆91Updated last year
- Tender: Accelerating Large Language Models via Tensor Decompostion and Runtime Requantization (ISCA'24)☆16Updated 11 months ago
- PyTorch implementation of PTQ4DiT https://arxiv.org/abs/2405.16005☆30Updated 7 months ago
- H2-LLM: Hardware-Dataflow Co-Exploration for Heterogeneous Hybrid-Bonding-based Low-Batch LLM Inference☆19Updated 2 months ago
- [HPCA'21] SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning☆94Updated 10 months ago
- ☆54Updated last year
- An Automatic Synthesis Tool for PIM-based CNN Accelerators.☆12Updated last year
- ☆11Updated 9 months ago
- Github repository of HPCA 2025 paper "UniNDP: A Unified Compilation and Simulation Tool for Near DRAM Processing Architectures"☆13Updated 6 months ago
- ViTALiTy (HPCA'23) Code Repository☆23Updated 2 years ago
- Curated collection of papers in MoE model inference☆200Updated 4 months ago
- 此项目是我个人对MIT 6.5940 课程作业的答案,学习笔记和心得。☆14Updated last year
- LLMA = LLM + Arithmetic coder, which use LLM to do insane text data compression. LLMA=大模型+算术编码,它能使用LLM对文本数据进行暴力的压缩,达到极高的压缩率。☆17Updated 7 months ago
- This repo contains the Assignments from Cornell Tech's ECE 5545 - Machine Learning Hardware and Systems offered in Spring 2023☆32Updated 2 years ago
- SpInfer: Leveraging Low-Level Sparsity for Efficient Large Language Model Inference on GPUs☆48Updated 3 months ago
- Vitis 部署加速器工作流介绍☆10Updated 5 months ago