dingyuqing05 / trt2022_wenetLinks
☆71Updated 2 years ago
Alternatives and similar repositories for trt2022_wenet
Users that are interested in trt2022_wenet are comparing it to the libraries listed below
Sorting:
- ☆75Updated 2 years ago
- ☆99Updated 3 years ago
- Serving Inside Pytorch☆160Updated 2 weeks ago
- simplify >2GB large onnx model☆59Updated 6 months ago
- Simple Dynamic Batching Inference☆145Updated 3 years ago
- ☆139Updated last year
- Use PyTorch model in C++ project☆139Updated 3 years ago
- ☆26Updated last year
- symmetric int8 gemm☆66Updated 5 years ago
- TensorRT 2022复赛方案: 首个基于Transformer的图像重建模型MST++的TensorRT模型推断优化☆139Updated 2 years ago
- ☆26Updated last year
- export llama to onnx☆126Updated 5 months ago
- A Toolkit to Help Optimize Large Onnx Model☆157Updated last year
- Whisper inference with TensorRT-LLM☆22Updated last year
- Transformer related optimization, including BERT, GPT☆59Updated last year
- Compare multiple optimization methods on triton to imporve model service performance☆52Updated last year
- ☆120Updated 2 years ago
- ☢️ TensorRT 2023复赛——基于TensorRT-LLM的Llama模型推断加速优化☆48Updated last year
- Offline Quantization Tools for Deploy.☆129Updated last year
- Triton Inferece Server Model Config and Client Scripts☆32Updated 3 years ago
- ☆58Updated 7 months ago
- Trans different platform's network to International Representation(IR)☆44Updated 7 years ago
- ☆36Updated 8 months ago
- ONNX2Pytorch☆162Updated 4 years ago
- MegEngine到其他框架的转换器☆70Updated 2 years ago
- Models and examples built with OneFlow☆97Updated 8 months ago
- ☆21Updated 3 years ago
- NART = NART is not A RunTime, a deep learning inference framework.☆37Updated 2 years ago
- 使用 cutlass 实现 flash-attention 精简版,具有教学意义☆42Updated 10 months ago
- ☆148Updated 5 months ago