liguodongiot / netron-flaskLinks
模型可视化工具netron的Flask版本
☆18Updated 3 years ago
Alternatives and similar repositories for netron-flask
Users that are interested in netron-flask are comparing it to the libraries listed below
Sorting:
- ☢️ TensorRT 2023复赛——基于TensorRT-LLM的Llama模型推断加速优化☆50Updated last year
- CUDA 编程指南学习☆31Updated 6 years ago
- run ChatGLM2-6B in BM1684X☆49Updated last year
- A high performance, high expansion, easy to use framework for AI application. 为AI应用的开发者提供一套统一的高性能、易用的编程框架,快速基于AI全栈服务、开发跨端边云的AI行业应用,支持GPU,…☆156Updated last year
- Music large model based on InternLM2-chat.☆22Updated 9 months ago
- 模型压缩的小白入门教程☆22Updated last year
- ncnn和pnnx格式编辑器☆137Updated 11 months ago
- NVIDIA TensorRT Hackathon 2023复赛选题:通义千问Qwen-7B用TensorRT-LLM模 型搭建及优化☆42Updated last year
- autoTVM神经网络推理代码优化搜索演示,基于tvm编译开源模型centerface,并使用autoTVM搜索最优推理代码, 最终部署编译为c++代码,演示平台是cuda,可以是其他平台,例如树莓派,安卓手机,苹果手机.Thi is a demonstration of …☆28Updated 4 years ago
- 大模型API性能指标比较 - 深入分析TTFT、TPS等关键指标☆19Updated last year
- Snapdragon Neural Processing Engine (SNPE) SDKThe Snapdragon Neural Processing Engine (SNPE) is a Qualcomm Snapdragon software accelerate…☆35Updated 3 years ago
- 跨平台的容器化Linux桌面环境☆73Updated 7 months ago
- 大模型部署实战:TensorRT-LLM, Triton Inference Server, vLLM☆26Updated last year
- PaddlePaddle Developer Community☆124Updated last week
- llm-inference is a platform for publishing and managing llm inference, providing a wide range of out-of-the-box features for model deploy…☆86Updated last year
- 关于自建AI推理引擎的手册,从0开始你需要知道的所有事情☆270Updated 3 years ago
- run chatglm3-6b in BM1684X☆40Updated last year
- DeepSparkHub selects hundreds of application algorithms and models, covering various fields of AI and general-purpose computing, to suppo…☆67Updated last week
- Tiny C++ LLM inference implementation from scratch☆66Updated 2 weeks ago
- 百度QA100万数据集☆48Updated last year
- 官方transformers源码解析。AI大模型时代,pytorch、transformer是新操作系统,其他都是运行在其上面的软件。☆17Updated 2 years ago
- Datasets, Transforms and Models specific to Computer Vision☆89Updated last year
- unify-easy-llm(ULM)旨在打造一个简易的一键式大模型训练工具,支持Nvidia GPU、Ascend NPU等不同硬件以及常用的大模型。☆57Updated last year
- Transformer related optimization, including BERT, GPT☆17Updated 2 years ago
- Edge : a computing engine base C++/C☆117Updated 5 months ago
- ☆125Updated last year
- ☆135Updated 7 months ago
- 关于书籍CUDA Programming使用了pycuda模块的Python版本的示例代码☆257Updated 5 years ago
- ☆36Updated last year
- deploy onnx models with TensorRT and LibTorch☆18Updated 3 years ago