AIFlowPlayer / LMDeploy-JetsonLinks
Deploying LLMs offline on the NVIDIA Jetson platform marks the dawn of a new era in embodied intelligence, where devices can function independently without continuous internet access.
☆98Updated last year
Alternatives and similar repositories for LMDeploy-Jetson
Users that are interested in LMDeploy-Jetson are comparing it to the libraries listed below
Sorting:
- 基于昇腾310芯片的大语言模型部署☆20Updated last year
- ☆47Updated 8 months ago
- ☢️ TensorRT 2023复赛——基于TensorRT-LLM的Llama模型推断加速优化☆49Updated last year
- llm-export can export llm model to onnx.☆299Updated 5 months ago
- An onnx-based quantitation tool.☆71Updated last year
- ☆61Updated last year
- 大模型部署实战:TensorRT-LLM, Triton Inference Server, vLLM☆26Updated last year
- ☆22Updated last year
- 训练一个对中文支持更好的LLaVA模型,并开源训练代码和数据。☆64Updated 10 months ago
- ☆142Updated last year
- 该代码与B站上的视频 https://www.bilibili.com/video/BV18L41197Uz/?spm_id_from=333.788&vd_source=eefa4b6e337f16d87d87c2c357db8ca7 相关联。☆69Updated last year
- [EMNLP 2024 Industry Track] This is the official PyTorch implementation of "LLMC: Benchmarking Large Language Model Quantization with a V…☆510Updated last week
- 基于InternLM2大模型的离线具身智能导盲犬☆101Updated last year
- A Light-Weight Framework for Open-Set Object Detection with Decoupled Feature Alignment in Joint Space☆87Updated 6 months ago
- NVIDIA TensorRT Hackathon 2023复赛选题:通义千问Qwen-7B用TensorRT-LLM模型搭建及优化☆42Updated last year
- ☆86Updated 9 months ago
- A CPU Realtime VLM in 500M. Surpassed Moondream2 and SmolVLM. Training from scratch with ease.☆220Updated 2 months ago
- Run generative AI models in sophgo BM1684X/BM1688☆224Updated last week
- mllm-npu: training multimodal large language models on Ascend NPUs☆90Updated 10 months ago
- run ChatGLM2-6B in BM1684X☆49Updated last year
- Serving Inside Pytorch☆163Updated last week
- simplify >2GB large onnx model☆59Updated 7 months ago
- A light llama-like llm inference framework based on the triton kernel.☆134Updated this week
- DashInfer is a native LLM inference engine aiming to deliver industry-leading performance atop various hardware architectures, including …☆259Updated last month
- ☆53Updated 2 weeks ago
- Large Language Model Onnx Inference Framework☆36Updated 6 months ago
- Llama3 Streaming Chat Sample☆22Updated last year
- Compare multiple optimization methods on triton to imporve model service performance☆52Updated last year
- llm deploy project based onnx.☆42Updated 9 months ago
- 一大波学习onnx的案例☆19Updated 9 months ago