intel / Training-and-Learning-SuiteLinks
Application Toolkit to perform E2E (End-To-End) Training inferencing
☆13Updated 2 years ago
Alternatives and similar repositories for Training-and-Learning-Suite
Users that are interested in Training-and-Learning-Suite are comparing it to the libraries listed below
Sorting:
- Achieve state of the art inference performance with modern accelerators on Kubernetes☆2,429Updated last week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆428Updated this week
- A unified library of SOTA model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresse…☆1,925Updated this week
- Generative AI Examples is a collection of GenAI examples such as ChatQnA, Copilot, which illustrate the pipeline capabilities of the Open…☆718Updated last week
- A Datacenter Scale Distributed Inference Serving Framework☆6,052Updated this week
- Reference implementations of MLPerf® inference benchmarks☆1,525Updated this week
- Edge Insights for Vision (eiv) is a package that helps to auto install Intel® GPU drivers and setup environment for Inference application…☆20Updated 4 months ago
- Tools for easier OpenVINO development/debugging☆10Updated 6 months ago
- Intel® AI Reference Models: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors and Inte…☆728Updated this week
- Supercharge Your LLM with the Fastest KV Cache Layer☆6,839Updated this week
- Kubernetes enhancements for Network Topology Aware Gang Scheduling & Autoscaling☆159Updated this week
- Cost-efficient and pluggable Infrastructure components for GenAI inference☆4,603Updated this week
- NCCL Tests☆1,423Updated this week
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆2,010Updated this week
- NVIDIA Inference Xfer Library (NIXL)☆876Updated this week
- ☆152Updated last month
- Tools for building GPU clusters☆1,416Updated last month
- This repository contains tutorials and examples for Triton Inference Server☆819Updated this week
- Collection of awesome LLM apps with AI Agents and RAG using OpenAI, Anthropic, Gemini and opensource models.☆94Updated 5 months ago
- ☆29Updated last year
- MIG Partition Editor for NVIDIA GPUs☆240Updated this week
- Run cloud native workloads on NVIDIA GPUs☆223Updated 2 weeks ago
- AI cloud native pipeline for confidential and sustainable computing☆39Updated 6 months ago
- Python client for the Run:ai REST API☆23Updated last month
- KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale☆1,111Updated this week
- A framework for efficient model inference with omni-modality models☆2,659Updated this week
- NVIDIA Data Center GPU Manager (DCGM) is a project for gathering telemetry and measuring the health of NVIDIA GPUs☆658Updated 2 months ago
- ☆32Updated last week
- A curated list of 100+ libraries and frameworks for AI engineers building with LLMs☆2,472Updated 2 months ago
- Qualcomm Cloud AI SDK (Platform and Apps) enable high performance deep learning inference on Qualcomm Cloud AI platforms delivering high …☆71Updated 2 months ago