ForceInjection / AI-fundermentalsLinks
AI 基础知识 - GPU 架构、CUDA 编程、大模型基础及AI Agent 相关知识
☆464Updated this week
Alternatives and similar repositories for AI-fundermentals
Users that are interested in AI-fundermentals are comparing it to the libraries listed below
Sorting:
- This repository organizes materials, recordings, and schedules related to AI-infra learning meetings.☆133Updated last week
- 一种任务级GPU算力分时调度的高性能深度学习训练平台☆696Updated last year
- ☆319Updated 2 months ago
- GLake: optimizing GPU memory management and IO transmission.☆480Updated 5 months ago
- A self-learning tutorail for CUDA High Performance Programing.☆735Updated 2 months ago
- LLM notes, including model inference, transformer model structure, and llm framework code analysis notes.☆819Updated this week
- Community maintained hardware plugin for vLLM on Ascend☆1,128Updated this week
- Learning Machine Learning, The Chinese Taoist Way☆438Updated 5 years ago
- ☆497Updated last week
- HAMi-core compiles libvgpu.so, which ensures hard limit on GPU in container☆214Updated this week
- ☆535Updated last year
- Open Source Landscapes and Insights Produced by AntOSS☆76Updated this week
- The IX device plugin is a DaemonSet for Kubernetes, which can help to expose the Iluvatar GPU in the Kubernetes cluster.☆17Updated last month
- ☆353Updated this week
- Using CRDs to manage GPU resources in Kubernetes.☆209Updated 2 years ago
- RTP-LLM: Alibaba's high-performance LLM inference engine for diverse applications.☆851Updated last week
- Omni_Infer is a suite of inference accelerators designed for the Ascend NPU platform, offering native support and an expanding feature se…☆71Updated last week
- LLM Inference benchmark☆426Updated last year
- Hooked CUDA-related dynamic libraries by using automated code generation tools.☆166Updated last year
- A kubernetes plugin which enables dynamically add or remove GPU resources for a running Pod☆127Updated 3 years ago
- Device-plugin for volcano vgpu which support hard resource isolation☆103Updated 2 months ago
- how to learn PyTorch and OneFlow☆451Updated last year
- ☆90Updated 5 months ago
- LLM全栈优质资源汇总☆629Updated 2 months ago
- ☆52Updated 2 months ago
- ☆124Updated 7 months ago
- OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)☆269Updated this week
- Run your deep learning workloads on Kubernetes more easily and efficiently.☆531Updated last year
- Kubernetes Operator for AI and Bigdata Elastic Training☆88Updated 8 months ago
- KV cache store for distributed LLM inference☆330Updated last week