basetenlabs / Workshop-TRT-LLMLinks
☆19Updated last year
Alternatives and similar repositories for Workshop-TRT-LLM
Users that are interested in Workshop-TRT-LLM are comparing it to the libraries listed below
Sorting:
- Fine-tune an LLM to perform batch inference and online serving.☆112Updated 2 months ago
- A miniature version of Modal☆20Updated last year
- An introduction to LLM Sampling☆79Updated 7 months ago
- A set of scripts and notebooks on LLM finetunning and dataset creation☆110Updated 10 months ago
- ☆124Updated 9 months ago
- Build Agentic workflows with function calling using open LLMs☆28Updated this week
- Google TPU optimizations for transformers models☆117Updated 6 months ago
- experiments with inference on llama☆104Updated last year
- LLM training in simple, raw C/CUDA☆15Updated 8 months ago
- A repository to unravel the language of GPUs, making their kernel conversations easy to understand☆188Updated 2 months ago
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.☆33Updated 2 months ago
- Set of scripts to finetune LLMs☆37Updated last year
- ML/DL Math and Method notes☆62Updated last year
- ☆79Updated last year
- Just a bunch of benchmark logs for different LLMs☆119Updated last year
- 🕹️ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.☆137Updated last year
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆49Updated last year
- Collection of scripts and notebooks for OpenAI's latest GPT OSS models☆222Updated this week
- Matrix (Multi-Agent daTa geneRation Infra and eXperimentation framework) is a versatile engine for multi-agent conversational data genera…☆81Updated last week
- a pipeline for using api calls to agnostically convert unstructured data into structured training data☆30Updated 10 months ago
- ☆23Updated 2 years ago
- Manage scalable open LLM inference endpoints in Slurm clusters☆268Updated last year
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆82Updated last year
- Fun project: LLM powered RAG Discord Bot that works seamlessly on CPU☆32Updated last year
- ☆48Updated 9 months ago
- 👷 Build compute kernels☆87Updated last week
- Doing simple retrieval from LLM models at various context lengths to measure accuracy☆102Updated last year
- Code for NeurIPS LLM Efficiency Challenge☆59Updated last year
- ☆20Updated 9 months ago
- Cray-LM unified training and inference stack.☆22Updated 6 months ago