NVIDIA / workbench-example-nemotron-finetune
An NVIDIA AI Workbench example project for fine-tuning a Nemotron-3 8B model
☆39Updated 7 months ago
Related projects ⓘ
Alternatives and complementary repositories for workbench-example-nemotron-finetune
- End-to-End LLM Guide☆97Updated 4 months ago
- ☆62Updated 4 months ago
- Unofficial implementation of https://arxiv.org/pdf/2407.14679☆36Updated 2 months ago
- ☆51Updated 6 months ago
- LLM reads a paper and produce a working prototype☆36Updated last week
- ☆41Updated 8 months ago
- Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench☆107Updated this week
- ☆49Updated this week
- A repository of Python scripts to scrape code contents of the public repositories of `huggingface`.☆43Updated 8 months ago
- ☆28Updated 8 months ago
- Lean implementation of various multi-agent LLM methods, including Iteration of Thought (IoT)☆53Updated 2 months ago
- ☆19Updated 3 months ago
- This is an NVIDIA AI Workbench example project that demonstrates an end-to-end model development workflow using Llamafactory.☆37Updated last month
- Official repository for RAGViz: Diagnose and Visualize Retrieval-Augmented Generation [EMNLP 2024]☆54Updated this week
- Self-host LLMs with vLLM and BentoML☆74Updated last week
- LLaMA 3 is one of the most promising open-source model after Mistral, we will recreate it's architecture in a simpler manner.☆104Updated 2 months ago
- Agentic RAG with Langchain, Qdrant and CrewAI☆38Updated 6 months ago
- ☆55Updated this week
- inference code for mixtral-8x7b-32kseqlen☆98Updated 11 months ago
- Beating the GAIA benchmark with Transformers Agents. 🚀☆62Updated 3 weeks ago
- An NVIDIA AI Workbench example project for fine-tuning a Mistral 7B model☆49Updated 5 months ago
- Evaluation of bm42 sparse indexing algorithm☆62Updated 4 months ago
- [EMNLP 2024] LongRAG: A Dual-perspective Retrieval-Augmented Generation Paradigm for Long-Context Question Answering☆80Updated 2 weeks ago
- ☆59Updated last month
- Eh, simple and works.☆27Updated 11 months ago
- This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows inste…☆116Updated 8 months ago
- Data preparation code for Amber 7B LLM☆83Updated 6 months ago
- Official Implementation of "Multi-Head RAG: Solving Multi-Aspect Problems with LLMs"☆176Updated 3 weeks ago
- a collection of resources around LLMs, aggregated for the workshop "Mastering LLMs: End-to-End Fine-Tuning and Deployment" by Dan Becker …☆106Updated 5 months ago