Andrei-Aksionov / nanoGPTplus
☆47Updated last year
Alternatives and similar repositories for nanoGPTplus:
Users that are interested in nanoGPTplus are comparing it to the libraries listed below
- A high-throughput and memory-efficient inference and serving engine for LLMs☆263Updated 7 months ago
- experiments with inference on llama☆104Updated 11 months ago
- Reimplementation of the task generation part from the Alpaca paper☆119Updated 2 years ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆82Updated last year
- inference code for mixtral-8x7b-32kseqlen☆100Updated last year
- Minimal code to train a Large Language Model (LLM).☆168Updated 2 years ago
- Experiments with generating opensource language model assistants☆97Updated last year
- The Next Generation Multi-Modality Superintelligence☆71Updated 8 months ago
- ☆199Updated last year
- Our open source implementation of MiniLMv2 (https://aclanthology.org/2021.findings-acl.188)☆61Updated last year
- QLoRA with Enhanced Multi GPU Support☆37Updated last year
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆76Updated 6 months ago
- This repository contains the code for dataset curation and finetuning of instruct variant of the Bilingual OpenHathi model. The resultin…☆23Updated last year
- Supervised instruction finetuning for LLM with HF trainer and Deepspeed☆35Updated last year
- Find the optimal model serving solution for 🤗 Hugging Face models 🚀☆43Updated last year
- Functional local implementations of main model parallelism approaches☆95Updated 2 years ago
- Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks☆208Updated last year
- A Multilingual Dataset for Parsing Realistic Task-Oriented Dialogs☆114Updated 2 years ago
- 📚 Datasets and models for instruction-tuning☆237Updated last year
- Train your own small bitnet model☆70Updated 6 months ago
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 lines☆198Updated last year
- ☆92Updated last year
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆115Updated 2 years ago
- Drop in replacement for OpenAI, but with Open models.☆152Updated 2 years ago
- Experiments on speculative sampling with Llama models☆126Updated last year
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆231Updated 6 months ago
- Simple implementation of Speculative Sampling in NumPy for GPT-2.☆95Updated last year
- TitanML Takeoff Server is an optimization, compression and deployment platform that makes state of the art machine learning models access…☆114Updated last year
- Aana SDK is a powerful framework for building AI enabled multimodal applications.☆47Updated this week
- FineTune LLMs in few lines of code (Text2Text, Text2Speech, Speech2Text)☆239Updated last year