microsoft / unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
β19,545Updated 3 weeks ago
Related projects: β
- π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.β15,839Updated this week
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.β30,165Updated last week
- CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an imageβ24,723Updated last month
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.β34,719Updated this week
- Fast and memory-efficient exact attentionβ13,401Updated this week
- A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) trainingβ19,845Updated last month
- Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"β10,327Updated last month
- π€ Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.β25,168Updated this week
- [NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.β19,294Updated last month
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β7,687Updated this week
- Build and share delightful machine learning apps, all in Python. π Star to support our work!β32,206Updated this week
- LAVIS - A One-stop Library for Language-Vision Intelligenceβ9,663Updated 3 weeks ago
- The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoiβ¦β46,708Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMsβ26,822Updated this week
- An open source implementation of CLIP.β9,782Updated last month
- RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the bestβ¦β12,397Updated 2 weeks ago
- Ongoing research training transformer models at scaleβ9,949Updated this week
- Making large AI models cheaper, faster and more accessibleβ38,614Updated this week
- The simplest, fastest repository for training/finetuning medium-sized GPTs.β36,216Updated last month
- Code and documentation to train Stanford's Alpaca models, and generate the data.β29,346Updated 2 months ago
- State-of-the-Art Text Embeddingsβ14,844Updated last week
- π€ Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.β132,113Updated this week
- This repository contains demos I made with the Transformers library by HuggingFace.β9,050Updated last month
- An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.β36,446Updated this week
- Train transformer language models with reinforcement learning.β9,288Updated this week
- A library for efficient similarity search and clustering of dense vectors.β30,511Updated this week
- Latest Advances on Multimodal Large Language Modelsβ11,722Updated this week
- Inference code for Llama modelsβ55,482Updated last month
- A playbook for systematically maximizing the performance of deep learning models.β26,385Updated 3 months ago
- Instruct-tune LLaMA on consumer hardwareβ18,537Updated last month