microsoft / unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
ā20,584Updated last week
Alternatives and similar repositories for unilm:
Users that are interested in unilm are comparing it to the libraries listed below
- Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"ā11,119Updated last month
- Fast and memory-efficient exact attentionā15,064Updated this week
- š¤ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.ā16,978Updated this week
- LAVIS - A One-stop Library for Language-Vision Intelligenceā10,161Updated last month
- š A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iā¦ā8,178Updated this week
- Ongoing research training transformer models at scaleā11,109Updated this week
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.ā36,255Updated this week
- CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an imageā26,971Updated 5 months ago
- š¤ Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.ā137,641Updated this week
- RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable)ā¦ā13,005Updated last week
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.ā30,809Updated last week
- Hackable and optimized Transformers building blocks, supporting a composable construction.ā8,910Updated this week
- Train transformer language models with reinforcement learning.ā10,609Updated this week
- An open source implementation of CLIP.ā10,804Updated last week
- [NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.ā21,096Updated 5 months ago
- State-of-the-Art Text Embeddingsā15,772Updated last week
- Build and share delightful machine learning apps, all in Python. š Star to support our work!ā35,268Updated this week
- Accessible large language models via k-bit quantization for PyTorch.ā6,522Updated this week
- A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) trainingā20,810Updated 5 months ago
- ImageBind One Embedding Space to Bind Them Allā8,476Updated 5 months ago
- This repository contains demos I made with the Transformers library by HuggingFace.ā9,799Updated this week
- Unsupervised text tokenizer for Neural Network-based text generation.ā10,479Updated last month
- BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)ā7,075Updated last year
- š¤ Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch and FLAX.ā27,123Updated this week
- The simplest, fastest repository for training/finetuning medium-sized GPTs.ā38,486Updated last month
- A concise but complete full-attention transformer with a set of promising experimental features from various papersā4,985Updated last week
- QLoRA: Efficient Finetuning of Quantized LLMsā10,168Updated 7 months ago
- The standard data-centric AI package for data quality and machine learning with messy, real-world data and labels.ā10,056Updated last week
- Foundation Architecture for (M)LLMsā3,038Updated 9 months ago
- Latest Advances on Multimodal Large Language Modelsā13,556Updated this week