haoliuhl / language-quantized-autoencoders
Language Quantized AutoEncoders
β102Updated 2 years ago
Alternatives and similar repositories for language-quantized-autoencoders:
Users that are interested in language-quantized-autoencoders are comparing it to the libraries listed below
- β127Updated 2 years ago
- Implementation of π» Mirasol, SOTA Multimodal Autoregressive model out of Google Deepmind, in Pytorchβ88Updated last year
- https://arxiv.org/abs/2209.15162β49Updated 2 years ago
- Code for paper "Diffusion Language Models Can Perform Many Tasks with Scaling and Instruction-Finetuning"β71Updated last year
- β91Updated last year
- PyTorch codes for the paper "An Empirical Study of Multimodal Model Merging"β38Updated last year
- The official code for paper "EasyGen: Easing Multimodal Generation with a Bidirectional Conditional Diffusion Model and LLMs"β73Updated 4 months ago
- [Arxiv] Aligning Modalities in Vision Large Language Models via Preference Fine-tuningβ80Updated 10 months ago
- β117Updated 2 years ago
- Sparkles: Unlocking Chats Across Multiple Images for Multimodal Instruction-Following Modelsβ43Updated 9 months ago
- Touchstone: Evaluating Vision-Language Models by Language Modelsβ82Updated last year
- β48Updated last year
- Official code for "pi-Tuning: Transferring Multimodal Foundation Models with Optimal Multi-task Interpolation", ICML 2023.β32Updated last year
- Holistic evaluation of multimodal foundation modelsβ43Updated 7 months ago
- β49Updated last year
- Online Adaptation of Language Models with a Memory of Amortized Contexts (NeurIPS 2024)β63Updated 7 months ago
- β126Updated last year
- Mind the Gap: Understanding the Modality Gap in Multi-modal Contrastive Representation Learningβ148Updated 2 years ago
- Implementation of Soft MoE, proposed by Brain's Vision team, in Pytorchβ270Updated 11 months ago
- Patching open-vocabulary models by interpolating weightsβ91Updated last year
- This repo contains evaluation code for the paper "BLINK: Multimodal Large Language Models Can See but Not Perceive". https://arxiv.orβ¦β115Updated 8 months ago
- M4 experiment logbookβ57Updated last year
- β71Updated 8 months ago
- [NeurIPS 2023] Make Your Pre-trained Model Reversible: From Parameter to Memory Efficient Fine-Tuningβ31Updated last year
- [ICLR 2023] Official implementation of Transnormer in our ICLR 2023 paper - Toeplitz Neural Network for Sequence Modelingβ79Updated 11 months ago
- Randomized Positional Encodings Boost Length Generalization of Transformersβ80Updated last year
- ChatBridge, an approach to learning a unified multimodal model to interpret, correlate, and reason about various modalities without relyβ¦β50Updated last year
- Toolkit for Elevater Benchmarkβ70Updated last year
- Implementation of Zorro, Masked Multimodal Transformer, in Pytorchβ97Updated last year
- Source code for the paper "Prefix Language Models are Unified Modal Learners"β43Updated last year