ChanganVR / action2soundLinks
Action2Sound: Ambient-Aware Generation of Action Sounds from Egocentric Videos
☆25Updated last year
Alternatives and similar repositories for action2sound
Users that are interested in action2sound are comparing it to the libraries listed below
Sorting:
- Offical code for the CVPR 2024 Paper: Separating the "Chirp" from the "Chat": Self-supervised Visual Grounding of Sound and Language☆86Updated last year
- The official implementation of V-AURA: Temporally Aligned Audio for Video with Autoregression (ICASSP 2025) (Oral)☆32Updated last year
- Official PyTorch implementation of "Conditional Generation of Audio from Video via Foley Analogies".☆93Updated 2 years ago
- Diff-Foley: Synchronized Video-to-Audio Synthesis with Latent Diffusion Models☆200Updated last year
- [🏆 IJCV 2025 & ACCV 2024 Best Paper Honorable Mention] Official pytorch implementation of the paper "High-Quality Visually-Guided Sound …☆27Updated 2 months ago
- Source code for "Synchformer: Efficient Synchronization from Sparse Cues" (ICASSP 2024)☆104Updated 4 months ago
- Ego4DSounds: A diverse egocentric dataset with high action-audio correspondence☆19Updated last year
- ☆47Updated 9 months ago
- This repo contains the official PyTorch implementation of AudioToken: Adaptation of Text-Conditioned Diffusion Models for Audio-to-Image …☆88Updated last year
- ☆48Updated last year
- Official source codes for the paper: EmoDubber: Towards High Quality and Emotion Controllable Movie Dubbing.☆34Updated 7 months ago
- EchoX: Towards Mitigating Acoustic-Semantic Gap via Echo Training for Speech-to-Speech LLMs☆46Updated 4 months ago
- [ECCV 2024 Oral] Audio-Synchronized Visual Animation☆57Updated last year
- a text-conditional diffusion probabilistic model capable of generating high fidelity audio.☆188Updated last year
- The official implementation of the IJCAI 2024 paper "MusicMagus: Zero-Shot Text-to-Music Editing via Diffusion Models".☆47Updated last year
- Official codes and models of the paper "Auffusion: Leveraging the Power of Diffusion and Large Language Models for Text-to-Audio Generati…☆193Updated last year
- Make-An-Audio-3: Transforming Text/Video into Audio via Flow-based Large Diffusion Transformers☆118Updated 8 months ago
- ☆62Updated 7 months ago
- official code for CVPR'24 paper Diff-BGM☆72Updated last year
- [ICML2023] Long-Term Rhythmic Video Soundtracker☆61Updated 6 months ago
- [NeurIPS 2024] Code, Dataset, Samples for the VATT paper “ Tell What You Hear From What You See - Video to Audio Generation Through Text”☆35Updated 6 months ago
- [CVPR 2024] Seeing and Hearing: Open-domain Visual-Audio Generation with Diffusion Latent Aligners☆155Updated last year
- ☆59Updated last year
- Official code for the paper "Understanding Co-speech Gestures in-the-wild"☆20Updated 2 months ago
- ☆21Updated 3 years ago
- ☆17Updated 2 years ago
- Towards Fine-grained Audio Captioning with Multimodal Contextual Cues☆86Updated 3 weeks ago
- [Interspeech 2023] Intelligible Lip-to-Speech Synthesis with Speech Units☆47Updated last year
- Diff-TTSG: Denoising probabilistic integrated speech and gesture synthesis☆40Updated 2 years ago
- Kling-Foley: Multimodal Diffusion Transformer for High-Quality Video-to-Audio Generation☆63Updated 7 months ago