Hassaan-Elahi / Dense-Captioning
Dense Captioning is a system of fully localized Deep Convolutional Neural networks to translate a video into natural language. It uses CNN (VGG16) for feature extraction from video and encoder-decoder models (LSTM & GRU) to generate descriptions utilizing transfer-learning approach.
☆11Updated 6 years ago
Alternatives and similar repositories for Dense-Captioning:
Users that are interested in Dense-Captioning are comparing it to the libraries listed below
- Word2VisualVec : Predicting Visual Features from Text for Image and Video Caption Retrieval☆69Updated 5 years ago
- Image Caption using keras, VGG16 pretrained model, CNN and RNN☆44Updated 5 years ago
- Chinese image to caption, based on VGG + LSTM + ATTENTION☆10Updated 6 years ago
- Novel Object Captioner - Captioning Images with diverse objects☆41Updated 7 years ago
- A reimplementation of Show and Tell☆15Updated 6 years ago
- Image Captioning with Keras☆63Updated 4 years ago
- A video captioning tool using S2VT method and attention mechanism (TensorFlow)☆15Updated 6 years ago
- 🎬 Video Captioning: ICCV '15 paper implementation☆47Updated 6 years ago
- Code and demos for our paper at ACM MM 2017☆62Updated 5 years ago
- Official Tensorflow Implementation of the paper "Bidirectional Attentive Fusion with Context Gating for Dense Video Captioning" in CVPR 2…☆149Updated 5 years ago
- Video Summarization (Attention Mechanism and Hierarchical LSTM)☆29Updated 7 years ago
- Course Project for CS771: Machine Learning☆25Updated 8 years ago
- A PyTorch reimplementation of FCSN in paper "Video Summarization Using Fully Convolutional Sequence Networks"☆117Updated last year
- TensorFlow Implementation of "Attention Clusters: Purely Attention Based Local Feature Integration for Video Classification".☆40Updated 6 years ago
- Codes for paper of "Attention-based LSTM with Semantic Consistency for Videos Captioning "☆19Updated 8 years ago
- Repository for image caption for Chinese☆28Updated 7 years ago
- [ACM MM 2017 & IEEE TMM 2020] This is the Theano code for the paper "Video Description with Spatial Temporal Attention"☆57Updated 4 years ago
- Pytorch implementation of audio-visual fusion video captioning model☆27Updated 6 years ago
- AAAI 2018 - Unsupervised video summarization with deep reinforcement learning (PyTorch)☆46Updated 5 years ago
- Image captioning models in PyTorch☆37Updated 4 years ago
- A PyTorch implementation of VSumPtrGAN☆39Updated last year
- PyTorch implementation of Chinese image captioning on AI_challenger dataset☆34Updated 5 years ago
- A deep learning model utilizing CNN and LSTM to recognize activity from video. (To be used for bench-marking hardware accelerator)☆8Updated 5 years ago
- AAAI 2018 - Unsupervised video summarization with deep reinforcement learning (Theano)☆138Updated 3 years ago
- ☆17Updated 7 years ago
- ☆33Updated 7 years ago
- A Pytorch implementation of "describing videos by exploiting temporal structure", ICCV 2015☆48Updated 2 years ago
- ☆27Updated 4 years ago
- Image Captioning based on Bottom-Up and Top-Down Attention model☆102Updated 6 years ago
- Source code for the paper "Speaking the Same Language: Matching Machine to Human Captions by Adversarial Training"☆66Updated 6 years ago