Paper Reading - CNN+CNN: Convolutional Decoders for Image Captioning

Posted zlian2016

tags:

篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了Paper Reading - CNN+CNN: Convolutional Decoders for Image Captioning相关的知识,希望对你有一定的参考价值。

Link of the Paper: https://arxiv.org/abs/1805.09019

Innovations:

  • The authors propose a CNN + CNN framework for image captioning. There are four modules in the framework: vision module ( VGG-16 ), which is adopted to "watch" images; language module, which is to model sentences; attention module, which connects the vision module with the language module; prediction module, which takes the visual features from the attention module and concepts from the language module as input and predicts the next word.

技术分享图片    技术分享图片    技术分享图片

General Points:

  • RNNs or LSTMs cannot be calculated in parallel and ignore the underlying hierarchical structure of a sentence.
  • Directly feeding the output of the CNN into the RNN treats objects in an image the same and ignores the salient objects when generating one word.
  • In both m-RNN and NIC, an image is represented by a single vector, which ignores different areas and objects in the image. A spatial attention mechanism is introduced into image captioning model in Show, attend and tell: Neural image caption generation with visual attention, which allows the model to pay attention to different areas at each time step.

以上是关于Paper Reading - CNN+CNN: Convolutional Decoders for Image Captioning的主要内容,如果未能解决你的问题,请参考以下文章

Paper ReadingR-CNN(V5)论文解读

CNN卷积神经网络的改进(15年最新paper)

paper:A Deep Learning based Stock Trading Model with 2-D CNN Trend Detection 2D CNN模型

[Paper Weekly]CNN采样方法:空间变换网络(STN)与可变形卷积网络(DCN)

Faster R-CNN论文详解

基于CNN的图像修复(CNN-based Image Inpainting)