Image Captioning论文合辑

Posted czhwust

tags:

篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了Image Captioning论文合辑相关的知识,希望对你有一定的参考价值。

 Image Caption

Automatically describing the content of an image

CV+NLP

数据集:Flickr8k,Flickr30k,MSCOCO,Visual Genome

评测指标:BLEU,METEOR,CIDEr,ROUGE

Learning to Evaluate Image Captioning(CVPR 2018)

技术分享图片

 

技术分享图片

 

Show and Tell: A Neural Image Caption Generator(CVPR 2015)

directly maximize the probability of the correct description given the image by using the following formulation:

技术分享图片

θ are the parameters of our model, I is an image, and S its correct transcription

技术分享图片

Encoder:Inception-V2

Decoder:LSTM
Inference:BeamSearch

 

Show, Attend and Tell: Neural Image Caption Generation with Visual Attention (ICML 2015)

Highlight:Attention Mechnism(Soft&Hard)

技术分享图片

 "Soft" attention:different parts,different subregions

"Hard" attention:only one subregion.Random choice

 

Sumary:

1.Attention involves focus of certain parts of input

2.Soft Attention is Deterministic.Hard attention is Stochastic.

3.Attention is used in NMT, AttnGAN, teaching machines to read.

Image Captioning with Semantic Attention(CVPR 2016)

 

SCA-CNN: Spatial and Channel-Wise Attention in Convolutional Networks for Image Captioning(CVPR 2017)

Highlight:Spatial and Channel-Wise Attention

技术分享图片

技术分享图片

Knowing When to Look: Adaptive Attention via a Visual Sentinel for Image Captioning(CVPR 2017)

Hightlight:Adaptive Attention

Semantic Compositional Networks for Visual Captioning(CVPR 2017)

技术分享图片技术分享图片

Deep Reinforcement Learning-based Image Captioning with Embedding Reward (CVPR 2017)

A decision-making framework for image captioning.

 A "policy network" and a "value network" to collaboratively generate captions.

 

 

 

Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering(CVPR 2018)

In the human visual system, attention can be focused volitionally by top-down signals determined by the current task(e.g.,looking for something), and automatically by bottom-up signals associated with unexpected, novel or salient stimuli.

top-down:attention mechanisms driven by non-visual or task-specific context; feature weights;

bottom-up:purely visual feed-forward attention mechanisms;based on Faster-RCNN proposes image regions (feature vector);

技术分享图片

 

技术分享图片

 

技术分享图片

 


以上是关于Image Captioning论文合辑的主要内容,如果未能解决你的问题,请参考以下文章

论文解读:Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering

Medical Image Report论文合辑

Paper Reading - Convolutional Image Captioning ( CVPR 2018 )

Paper Reading - CNN+CNN: Convolutional Decoders for Image Captioning

专知荟萃24视频描述生成(Video Captioning)知识资料全集(入门/进阶/论文/综述/代码/专家,附查看)

图像描述文献阅读Comprehending and Ordering Semantics for Image Captioning