Skip to content
Zhengyuan Zhu
Go back

Demo-Driven Learning: Image Caption

Introduction to demo

Source Code:image_captioning_with_attention

Show, Attend and Tell: Neural Image Caption Generation with Visual Attention.

Goal of this end2end model

Dateset

MS-COCO:This dataset contains >82,000 images, each of which has been annotated with at least 5 different captions.

Frame work of demo:

Preprocess the images using InceptionV3: extract features from the last convolutional layer.

Preprocess and tokenize the captions

create a tf.data dataset to use for training our model.

Model

Training

Caption

Problems undesirable

Version

GPU lose connect

Reference


Share this post on:

Previous Post
Paper Note: A Diversity-Promoting Objective Function for Neural Conversation Models
Next Post
Deep Learning Module Documentation Memo
Jack the orange tabby cat
I'm Jack 🧡
Luna the tuxedo cat
I'm Luna! 🖤