IMAGE CAPTIONING USING PHRASE-BASED HIERARCHICAL LSTM MODEL

Similar documents

Recurrent neural network grammars. Slide credits: Chris Dyer, Adhiguna Kuncoro

Genera&on of Image Descrip&ons. Tambet Ma&isen

[Boston March for Science 2017 photo Hendrik Strobelt]

Week 42: Siamese Network: Architecture and Applications in Visual Object Tracking. Yuanwei Wu

CSE 408 Multimedia Info Sys.

Reasoning with Neural Networks

Moving toward formalisation COMP62342

CS6501: Deep Learning for Visual Recognition. CNN Architectures

Second Interna,onal Workshop on Parts and A5ributes ECCV 2012, Firenze, Italy October, 2012 Discovering a Lexicon of Parts and Attributes

Sentences and pictures: not just more words and pictures

Subdomain Entry Vocabulary Modules Evaluation

The online processing of semantic and pragmatic content

Semantics. These slides were produced by Hadas Kotek.

X-bar Node Flavors Introduction to syntax. Noun Phrase

Moving towards formalisation COMP62342

What kind of Theory do we need for English Syntax? Are languages finite? Could we list all the sentences of English?

Development of intelligent systems (RInS) Object recognition with Convolutional Neural Networks

Long, blue, spiky-edged shadows crept out across the snowfields, while a rosy glow, at first scarce discernible, gradually deepened and suffused

Natural Language Processing (NLP)

Dynamic Programming for Linear Time Incremental Parsing

Chapter 6: Extending Theory

The integration of dogs into collaborative humanrobot. - An applied ethological approach - PhD Thesis. Linda Gerencsér Supervisor: Ádám Miklósi

Grade 5 English Language Arts

Read Brown Bear, Brown Bear, What Do You See? Read the book and talk about all the animals!

Semantically-driven Automatic Creation of Training Sets for Object Recognition

On Deriving Aspectual Sense

STUDY BEHAVIOR OF CERTAIN PARAMETERS AFFECTING ASSESSMENT OF THE QUALITY OF QUAIL EGGS BY COMPUTER VISION SYSTEM

A few applications of natural language processing

CONNECTION TO LITERATURE

An Introduction to Formal Logic

BioSci 110, Fall 08 Exam 2

Graphics libraries, PCS Symbols, Animations and Clicker 5

FIRST TERM READING REVISION PAPER ENGLISH LANGUAGE GRADE 3

FPGA-based Emotional Behavior Design for Pet Robot

Hunting Zika Virus using Machine Learning

ACCUPLACER Sample Questions for Students

The Kaggle Competitions: An Introduction to CAMCOS Fall 2015

Entailment above the word level in distributional semantics

Noun Underline each noun in the sentences below. Then, write C above the noun if it is a common noun, or P if it is a proper noun.

ESL Writing & Computerized Accuplacer ESL (Reading, Listening, Language Use)

Biology 164 Laboratory

Bayesian Analysis of Population Mixture and Admixture

ENGLISH LANGUAGE GRADE 3 TERM END READING REVISION

Chapter 6: Extending Theory

Grade 3, Prompt for Opinion Writing

The Distorting Mirror

Caillou and Gilbert Written by Joceline Sanschagrin Illustrated by Cinar Animation

My Journey To Penang, Malaysia

Characteristics of the Text Genre Fantasy Text Structure Simple fi rst-person narrative, with story carried by pictures Content

Design of 32 bit Parallel Prefix Adders

Activity 1: Changes in beak size populations in low precipitation

Level 11. Book g. Level 11. Word Count 210 Text Type Information report High Frequency Word/s Introduced. The Snail Race Outside Games

My world. 1º, 2º, 3º Primary

INSPIRE A WRITING REVOLUTION! /

I will learn to talk about. groups of animals animal characteristics animal habitats. Unit Unit 7

Lacey Blocker Vernon Parish Teacher Leader NBCT

A Dog s Life. Unit 7. Speaking. Vocabulary - Dogs. Dog breeds: poodle husky German shepherd Labrador Yorkshire terrier

Lesson Objectives. Core Content Objectives. Language Arts Objectives

Expanded noun phrases and verbs to describe an underwater world

Results for: HABIBI 30 MARCH 2017

Prof Michael O Neill Introduction to Evolutionary Computation

Accounting for the causal link between free adjuncts and their host clauses

Call of the Wild. Investigating Predator/Prey Relationships

3-35. A House for Hermit Crab

Jefferson County High School Course Syllabus

Discussion and Activity Guide for. Orville: A Dog Story Written by Haven Kimmel, illustrated by Robert Andrew Parker

Boosting Biomedical Entity Extraction by Using Syntactic Patterns for Semantic Relation Discovery

Where Is My Puppy? Retrieving Lost Dogs by Facial Features

Grade 5, Prompt for Opinion Writing Common Core Standard W.CCR.1

Caring and. sharing. We love Hong Kong. 2 Small houses News report. 3 Food in a basin Fun and games Description. 4 Computer Jobs Biography

8A READ-ALOUD. How Turtle Cracked His Shell. Lesson Objectives. Language Arts Objectives. Core Vocabulary

Performance Task: Lizards, Lizards, Everywhere!

Teacher Guide Teacher Answer Key and Kentucky Core Academic Standards for RPA 1 Grade 3

Teacher s Notes. Level 3. Did you know? Pearson English Kids Readers. Teacher s Notes. Summary of the story. Background information

Attributing the Bixby Letter: A case of historical disputed authorship

Perplexity of n-gram and dependency language models

Have something to say, and say it as clearly as you can. That is the only secret of style.

Grade 2 English Language Arts

Following the huge success of our first series of Brilliant Books, we are very proud to present MORE BRILLIANT BOOKS!

SUBJECT, SUBJECT + PREDICATE, PREDICATE USING COMPOUND SUBJECTS AND PREDICATES

VISUALIZING TEXT. Petra Isenberg

VISUALIZING TEXT. Petra Isenberg

A Creature Went Walking A Lesson for Gr. 4-6

TABLE OF CONTENTS. 4. VIP PETCARE COLORS a. Core Colors b. Accent Colors. 5. VIP PETCARE FONTS a. Font Guidelines

اإلجابت على الورقت وفسها

Lab 7. Evolution Lab. Name: General Introduction:

Kentucky Academic Standards

English 11H Mrs. V. Pechstein

Design of Low Power and High Speed Carry Select Adder Using Brent Kung Adder

Let s Talk Turkey Selection Let s Talk Turkey Expository Thinking Guide Color-Coded Expository Thinking Guide and Summary

Sample Paper HEC CAT A

Cat Swarm Optimization

Biol 160: Lab 7. Modeling Evolution

Figure 1 Background Information to the phylum Arthropoda and appears to not have changed for

Saint Bernards. and Other Working Dogs. by Holly Schroeder illustrated by Troy Howell. Scott Foresman Reading Street 2.2.5

The weekly passage discussed issues related to dog ownership. Here is some information that might be helpful to students less familiar the topic.

Trapped in a Sea Turtle Nest

Theme and Rheme of Main Character Script in Hachiko Movie. *Tohom Marthin Donius Pasaribu and ** Sumarsih The State University of Medan (UNIMED)

@DEVONPERSING DESIGNING FOR ACCESSIBILITY

Transcription:

IMAGE CAPTIONING USING PHRASE-BASED HIERARCHICAL LSTM MODEL 1 Chee Seng Chan PhD SMIEEE 23 October 2017 Nvidia AI Conference, Singapore email: cs.chan@um.edu.my

INTRODUCTION Aim: Automatic generate a full sentence describing an image. Motivated by the significant progress of image classification and statistical language model. Applications: Early childhood educations Scene understanding for the visual impairments Image retrievals Two children are playing on a swing made out of a tire. 2

BACKGROUNDS Processing of Image, I: Represented as a vector using feature learning algorithm such as convolutional neural network (CNN) Processing of Language: Each sentence is equivalent to a sequence of words. A statistical model is trained to predict the conditional probability of next word given all previous words Multimodal Embedding Prediction of next word also conditioned on image 3

BACKGROUNDS Sequence is learned with Recurrent Neural Network (RNN). The most popular variant of RNN is Long Short-Term Memory (LSTM). 4

PROBLEM STATEMENT Conventional models treat a sentence as a sequence of words. All other linguistic syntax and structure are disregarded. Sentence structure is one of the most prominent characteristic of sentence! Two dogs are running in the snow. NP VP PP NP NP = noun phrase VP = verb phrase PP = prepositional phrase 5

PROBLEM STATEMENT Quoted on Victor Yngve [14] (an influential contributor in linguistic theory): language structure involving, in some form or other, a phrase structure hierarchy, or immediate constituent organization Example: S VP VP PP NP NP the dogs are running in the snow Phrase structure grammar ROOT running aux nmod:in are nsubj snow case dogs det det in the the Dependency grammar 6

RESEARCH INTEREST & OBJECTIVE Is it really okay to treat sentence as only sequence of words, while disregarding any other important characteristic of sentence such as structure? 1. Design of phrase-based model for image captioning. This is one of the most earliest work after PbIC[13]. 2. Investigate on its performance as compared to a pure sequence model. 7

DESIGN MOTIVATION A young girl wearing a yellow shirt with a blue backpack is walking next to a fence covered with a blue plastic cover. Noun phrases form most of an image caption. They have similar syntactic role They have strong relation with the image. 8

CONVENTIONAL VS. PROPOSAL Sentence: A motorcyclist on the street. conventional proposal 9

RELATED WORKS Methods Details (Red words are their cons) References Template based Generate sentence from a fix template. Sentence generated is rigid. Composition Method Stitch up image relevant phrases to form a sentence. Computational cost is high. Neural Network Trained to predict sequence. Only model words sequence. 1-4 5-7 mrnn [8], NIC [9], DeepVS [10], LCRNN [12] The closest work is Phrase based Image Captioning PbIC[13] proposed by Lebret et al. They encode each sentence as phrase sequence only while my proposal is to encode as sequence of phrase and words. They use simpler model. 10

PROPOSED MODEL Training Data: image sentence pair Phrase Chunking Encode Image & Phrases Encode Image & Sentence Training Generate Caption 11

PROPOSED MODEL: 1) PHRASE CHUNKING Approach to identify the constituents of a sentence. Extract only noun phrase prominent in image description Dependency parse * with selected relations: det determiner (e.g.: a man ) amod - adjective modifier (e.g.: green shirt ) nummod - numeric modifier (e.g.: two dogs ) compound - compound (e.g.: basketball court ) advmod - adverbial modifier, when modifying meaning of adjective (e.g.: dimly lit room ) nmod:of & nmod:poss - nominal modifier for possessive alteration (e.g.: his hand ) 12 *Stanford CoreNLP Software - https://stanfordnlp.github.io/corenlp/

PROPOSED MODEL: 1) PHRASE CHUNKING Chunking from dependency parse 13

PROPOSED MODEL: 2) COMPOSITIONAL VECTOR OF PHRASE Our proposed architecture is the hierarchical counterpart of NIC model proposed by Vinyals et al [9] 14 Phrases: the man, the gray shirt, sandals, the large tricycle

PROPOSED MODEL: 3) SENTENCE ENCODING Sentence: The man in the gray shirt and sandals is pulling the large tricycle. A phrase token is added into the corpus for prediction 15

TRAINING Objective function: Perplexity: j / M = index / total no of training sentence pt p / pt s = probability distribution over words on the particular time step for phrase / sentence t p / P = time step / total no. of time step in phrase t s / Q = time step / total no. of time step in sentence i / R = index / total no. of phrase in sentence I 16

TRAINING PHRASE SELECTION OBJECTIVE Objective function: Cost of phrase selection objective: = trainable parameters = hidden output at t s for input k = label of input k at t s = normalizing constant based on = index / total no of inputs at t s = set of t s which the input is phrase 17

GRAPHICAL ILLUSTRATION: SENTENCE GENERATION (PHRASE LEVEL) CNN & Img Embedding LSTM #START# LSTM a the two its three LSTM dogs snow brown beach dog LSTM dog LSTM Selected Phrases: a two brown dogs dog the two snow dogs the beach snow the a beach dog a dog pt pt pt pt a the two its three K candidates atwo brown dogs the dog snow a black brown the large beach red a dog the snow beach two dogs brown its three K candidates two dogs (#END#) the snow (#END#) a brown dog the beach (#END#) a dog (#END#) K 2 candidates a brown dog (#END#) 18

GRAPHICAL ILLUSTRATION: SENTENCE GENERATION (SENTENCE LEVEL) #START# a brown dog two dogs the snow the beach a dog dog a there brown CNN & Img Embedding LSTM LSTM LSTM Two dogs play in the snow. pt pt #PHRASE# dog a there brown 1 1 0 0 1 a brown dog is a dog runs two dogs are a dog is a brown dog runs K Selected Phrases: a brown dog two dogs the snow the beach a dog 19

EXPERIMENT Tested on Flickr8k and Flickr30k datasets. Each image is annotated with five descriptions by human. 1k of images are used for validation and another 1k of images are used for testing, while the rest are for training (consistent with state-of-the-art). A woman in a red coat with a man in a white and black coat and a black dog in the snow. Two people and a dog are in the snow. Two people are interacting with a dog that has bitten an object one of them is holding. Two people are walking up a snowy hill with a dog. Two people playing on a snowy hill. 20

QUALITATIVE RESULTS (PHRASE) Phrase generation: 21

QUALITATIVE RESULTS (SENTENCE) (baseline) (proposed) (human) (baseline) (proposed) 22 (human)

MORE RESULTS (SENTENCES WITH SAME OBJECT(S)) 23

MORE RESULTS (SENTENCES WITH SAME SCENE) 24

QUALITATIVE RESULTS (POOR EXAMPLES) 25

QUANTITATIVE RESULTS Evaluation metric: BLEU Measure n-grams precision quality between generated caption and reference sentences (human). Our proposed model Our proposed model 26

MORE ANALYSIS BY COMPARING WITH BASELINE Given same amount of training data, and same set of test image, and same set of setting in training: Our model can generate sentence formed with more variety of words in the training corpus. What is the minimum time a word should appears in training data, so the model can generate sentence using that word? Our model (phi-lstm) = 81 Baseline (NIC) = 93 27

CONCLUSION Proposed of hierarchical phrase-based LSTM model to generate image description. Hierarchical model vs pure sequential model: Able to generate better description Can learn with less data Published in ACCV 2016, extension to journal. Future works Experiments on MSCOCO dataset Evaluation on more types of automatic evaluation metrics such as ROUGE, METEOR, CIDEr Apply on image sentence bi-directional retrieval Tackle problem in poor results 28

REFERENCES 1. Farhadi, A., Hejrati, M., Sadeghi, M.A., Young, P., Rashtchian, C., Hockenmaier, J., Forsyth, D.: Every picture tells a story: Generating sentences from images. In: ECCV 2010. 2. Kulkarni, G., Premraj, V., Dhar, S., Li, S., Choi, Y., Berg, A.C., Berg, T.L.: Baby talk: Understanding and generating image descriptions. In: CVPR 2011. 3. Yang, Y., Teo, C.L., Daum e III, H., Aloimonos, Y.: Corpus-guided sentence generation of natural images. In: EMNLP 2011. 4. Mitchell, M., Han, X., Dodge, J., Mensch, A., Goyal, A., Berg, A., Yamaguchi, K., Berg, T., Stratos, K., Daum e III, H.: Midge: Generating image descriptions from computer vision detections. In: EACL 2012. 5. Kuznetsova, P., Ordonez, V., Berg, T.L., Choi, Y.: Treetalk: Composition and compression of trees for image descriptions. TACL 2014. 29

REFERENCES 6. Li, S., Kulkarni, G., Berg, T.L., Berg, A.C., Choi, Y.: Composing simple image descriptions using web-scale n-grams. In: CONLL 2011. 7. Kuznetsova, P., Ordonez, V., Berg, A.C., Berg, T.L., Choi, Y.: Collective generation of natural image descriptions. In: ACL 2012. 8. Mao, J., Xu, W., Yang, Y., Wang, J., Huang, Z., Yuille, A.: Deep captioning with multimodal recurrent neural networks (m-rnn). In: ICLR 2015. 9. Vinyals, O., Toshev, A., Bengio, S., Erhan, D.: Show and tell: A neural image caption generator. In: CVPR 2015. 10. Karpathy, A., Fei-Fei, L.: Deep visual-semantic alignments for generating image descriptions. In: CVPR 2015. 30

REFERENCES 11. Kiros, R., Salakhutdinov, R., Zemel, R.S.: Unifying visual-semantic embeddings with multimodal neural language models. arxiv preprint arxiv:1411.2539 (2014) 12. Donahue, J., Anne Hendricks, L., Guadarrama, S., Rohrbach, M., Venugopalan, S., Saenko, K., Darrell, T.: Long-term recurrent convolutional networks for visual recognition and description. In: CVPR 2015. 13. Lebret, R., Pinheiro, P.O., Collobert, R.: Phrase-based image captioning. ICML 2015 14. Yngve, V.: A model and an hypothesis for language structure. Proceedings of the American Philosophical Society 104 (1960) 444 466 31

THE END Q & A? Chee Seng Chan PhD SMIEEE University of Malaya, Malaysia www.cs-chan.com Full Paper: Tan, Y. H., & Chan, C. S. (2016, November). phi-lstm: A phrase-based hierarchical LSTM model for image captioning. In Asian Conference on Computer Vision (ACCV), pp. 101-117.