Encoder-Decoder Architecture: I utilized transfer learning to generate image features by feeding the images into a pre-trained Inception-v3 model, then caption sequences were generated by inputting text into a word embedding layer and feeding the resultant output into a subsequent long short-term memory (LSTM) layer Data: Flicker8k dataset https://www.kaggle.com/hsankesara/flickr-image-dataset
-
Notifications
You must be signed in to change notification settings - Fork 0
quynhanh12345/ImageCaptioning
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
About
No description, website, or topics provided.
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published