Skip to main content

Image2Text Dashboard

Dashboard for Experiments
Created on March 2|Last edited on April 12

Training πŸ› 

Data πŸ’Ύ

The data input for training is
  • an image whose features are extracted via InceptionV3.
  • caption which is converted into a word vector with index = <position in sentence> and value = <id of word as defined by the constructed vocabulary>
We show here the original images alongside the caption and the associated relative file name as defined in our artifact structure
For ✨flair✨ I added an interactive way to view the images based on the word vectors of the caption (may not make the most sense but it's fun!)

ο»Ώ
ο»Ώ
name.caption
image
1
COCO_train2014_000000321666.jpg
2
COCO_train2014_000000460039.jpg
name.name
ο»Ώ

Losses πŸ“‰

ο»Ώ
2k4k6kStep0.40.60.81
02k4k6kStep20406080100
Run: train-coco2014-attention-model
1
ο»Ώ
ο»Ώ

Gradients 🎱

Evaluation πŸ“Έβ†’βœπŸ½

Predictions πŸ™ŒπŸ½

For each prediction, we also include the attention plot to allow us to better interpret why the model may be making the predictions it made. Take a look! πŸ”
πŸ’‘
Logging this as a table makes it very easy for the βœπŸ½β†’πŸ—£ team to pull and run their inference on!

ο»Ώ
ο»Ώ
image
real_caption
predicted_caption
attention_plot
1
COCO_train2014_000000397526.jpg
2
COCO_train2014_000000480861.jpg
name
File<(table)>
File<(table)>
File<(table)>