Hyeonwoo Noh | Bohyung Han | ||
POSTECH |
We propose a novel algorithm for visual question answering based on a recurrent deep neural network, where every module in the network corresponds to a complete answering unit with attention mechanism by itself. The network is optimized by minimizing loss aggregated from all the units, which share model parameters while receiving different information to compute attention probability. For training, our model attends to a region within image feature map, updates its memory based on the question and attended image feature, and answers the question based on its memory state. This procedure is performed to compute loss in each step. The motivation of this approach is our observation that multi-step inferences are often required to answer questions while each problem may have a unique desirable number of steps, which is difficult to identify in practice. Hence, we always make the first unit in the network solve problems, but allow it to learn the knowledge from the rest of units by backpropagation unless it degrades the model. To implement this idea, we early-stop training each unit as soon as it starts to overfit. Note that, since more complex models tend to overfit on easier questions quickly, the last answering unit in the unfolded recurrent neural network is typically killed first while the first one remains last. We make a single-step prediction for a new question using the shared model. This strategy works better than the other options within our framework since the selected model is trained effectively from all units without overfitting. The proposed algorithm achieves the state-of-the-art performance on the standard benchmark dataset without data augmentation.
The most interesting observation from our paper is that training recurrent answering units with multi-step joint loss minimization and earlystopping improves VQA accuracy of a single-step answering unit. This effect cannot be expected from the previous approaches based on multi-step attention based methods [1,2,3], because they use independent weight parameters for each attention steps. Interestingly, our single-step answering unit trained with proposed training method outperforms other approaches based on multi-step attention and prediction.
Image | Ours_FULL | Ours_SS | Question / Answer | Image | Ours_FULL | Ours_SS | Question / Answer |
Q: What color is the man's hat?
|
Q: Is it a rainy dat?
|
||||||
Q: What color is her hair?
|
Q: What is the pattern of the woman's dress?
|
||||||
Q: Can you see the man's hands?
|
Q: Which sign is the man showing?
|
||||||
Q: What is the girl walking with?
|
Q: What kind of bedspread is that?
|
The proposed algorithm outperforms other multi-step attention based models on VQA datasets[4].
Table 1: Comparison with other multi-step attention based models on VQA test-dev [4].
Open-Ended | Multiple-Choice | |||||||
All | Y/N | Num | Others | All | Y/N | Num | Others | |
SAN (VGG)[1] | 58.7 | 79.3 | 36.6 | 46.1 | - | - | - | - |
DMN (VGG) [2] | 60.3 | 80.5 | 36.8 | 48.3 | - | - | - | - |
HieCoAtt (VGG) [3] | 60.5 | 79.6 | 38.4 | 49.1 | 64.9 | 79.7 | 40.1 | 57.9 |
Ours_FULL (VGG) | 61.3 | 81.5 | 37.0 | 49.6 | 66.1 | 81.5 | 39.5 | 58.9 |
HieCoAtt (ResNet) [3] | 61.8 | 79.7 | 38.7 | 51.7 | 65.8 | 79.7 | 40.0 | 59.8 |
Ours_ResNet (ResNet) | 63.3 | 81.9 | 39.0 | 53.0 | 67.7 | 81.9 | 41.1 | 61.5 |
Check out GitHub repository: RAU_VQA GitHub Repository