Publication: Vision Matters When It Should: Sanity Checking Multimodal Machine Translation Models
Vision Matters When It Should: Sanity Checking Multimodal Machine Translation Models
Date
Date
Date
Citations
Li, J., Ataman, D., & Sennrich, R. (2021). Vision Matters When It Should: Sanity Checking Multimodal Machine Translation Models. 8556–8562. https://aclanthology.org/2021.emnlp-main.673
Abstract
Abstract
Abstract
Multimodal machine translation (MMT) systems have been shown to outperform their text-only neural machine translation (NMT) counterparts when visual context is available. However, recent studies have also shown that the performance of MMT models is only marginally impacted when the associated image is replaced with an unrelated image or noise, which suggests that the visual context might not be exploited by the model at all. We hypothesize that this might be caused by the nature of the commonly used evaluation benchmark, also known as
Additional indexing
Creators (Authors)
Event Title
Event Title
Event Title
Event Location
Event Location
Event Location
Event Country
Event Country
Event Country
Event Start Date
Event Start Date
Event Start Date
Event End Date
Event End Date
Event End Date
Publisher
Publisher
Publisher
Page range/Item number
Page range/Item number
Page range/Item number
Page end
Page end
Page end
Item Type
Item Type
Item Type
In collections
Dewey Decimal Classifikation
Dewey Decimal Classifikation
Dewey Decimal Classifikation
Language
Language
Language
Date available
Date available
Date available
OA Status
OA Status
OA Status
Free Access at
Free Access at
Free Access at
Citations
Li, J., Ataman, D., & Sennrich, R. (2021). Vision Matters When It Should: Sanity Checking Multimodal Machine Translation Models. 8556–8562. https://aclanthology.org/2021.emnlp-main.673