Please use this identifier to cite or link to this item: http://dr.iiserpune.ac.in:8080/xmlui/handle/123456789/7139
Full metadata record
DC FieldValueLanguage
dc.contributor.authorKhare, Yash-
dc.contributor.authorBAGAL, VIRAJ-
dc.contributor.authorMathew, Minesh-
dc.contributor.authorDevi, Adithi-
dc.contributor.authorPriyakumar, U Deva-
dc.contributor.authorJawahar, C.V.-
dc.coverage.spatialNice, Franceen_US
dc.date.accessioned2022-06-21T05:17:00Z-
dc.date.available2022-06-21T05:17:00Z-
dc.date.issued2021-05-
dc.identifier.citation2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI).en_US
dc.identifier.urihttps://ieeexplore.ieee.org/document/9434063/authorsen_US
dc.identifier.urihttp://dr.iiserpune.ac.in:8080/xmlui/handle/123456789/7139-
dc.description.abstractImages in the medical domain are fundamentally different from the general domain images. Consequently, it is infeasible to directly employ general domain Visual Question Answering (VQA) models for the medical domain. Additionally, medical image annotation is a costly and time-consuming process. To overcome these limitations, we propose a solution inspired by self-supervised pretraining of Transformer-style architectures for NLP, Vision, and Language tasks. Our method involves learning richer medical image and text semantic representations using Masked Vision-Language Modeling as the pretext task on a large medical image + caption dataset. The proposed solution achieves new state-of-the-art performance on two VQA datasets for radiology images - VQA-Med 2019 and VQA-RAD, outperforming even the ensemble models of previous best solutions. Moreover, our solution provides attention maps which help in model interpretability.en_US
dc.language.isoenen_US
dc.publisherIEEEen_US
dc.subjectChemistryen_US
dc.subject2021en_US
dc.titleMMBERT: Multimodal BERT Pretraining for Improved Medical VQAen_US
dc.typeConference Papersen_US
dc.contributor.departmentDept. of Chemistryen_US
dc.identifier.doihttps://doi.org/10.1109/ISBI48211.2021.9434063en_US
dc.publication.originofpublisherForeignen_US
Appears in Collections:CONFERENCE PAPERS

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.