TY - GEN
T1 - Multi-Modal Learning Using Physicians Diagnostics for Optical Coherence Tomography Classification
AU - Logan, Yash Yee
AU - Kokilepersaud, Kiran
AU - Kwon, Gukyeong
AU - Alregib, Ghassan
AU - Wykoff, Charles
AU - Yu, Hannah
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - In this paper, we propose a framework that incorporates experts diagnostics and insights into the analysis of Optical Coherence Tomography (OCT) using multi-modal learning. To demonstrate the effectiveness of this approach, we create a medical diagnostic attribute dataset to improve disease classification using OCT. Although there have been successful attempts to deploy machine learning for disease classification in OCT, such methodologies lack the experts insights. We argue that injecting ophthalmological assessments as another supervision in a learning framework is of great importance for the machine learning process to perform accurate and interpretable classification. We demonstrate the proposed framework through comprehensive experiments that compare the effectiveness of combining diagnostic attribute features with latent visual representations and show that they surpass the state-of-the-art approach. Finally, we analyze the proposed dual-stream architecture and provide an insight that determine the components that contribute most to classification performance.
AB - In this paper, we propose a framework that incorporates experts diagnostics and insights into the analysis of Optical Coherence Tomography (OCT) using multi-modal learning. To demonstrate the effectiveness of this approach, we create a medical diagnostic attribute dataset to improve disease classification using OCT. Although there have been successful attempts to deploy machine learning for disease classification in OCT, such methodologies lack the experts insights. We argue that injecting ophthalmological assessments as another supervision in a learning framework is of great importance for the machine learning process to perform accurate and interpretable classification. We demonstrate the proposed framework through comprehensive experiments that compare the effectiveness of combining diagnostic attribute features with latent visual representations and show that they surpass the state-of-the-art approach. Finally, we analyze the proposed dual-stream architecture and provide an insight that determine the components that contribute most to classification performance.
KW - Autoencoder
KW - Diagnostic Attributes
KW - Latent Representation
KW - Multi-modal Learning
KW - OCT
UR - http://www.scopus.com/inward/record.url?scp=85129679378&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85129679378&partnerID=8YFLogxK
U2 - 10.1109/ISBI52829.2022.9761432
DO - 10.1109/ISBI52829.2022.9761432
M3 - Conference contribution
AN - SCOPUS:85129679378
T3 - Proceedings - International Symposium on Biomedical Imaging
BT - ISBI 2022 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 19th IEEE International Symposium on Biomedical Imaging, ISBI 2022
Y2 - 28 March 2022 through 31 March 2022
ER -