TY - GEN
T1 - WaveFormer
T2 - 2024 International Joint Conference on Neural Networks, IJCNN 2024
AU - Chen, Yanlin
AU - Zhang, Xiaoqing
AU - Wang, Tianao
AU - Tang, Chen
AU - Ye, Haili
AU - Liu, Jiang
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Pathology symptoms of Parkinson disease (PD) are different from those of retinal diseases in the retinal layers, which are subtle. However, segmenting pathology information of PD from retinal layers automatically based on optical coherence tomography (OCT) images has not been studied before. Although existing Transformer-based segmentation methods have achieved good segmentation results, they have limitations in capturing local context information. Convolutional neural networks (CNNs) can construct local context dependencies among pixels, which is complementary to Transformers. Particularly, edge information extraction is significant for accurate retinal layer segmentation, which is ignored by both Transformers and CNNs but can be captured by frequency domain learning methods. To fully leverage the advantages of Transformers, CNNs, and frequency domain learning methods, we propose a Wavelet Transformer (WaveFormer) for retinal layer segmentation based on OCT images. In the WaveFormer, we design a Wavelet Spatial Attention block to exploit the potential of frequency information. Based on these advantages, WaveFormer can be data-efficient in limited OCT images of PD. The experimental results on the OCT-PD segmentation dataset show that our WaveFormer outperforms existing Transformers and CNNs. For example, WaveFormer outperforms Swin-UNet by 3.41% of IoU.
AB - Pathology symptoms of Parkinson disease (PD) are different from those of retinal diseases in the retinal layers, which are subtle. However, segmenting pathology information of PD from retinal layers automatically based on optical coherence tomography (OCT) images has not been studied before. Although existing Transformer-based segmentation methods have achieved good segmentation results, they have limitations in capturing local context information. Convolutional neural networks (CNNs) can construct local context dependencies among pixels, which is complementary to Transformers. Particularly, edge information extraction is significant for accurate retinal layer segmentation, which is ignored by both Transformers and CNNs but can be captured by frequency domain learning methods. To fully leverage the advantages of Transformers, CNNs, and frequency domain learning methods, we propose a Wavelet Transformer (WaveFormer) for retinal layer segmentation based on OCT images. In the WaveFormer, we design a Wavelet Spatial Attention block to exploit the potential of frequency information. Based on these advantages, WaveFormer can be data-efficient in limited OCT images of PD. The experimental results on the OCT-PD segmentation dataset show that our WaveFormer outperforms existing Transformers and CNNs. For example, WaveFormer outperforms Swin-UNet by 3.41% of IoU.
UR - http://www.scopus.com/inward/record.url?scp=85204955654&partnerID=8YFLogxK
U2 - 10.1109/IJCNN60899.2024.10650098
DO - 10.1109/IJCNN60899.2024.10650098
M3 - Conference contribution
AN - SCOPUS:85204955654
T3 - Proceedings of the International Joint Conference on Neural Networks
BT - 2024 International Joint Conference on Neural Networks, IJCNN 2024 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 30 June 2024 through 5 July 2024
ER -