- Qiupu Chen ORCID:orcid.org/0000-0002-7423-009013,14,
- Yimou Wang13,14,
- Jun Xu ORCID:orcid.org/0009-0001-4016-408013,14 &
- …
- Qiankun Li ORCID:orcid.org/0000-0001-5121-168213,14
Part of the book series:Lecture Notes in Computer Science ((LNCS,volume 14623))
Included in the following conference series:
78Accesses
Abstract
Accurate 3D tooth segmentation plays a pivotal role in the realm of computer-aided dental diagnosis and treatment. Despite its significance, the scarcity of labeled 3D tooth data poses a substantial challenge. This paper details our contribution to the MICCAI 2023 semi-supervised teeth segmentation challenge, aimed at enhancing the precision of tooth segmentation through the application of deep learning within a semi-supervised learning framework. Our approach leverages the nn-UNet architecture, incorporating innovative modifications to improve segmentation performance. Notably, we introduce two novel components: an axial attention mechanism module and a positional correction module. The axial attention mechanism enhances the model’s ability to capture contextual information in axial slices, contributing to improved segmentation accuracy. Simultaneously, the positional correction module solves the problem of incorrect segmentation of bony structures similar to tooth morphology and density. In the context of semi-supervised learning, where labeled data is limited, we propose a robust selection methodology for pseudo labels. This methodology considers the stability of pseudo labels across re-training iterations, ensuring the reliability of the learning process. The integration of these components and methodologies collectively enhances the model’s adaptability to the challenges posed by limited labeled data. The proposed model ranked fifth in the final ranking on unseen test data, with a score of 81.47%. The code is available athttps://github.com/qpuchen/nnUNet_att_position_correction.
Q. Chen and Y. Wang—Equal contributions.
This is a preview of subscription content,log in via an institution to check access.
Access this chapter
Subscribe and save
- Get 10 units per month
- Download Article/Chapter or eBook
- 1 Unit = 1 Article or 1 Chapter
- Cancel anytime
Buy Now
- Chapter
- JPY 3498
- Price includes VAT (Japan)
- eBook
- JPY 5719
- Price includes VAT (Japan)
- Softcover Book
- JPY 7149
- Price includes VAT (Japan)
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Chen, Y., et al.: Automatic segmentation of individual tooth in dental CBCT images from tooth surface map by a multi-task FCN. IEEE Access8, 97296–97309 (2020)
Cui, W et al.: Ctooth+: a large-scale dental cone beam computed tomography dataset and benchmark for tooth volume segmentation. In: MICCAI Workshop on Data Augmentation, Labelling, and Imperfections, pp. 64–73. Springer (2022).https://doi.org/10.1007/978-3-031-17027-0_7
Cui, W., et al.: Ctooth: a fully annotated 3d dataset and benchmark for tooth volume segmentation on cone beam computed tomography images. In: International Conference on Intelligent Robotics and Applications, pp. 191–200. Springer (2022).https://doi.org/10.1007/978-3-031-13841-6_18
Hao, J., et al.: Ai-enabled automatic multimodal fusion of cone-beam ct and intraoral scans for intelligent 3d tooth-bone reconstruction and clinical applications. arXiv preprintarXiv:2203.05784 (2022)
Ho, J., Kalchbrenner, N., Weissenborn, D., Salimans, T.: Axial attention in multidimensional transformers. arXiv preprintarXiv:1912.12180 (2019)
Huang, Z., et al.: Revisiting nnu-net for iterative pseudo labeling and efficient sliding window inference. In: MICCAI Challenge on Fast and Low-Resource Semi-supervised Abdominal Organ Segmentation, pp. 178–189. Springer (2022).https://doi.org/10.1007/978-3-031-13841-6_18
Isensee, F., Jaeger, P.F., Kohl, S.A., Petersen, J., Maier-Hein, K.H.: nnu-net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods18(2), 203–211 (2021)
Jang, T.J., Kim, K.C., Cho, H.C., Seo, J.K.: A fully automated method for 3d individual tooth identification and segmentation in dental cbct. IEEE Trans. Pattern Anal. Mach. Intell.44(10), 6562–6568 (2021)
Liu, Z., et al.: Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 10012–10022 (2021)
Luu, H.M., Park, S.H.: Extending nn-unet for brain tumor segmentation. In: International MICCAI Brainlesion Workshop, pp. 173–186. Springer (2021).https://doi.org/10.1007/978-3-031-09002-8_16
Polizzi, A., et al.: Tooth automatic segmentation from cbct images: a systematic review. Clin. Oral Investi., 1–16 (2023)
Tang, Y., et al.: Self-supervised pre-training of swin transformers for 3d medical image analysis. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 20730–20740 (2022)
Uthman, A., Salman, B., Aldeen, H.S., Marei, H., Al-Bayati, S.F., Al-Rawi, N.H.: Morphometric analysis of odontoid process among arab population: a retrospective cone beam ct study. PeerJ11, e15411 (2023)
Vaswani, A., et al.: Attention is all you need. Adv. Neural Inform. Process. Syst.30 (2017)
Wang, H., Minnema, J., Batenburg, K.J., Forouzanfar, T., Hu, F.J., Wu, G.: Multiclass cbct image segmentation for orthodontics with deep learning. J. Dent. Res.100(9), 943–949 (2021)
Zhou, H.Y., Guo, J., Zhang, Y., Yu, L., Wang, L., Yu, Y.: nnformer: Interleaved transformer for volumetric segmentation. arXiv preprintarXiv:2109.03201 (2021)
Zhou, Z., Rahman Siddiquee, M.M., Tajbakhsh, N., Liang, J.: UNet++: a nested U-Net architecture for medical image segmentation. In: Stoyanov, D., et al. (eds.) DLMIA/ML-CDS -2018. LNCS, vol. 11045, pp. 3–11. Springer, Cham (2018).https://doi.org/10.1007/978-3-030-00889-5_1
Acknowledgements
The authors of this paper declare that the segmentation method they implemented for participation in the STS 2023 challenge has not used any pre-trained models nor additional datasets other than those provided by the organizers. The proposed solution is fully automatic without any manual intervention. We thank all the data owners for making the X-ray images and CT scans publicly available and Alibaba Cloud for hosting the challenge platform.
Author information
Authors and Affiliations
HFIPS, Chinese Academy of Sciences, Hefei, Anhui, China
Qiupu Chen, Yimou Wang, Jun Xu & Qiankun Li
University of Science and Technology of China, Hefei, China
Qiupu Chen, Yimou Wang, Jun Xu & Qiankun Li
- Qiupu Chen
You can also search for this author inPubMed Google Scholar
- Yimou Wang
You can also search for this author inPubMed Google Scholar
- Jun Xu
You can also search for this author inPubMed Google Scholar
- Qiankun Li
You can also search for this author inPubMed Google Scholar
Corresponding author
Correspondence toQiankun Li.
Editor information
Editors and Affiliations
Communication University of Zhejiang, Hangzhou, China
Yaqi Wang
Hangzhou Dianzi University, Hangzhou, China
Xiaodiao Chen
Shanghai Jiao Tong University, Shanghai, China
Dahong Qian
Hangzhou Dianzi University, Hangzhou, China
Fan Ye
Sichuan University, Chengdu, China
Shuai Wang
Shenzhen University, Shenzhen, China
Hongyuan Zhang
Rights and permissions
Copyright information
© 2025 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Chen, Q., Wang, Y., Xu, J., Li, Q. (2025). Semi-supervised 3D Tooth Segmentation Using nn-UNet with Axial Attention and Positional Correction. In: Wang, Y., Chen, X., Qian, D., Ye, F., Wang, S., Zhang, H. (eds) Semi-supervised Tooth Segmentation. STS 2023. Lecture Notes in Computer Science, vol 14623. Springer, Cham. https://doi.org/10.1007/978-3-031-72396-4_9
Download citation
Published:
Publisher Name:Springer, Cham
Print ISBN:978-3-031-72395-7
Online ISBN:978-3-031-72396-4
eBook Packages:Computer ScienceComputer Science (R0)
Share this paper
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative