dc.contributor.author | Zhuang, Mingrui | |
dc.contributor.author | Chen, Zhonghua | |
dc.contributor.author | Wang, Hongkai | |
dc.contributor.author | Tang, Hong | |
dc.contributor.author | He, Jiang | |
dc.contributor.author | Qin, Bobo | |
dc.contributor.author | Yang, Yuxin | |
dc.contributor.author | Jin, Xiaoxian | |
dc.contributor.author | Yu, Mengzhu | |
dc.contributor.author | Jin, Baitao | |
dc.contributor.author | Li, Taijing | |
dc.contributor.author | Kettunen, Lauri | |
dc.date.accessioned | 2022-09-02T08:54:54Z | |
dc.date.available | 2022-09-02T08:54:54Z | |
dc.date.issued | 2023 | |
dc.identifier.citation | Zhuang, M., Chen, Z., Wang, H., Tang, H., He, J., Qin, B., Yang, Y., Jin, X., Yu, M., Jin, B., Li, T., & Kettunen, L. (2023). Efficient contour-based annotation by iterative deep learning for organ segmentation from volumetric medical images. <i>International Journal of Computer Assisted Radiology and Surgery</i>, <i>18</i>(2), 379-394. <a href="https://doi.org/10.1007/s11548-022-02730-z" target="_blank">https://doi.org/10.1007/s11548-022-02730-z</a> | |
dc.identifier.other | CONVID_155803705 | |
dc.identifier.uri | https://jyx.jyu.fi/handle/123456789/82925 | |
dc.description.abstract | Purpose
Training deep neural networks usually require a large number of human-annotated data. For organ segmentation from volumetric medical images, human annotation is tedious and inefficient. To save human labour and to accelerate the training process, the strategy of annotation by iterative deep learning recently becomes popular in the research community. However, due to the lack of domain knowledge or efficient human-interaction tools, the current AID methods still suffer from long training time and high annotation burden.
Methods
We develop a contour-based annotation by iterative deep learning (AID) algorithm which uses boundary representation instead of voxel labels to incorporate high-level organ shape knowledge. We propose a contour segmentation network with a multi-scale feature extraction backbone to improve the boundary detection accuracy. We also developed a contour-based human-intervention method to facilitate easy adjustments of organ boundaries. By combining the contour-based segmentation network and the contour-adjustment intervention method, our algorithm achieves fast few-shot learning and efficient human proofreading.
Results
For validation, two human operators independently annotated four abdominal organs in computed tomography (CT) images using our method and two compared methods, i.e. a traditional contour-interpolation method and a state-of-the-art (SOTA) convolutional network (CNN) method based on voxel label representation. Compared to these methods, our approach considerably saved annotation time and reduced inter-rater variabilities. Our contour detection network also outperforms the SOTA nnU-Net in producing anatomically plausible organ shape with only a small training set.
Conclusion
Taking advantage of the boundary shape prior and the contour representation, our method is more efficient, more accurate and less prone to inter-operator variability than the SOTA AID methods for organ segmentation from volumetric medical images. The good shape learning ability and flexible boundary adjustment function make it suitable for fast annotation of organ structures with regular shape. | en |
dc.format.mimetype | application/pdf | |
dc.language.iso | eng | |
dc.publisher | Springer | |
dc.relation.ispartofseries | International Journal of Computer Assisted Radiology and Surgery | |
dc.rights | CC BY 4.0 | |
dc.subject.other | medical image annotation | |
dc.subject.other | deep learning | |
dc.subject.other | organ segmentation | |
dc.subject.other | interactive segmentation | |
dc.title | Efficient contour-based annotation by iterative deep learning for organ segmentation from volumetric medical images | |
dc.type | article | |
dc.identifier.urn | URN:NBN:fi:jyu-202209024458 | |
dc.contributor.laitos | Informaatioteknologian tiedekunta | fi |
dc.contributor.laitos | Faculty of Information Technology | en |
dc.contributor.oppiaine | Laskennallinen tiede | fi |
dc.contributor.oppiaine | Secure Communications Engineering and Signal Processing | fi |
dc.contributor.oppiaine | Computing, Information Technology and Mathematics | fi |
dc.contributor.oppiaine | Computational Science | en |
dc.contributor.oppiaine | Secure Communications Engineering and Signal Processing | en |
dc.contributor.oppiaine | Computing, Information Technology and Mathematics | en |
dc.type.uri | http://purl.org/eprint/type/JournalArticle | |
dc.type.coar | http://purl.org/coar/resource_type/c_2df8fbb1 | |
dc.description.reviewstatus | peerReviewed | |
dc.format.pagerange | 379-394 | |
dc.relation.issn | 1861-6410 | |
dc.relation.numberinseries | 2 | |
dc.relation.volume | 18 | |
dc.type.version | publishedVersion | |
dc.rights.copyright | © The Author(s) 2022 | |
dc.rights.accesslevel | openAccess | fi |
dc.subject.yso | lääketieteellinen tekniikka | |
dc.subject.yso | algoritmit | |
dc.subject.yso | syväoppiminen | |
dc.format.content | fulltext | |
jyx.subject.uri | http://www.yso.fi/onto/yso/p13486 | |
jyx.subject.uri | http://www.yso.fi/onto/yso/p14524 | |
jyx.subject.uri | http://www.yso.fi/onto/yso/p39324 | |
dc.rights.url | https://creativecommons.org/licenses/by/4.0/ | |
dc.relation.doi | 10.1007/s11548-022-02730-z | |
jyx.fundinginformation | This work was supported in part by the National Key Research and Development Program No. 2020YFB1711500, 2020YFB1711501 and 2020YFB1711503, the general program of National Natural Science Fund of China (No. 81971693, 61971445 and 61971089), Dalian City Science and Technology Innovation Funding (No. 2018J12GX042), the Fundamental Research Funds for the Central Universities (No. DUT19JC01 and DUT20YG122), the funding of Liaoning Key Lab of IC & BME System and Dalian Engineering Research Center for Artificial Intelligence in Medical Imaging. | |
dc.type.okm | A1 | |