-
Je něco špatně v tomto záznamu ?
DeepFuse: A multi-rater fusion and refinement network for computing silver-standard annotations
CE. Akbaş, V. Ulman, M. Maška, M. Kozubek
Jazyk angličtina Země Spojené státy americké
Typ dokumentu časopisecké články
- MeSH
- lidé MeSH
- neuronové sítě * MeSH
- počítačové zpracování obrazu * metody MeSH
- Check Tag
- lidé MeSH
- Publikační typ
- časopisecké články MeSH
Achieving a reliable and accurate biomedical image segmentation is a long-standing problem. In order to train or adapt the segmentation methods or measure their performance, reference segmentation masks are required. Usually gold-standard annotations, i.e. human-origin reference annotations, are used as reference although they are very hard to obtain. The increasing size of the acquired image data, large dimensionality such as 3D or 3D + time, limited human expert time, and annotator variability, typically result in sparsely annotated gold-standard datasets. Reliable silver-standard annotations, i.e. computer-origin reference annotations, are needed to provide dense segmentation annotations by fusing multiple computer-origin segmentation results. The produced dense silver-standard annotations can then be either used as reference annotations directly, or converted into gold-standard ones with much lighter manual curation, which saves experts' time significantly. We propose a novel full-resolution multi-rater fusion convolutional neural network (CNN) architecture for biomedical image segmentation masks, called DeepFuse, which lacks any down-sampling layers. Staying everywhere at the full resolution enables DeepFuse to fully benefit from the enormous feature extraction capabilities of CNNs. DeepFuse outperforms the popular and commonly used fusion methods, STAPLE, SIMPLE and other majority-voting-based approaches with statistical significance on a wide range of benchmark datasets as demonstrated on examples of a challenging task of 2D and 3D cell and cell nuclei instance segmentation for a wide range of microscopy modalities, magnifications, cell shapes and densities. A remarkable feature of the proposed method is that it can apply specialized post-processing to the segmentation masks of each rater separately and recover under-segmented object parts during the refinement phase even if the majority of inputs vote otherwise. Thus, DeepFuse takes a big step towards obtaining fast and reliable computer-origin segmentation annotations for biomedical images.
Citace poskytuje Crossref.org
- 000
- 00000naa a2200000 a 4500
- 001
- bmc25015466
- 003
- CZ-PrNML
- 005
- 20250731091011.0
- 007
- ta
- 008
- 250708s2025 xxu f 000 0|eng||
- 009
- AR
- 024 7_
- $a 10.1016/j.compbiomed.2025.110186 $2 doi
- 035 __
- $a (PubMed)40279971
- 040 __
- $a ABA008 $b cze $d ABA008 $e AACR2
- 041 0_
- $a eng
- 044 __
- $a xxu
- 100 1_
- $a Akbaş, Cem Emre $u Masaryk University, Centre for Biomedical Image Analysis, Faculty of Informatics, Brno, 60200, Czech Republic
- 245 10
- $a DeepFuse: A multi-rater fusion and refinement network for computing silver-standard annotations / $c CE. Akbaş, V. Ulman, M. Maška, M. Kozubek
- 520 9_
- $a Achieving a reliable and accurate biomedical image segmentation is a long-standing problem. In order to train or adapt the segmentation methods or measure their performance, reference segmentation masks are required. Usually gold-standard annotations, i.e. human-origin reference annotations, are used as reference although they are very hard to obtain. The increasing size of the acquired image data, large dimensionality such as 3D or 3D + time, limited human expert time, and annotator variability, typically result in sparsely annotated gold-standard datasets. Reliable silver-standard annotations, i.e. computer-origin reference annotations, are needed to provide dense segmentation annotations by fusing multiple computer-origin segmentation results. The produced dense silver-standard annotations can then be either used as reference annotations directly, or converted into gold-standard ones with much lighter manual curation, which saves experts' time significantly. We propose a novel full-resolution multi-rater fusion convolutional neural network (CNN) architecture for biomedical image segmentation masks, called DeepFuse, which lacks any down-sampling layers. Staying everywhere at the full resolution enables DeepFuse to fully benefit from the enormous feature extraction capabilities of CNNs. DeepFuse outperforms the popular and commonly used fusion methods, STAPLE, SIMPLE and other majority-voting-based approaches with statistical significance on a wide range of benchmark datasets as demonstrated on examples of a challenging task of 2D and 3D cell and cell nuclei instance segmentation for a wide range of microscopy modalities, magnifications, cell shapes and densities. A remarkable feature of the proposed method is that it can apply specialized post-processing to the segmentation masks of each rater separately and recover under-segmented object parts during the refinement phase even if the majority of inputs vote otherwise. Thus, DeepFuse takes a big step towards obtaining fast and reliable computer-origin segmentation annotations for biomedical images.
- 650 _2
- $a lidé $7 D006801
- 650 12
- $a počítačové zpracování obrazu $x metody $7 D007091
- 650 12
- $a neuronové sítě $7 D016571
- 655 _2
- $a časopisecké články $7 D016428
- 700 1_
- $a Ulman, Vladimír $u Masaryk University, Centre for Biomedical Image Analysis, Faculty of Informatics, Brno, 60200, Czech Republic; IT4Innovations, VSB - Technical University of Ostrava, Ostrava, 70800, Czech Republic
- 700 1_
- $a Maška, Martin $u Masaryk University, Centre for Biomedical Image Analysis, Faculty of Informatics, Brno, 60200, Czech Republic
- 700 1_
- $a Kozubek, Michal $u Masaryk University, Centre for Biomedical Image Analysis, Faculty of Informatics, Brno, 60200, Czech Republic. Electronic address: kozubek@fi.muni.cz
- 773 0_
- $w MED00001218 $t Computers in biology and medicine $x 1879-0534 $g Roč. 192, č. Pt A (2025), s. 110186
- 856 41
- $u https://pubmed.ncbi.nlm.nih.gov/40279971 $y Pubmed
- 910 __
- $a ABA008 $b sig $c sign $y - $z 0
- 990 __
- $a 20250708 $b ABA008
- 991 __
- $a 20250731091006 $b ABA008
- 999 __
- $a ok $b bmc $g 2366354 $s 1252591
- BAS __
- $a 3
- BAS __
- $a PreBMC-MEDLINE
- BMC __
- $a 2025 $b 192 $c Pt A $d 110186 $e 20250424 $i 1879-0534 $m Computers in biology and medicine $n Comput Biol Med $x MED00001218
- LZP __
- $a Pubmed-20250708