Back to Search Start Over

Standardized assessment of automatic segmentation of white matter hyperintensities; results of the WMH segmentation challenge

Authors :
Universitat Politècnica de Catalunya. Doctorat en Teoria del Senyal i Comunicacions
Universitat Politècnica de Catalunya. Departament de Teoria del Senyal i Comunicacions
Universitat Politècnica de Catalunya. GPI - Grup de Processament d'Imatge i Vídeo
Kuijf, Hugo J.
Biesbroek, J. Matthijs
De Bresser, Jeroen
Casamitjana Díaz, Adrià
Vilaplana Besler, Verónica
Universitat Politècnica de Catalunya. Doctorat en Teoria del Senyal i Comunicacions
Universitat Politècnica de Catalunya. Departament de Teoria del Senyal i Comunicacions
Universitat Politècnica de Catalunya. GPI - Grup de Processament d'Imatge i Vídeo
Kuijf, Hugo J.
Biesbroek, J. Matthijs
De Bresser, Jeroen
Casamitjana Díaz, Adrià
Vilaplana Besler, Verónica
Publication Year :
2019

Abstract

© 2019 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes,creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.<br />Quantification of white matter hyperintensities (WMH) of presumed vascular origin is of key importance in many neurological research studies. Advanced measurements are obtained from manual segmentations on brain MR images, which is a laborious procedure. Automatic WMH segmentation methods exist, but a standardized comparison of such methods is lacking. We organized a scientific challenge, in which developers could evaluate their method on a standardized multi-center/-scanner image dataset, giving an objective comparison: the WMH Segmentation Challenge (http://wmh.isi.uu.nl/). Sixty T1+FLAIR images from three MR scanners were released with manual WMH segmentations. A secret test set of 110 images from five MR scanners was used for evaluation. Methods had to be containerized and submitted to the challenge organizers. Five evaluation metrics were used to rank the methods: (1) Dice Similarity Coefficient, (2) modified Hausdorff distance (95th percentile), (3) absolute percentage volume difference, (4) sensitivity for detecting individual lesions, and (5) F1-score for individual lesions. Additionally, methods were ranked on their inter-scanner robustness. Twenty participants submitted their method for evaluation. This paper provides a detailed analysis of the results. In brief, there is a cluster of four methods that rank significantly better than the other methods. There is one clear winner, which also has the best inter-scanner robustness. The challenge remains open for future submissions and provides a public platform for method evaluation.<br />Peer Reviewed<br />Postprint (author's final draft)

Details

Database :
OAIster
Notes :
13 p., application/pdf, English
Publication Type :
Electronic Resource
Accession number :
edsoai.on1141697742
Document Type :
Electronic Resource