Êtes-vous un étudiant de l'EPFL à la recherche d'un projet de semestre?
Travaillez avec nous sur des projets en science des données et en visualisation, et déployez votre projet sous forme d'application sur Graph Search.
Purpose: We present a 3-dimensional patient-specific eye model from magnetic resonance imaging (MRI) for proton therapy treatment planning of uveal melanoma (UM). During MRI acquisition of UM patients, the point fixation can be difficult and, together with physiological blinking, can introduce motion artifacts in the images, thus challenging the model creation. Furthermore, the unclear boundary of the small objects (eg, lens, optic nerve) near the muscle or of the tumors with hemorrhage and tantalum clips can limit model accuracy. Methods and Materials: A dataset of 37 subjects, including 30 healthy eyes of volunteers and 7 eyes of UM patients, was investigated. In our previous work, active shape model was successfully applied to retinoblastoma eye segmentation in T1-weighted 3T MRI. Here, we evaluate this method in a more challenging setting, based on 1.5T MRI acquisition and different datasets of awake adult eyes with UM. The lens and cornea together with the sclera, vitreous humor, and optic nerve were automatically segmented and validated against manual delineations of a senior ocular radiation oncologist, in terms of the Dice similarity coefficient and Hausdorff distance. Results: Leave-one-out cross validation (mixing both volunteers and UM patients) yielded median Dice similarity coefficient values (respective of Hausdorff distance) of 94.5% (1.64 mm) for the sclera, 92.2% (1.73 mm) for the vitreous humor, 88.3% (1.09 mm) for the lens, and 81.9% (1.86 mm) for the optic nerve. The average computation time for an eye was 10 seconds. Conclusions: To our knowledge, our work is the first attempt to automatically segment adult eyes, including patients with UM. Our results show that automated active shape model segmentation can succeed in the presence of motion, tumors, and tantalum clips. These results are promising for inclusion in clinical practice. (C) 2018 Elsevier Inc. All rights reserved.
Christophe Moser, Timothé Laforest, Mathieu Künzi