Show simple item record

Development of Deep Learning Models for Deformable Image Registration (DIR) in the Head and Neck Region

dc.contributor.advisor Ren, Lei
dc.contributor.author Amini, Ala
dc.date.accessioned 2020-06-09T17:45:48Z
dc.date.available 2021-06-01T08:17:18Z
dc.date.issued 2020
dc.identifier.uri https://hdl.handle.net/10161/20830
dc.description Master's thesis
dc.description.abstract <p>Deformable image registration (DIR) is the process of registering two or more images to a reference image by minimizing local differences across the entire image. DIR is conventionally performed using iterative optimization-based methods, which are time-consuming and require manual parameter tuning. Recent studies have shown that deep learning methods, most importantly convolutional neural networks (CNNs), can be employed to address the DIR problem. In this study, we propose two deep learning frameworks to perform the DIR task in an unsupervised approach for CT-to-CT deformable registration of the head and neck region. Given that head and neck cancer patients might undergo severe weight loss over the course of their radiation therapy treatment, DIR in this region becomes an important task. The first proposed deep learning framework contains two scales, where both scales are based on freeform deformation, and are trained based on minimizing a dissimilarity intensity-based metrics, while encouraging the deformed vector field (DVF) smoothness. The two scales were first trained separately in a sequential manner, and then combined in a two-scale joint training framework for further optimization. We then developed a transfer learning technique to improve the DIR accuracy of the proposed deep learning networks by fine-tuning a pre-trained group-based model into a patient-specific model to optimize its performance for individual patients. We showed that by utilizing as few as two prior CT scans of a patient, the performance of the pretrained model described above can be improved yielding more accurate DIR results for individual patients. The second proposed deep learning framework, which also consists of two scales, is a hybrid DIR method using B-spline deformation modeling and deep learning. In the first scale, deformation of control points are learned by deep learning and initial DVF is estimated using B-spline interpolation to ensure smoothness of the initial estimation. Second scale model of the second framework is the same as that in the first framework. In our study, the networks were trained and evaluated using public TCIA HNSCC-3DCT for the head and neck region. We showed that our DIR results of our proposed networks are comparable to conventional DIR methods while being several orders of magnitude faster (about 2 to 3 seconds), making it highly applicable for clinical applications.</p>
dc.subject Computer science
dc.subject Medical imaging
dc.subject Computer Vision
dc.subject Deep Learning
dc.subject Deformable Image Registration
dc.title Development of Deep Learning Models for Deformable Image Registration (DIR) in the Head and Neck Region
dc.type Master's thesis
dc.department Medical Physics
duke.embargo.months 11.736986301369862


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record