In this work, the effect of spatial gradients in the training data on deep learning-based QSM is explored. We observe that deep learning-based QSM underestimates the susceptibility values when spatial gradients differ between training and test data. For demonstration, three types of networks were trained by using different spatial gradients of training images and evaluated on test data with varying spatial gradients. The results indicate that expanding the spatial gradient distribution of training data improves the performance of deep learning-based QSM. Furthermore, we demonstrate that augmenting spatial gradients may improve deep-learning based QSM to work for various image resolutions.
This abstract and the presentation materials are available to members only; a login is required.