Korkmaz, YılmazYurt, MahmutDar, Salman Ul HassanÖzbey, MuzafferÇukur, Tolga2022-01-272022-01-272021978-3-030-88551-90302-9743http://hdl.handle.net/11693/76822Conference Name: International Workshop on Machine Learning for Medical Image Reconstruction, MLMIR 2021Date of Conference: 25 September 2021Supervised training of deep network models for MRI reconstruction requires access to large databases of fully-sampled MRI acquisitions. To alleviate dependency on costly databases, unsupervised learning strategies have received interest. A powerful framework that eliminates the need for training data altogether is the deep image prior (DIP). To do this, DIP inverts randomly-initialized models to infer network parameters most consistent with the undersampled test data. However, existing DIP methods leverage convolutional backbones, suffering from limited sensitivity to long-range spatial dependencies and thereby poor model invertibility. To address these limitations, here we propose an unsupervised MRI reconstruction based on a novel generative vision transformer (GVTrans). GVTrans progressively maps low-dimensional noise and latent variables onto MR images via cascaded blocks of cross-attention vision transformers. Cross-attention mechanism between latents and image features serve to enhance representational learning of local and global context. Meanwhile, latent and noise injections at each network layer permit fine control of generated image features, improving model invertibility. Demonstrations are performed for scan-specific reconstruction of brain MRI data at multiple contrasts and acceleration factors. GVTrans yields superior performance to state-of-the-art generative models based on convolutional neural networks (CNNs).EnglishMRI reconstructionTransformerGenerativeAttentionUnsupervisedDeep MRI reconstruction with generative vision transformerConference Paper10.1007/978-3-030-88552-6_6978-3-030-88552-61611-3349