BackgroundCurrent commercially available hybrid magnetic resonance linear accelerators (MR‐Linac) use 2D+t cine MR imaging to provide intra‐fractional motion monitoring. However, given the limited temporal resolution of cine MR imaging, target intra‐frame motion deterioration effects, resulting in effective time latency and motion artifacts in the image domain, can be appreciable, especially in the case of fast breathing.PurposeThe aim of this work is to investigate intra‐frame motion deterioration effects in MR‐guided radiotherapy (MRgRT) by simulating the motion‐corrupted image acquisition, and to explore the feasibility of deep‐learning‐based compensation approaches, relying on the intra‐frame motion information which is spatially and temporally encoded in the raw data (k‐space).MethodsAn intra‐frame motion model was defined to simulate motion‐corrupted MR images, with 4D anthropomorphic digital phantoms being exploited to provide ground truth 2D+t cine MR sequences. A total number of 10 digital phantoms were generated for lung cancer patients, with randomly selected eight patients for training or validation and the remaining two for testing. The simulation code served as the data generator, and a dedicated motion pattern perturbation scheme was proposed to build the intra‐frame motion database, where three degrees of freedom were designed to guarantee the diversity of intra‐frame motion trajectories, enabling a thorough exploration in the domain of the potential anatomical structure positions. U‐Nets with three types of loss functions: L1 or L2 loss defined in image or Fourier domain, referred to as NNImgLoss‐L1, NNFloss‐L1 and NNL2‐Loss were trained to extract information from the motion‐corrupted image and used to estimate the ground truth final‐position image, corresponding to the end of the acquisition. Images before and after compensation were evaluated in terms of (i) image mean‐squared error (MSE) and mean absolute error (MAE), and (ii) accuracy of gross tumor volume (GTV) contouring, based on optical‐flow image registration.ResultsImage degradation caused by intra‐frame motion was observed: for a linearly and fully acquired Cartesian readout k‐space trajectory, intra‐frame motion resulted in an imaging latency of approximately 50% of the acquisition time; in comparison, the motion artifacts exhibited only a negligible contribution to the overall geometric errors. All three compensation models led to a decrease in image MSE/MAE and GTV position offset compared to the motion‐corrupted image. In the investigated testing dataset for GTV contouring, the average dice similarity coefficients (DSC) improved from 88% to 96%, and the 95th percentile Hausdorff distance (HD95) dropped from 4.8 mm to 2.1 mm. Different models showed slight performance variations across different intra‐frame motion amplitude categories: NNImgLoss‐L1 excelled for small/medium amplitudes, whereas NNFloss‐L1 demonstrated higher DSC median values at larger amplitudes. The saliency maps of the motion‐corrupted image highlighted the major contribution of the later acquired k‐space data, as well as the edges of the moving anatomical structures at their final positions, during the model inference stage.ConclusionsOur results demonstrate the deep‐learning‐based approaches have the potential to compensate for intra‐frame motion by utilizing the later acquired data to drive the convergence of the earlier acquired k‐space components.