It is challenging to apply depth maps generated from sparse laser scan data to computer vision tasks, such as robot vision and autonomous driving, because of the sparsity and noise in the data. To overcome this problem, depth completion tasks have been proposed to produce a dense depth map from sparse LiDAR data and a single RGB image. In this study, we developed a deep convolutional architecture with cross guidance for multi-modal feature fusion to compensate for the lack of representation power of their modality. Two encoders, which are part of the proposed architecture, receive different modalities as inputs. They interact with each other by exchanging information in each stage through the attention mechanism during encoding. We also propose a residual atrous spatial pyramid block, comprising multiple dilated convolutions with different dilation rates, which are used to derive highly significant features. The experimental results of the KITTI depth completion benchmark dataset demonstrate that the proposed architecture shows higher performance than that of the other models trained in a two-dimensional space without pre-training or finetuning other datasets. INDEX TERMS Depth estimation, depth completion, LiDAR data, cross guidance, multi-scale dilated convolutional block. Recently, artificial neural network models with deep learning have been used in state-of-the-art technologies of pattern recognition and machine learning. In particular, convolutional neural networks (CNNs) exhibit excellent performance in many computer vision tasks. While conventional CNNs [3]-[5] comprise blocks of stacking convolution