We consider solving the low rank matrix sensing problem with Factorized Gradient Descend (FGD) method when the true rank is unknown and over-specified, which we refer to as over-parameterized matrix sensing. If the ground truth signal X * ∈ R d * d is of rank r, but we try to recover it using FF where F ∈ R d * k and k > r, the existing statistical analysis falls short, due to a flat local curvature of the loss function around the global maxima. By decomposing the factorized matrix F into separate column spaces to capture the effect of extra ranks, we show that F t F t − X * 2 F converges to a statistical error of Õ kdσ 2 /n after Õ( σr σ n d ) number of iterations where F t is the output of FGD after t iterations, σ 2 is the variance of the observation noise, σ r is the r-th largest eigenvalue of X * , and n is the number of sample. Our results, therefore, offer a comprehensive picture of the statistical and computational complexity of FGD for the over-parameterized matrix sensing problem.