To enhance the convergence speed and calculation precision of the grey wolf optimization algorithm (GWO), this paper proposes a dynamic generalized opposition-based grey wolf optimization algorithm (DOGWO). A dynamic generalized opposition-based learning strategy enhances the diversity of search populations and increases the potential of finding better solutions which can accelerate the convergence speed, improve the calculation precision, and avoid local optima to some extent. Furthermore, 23 benchmark functions were employed to evaluate the DOGWO algorithm. Experimental results show that the proposed DOGWO algorithm could provide very competitive results compared with other analyzed algorithms, with a faster convergence speed, higher calculation precision, and stronger stability.