Our environment has been significantly impacted by climate change. According to previous research, insect catastrophes induced by global climate change killed many trees, inevitably contributing to forest fires. The condition of the forest is an essential indicator of forest fires. Analysis of aerial images of a forest can detect deceased and living trees at an early stage. Automated forest health diagnostics are crucial for monitoring and preserving forest ecosystem health. Combining Modified Generative Adversarial Networks (MGANs) and YOLOv5 (You Only Look Once version 5) is presented in this paper as a novel method for assessing forest health using aerial images. We also employ the Tabu Search Algorithm (TSA) to enhance the process of identifying and categorizing unhealthy forest areas. The proposed model provides synthetic data to supplement the limited labeled dataset, thereby resolving the frequent issue of data scarcity in forest health diagnosis tasks. This improvement enhances the model's ability to generalize to previously unobserved data, thereby increasing the overall precision and robustness of the forest health evaluation. In addition, YOLOv5 integration enables real-time object identification, enabling the model to recognize and pinpoint numerous tree species and potential health issues with exceptional speed and accuracy. The efficient architecture of YOLOv5 enables it to be deployed on devices with limited resources, enabling forest-monitoring applications on-site. We use the TSA to enhance the identification of unhealthy forest areas. The TSA method effectively investigates the search space, ensuring the model converges to a near-optimal solution, improving disease detection precision and decreasing false positives. We evaluated our MGAN-YOLOv5 method using a large dataset of aerial images of diverse forest habitats. The experimental results demonstrated impressive performance in diagnosing forest health automatically, achieving a detection precision of 98.66%, recall of 99.99%, F1 score of 97.77%, accuracy of 99.99%, response time of 3.543 ms and computational time of 5.987 ms. Significantly, our method outperforms all the compared target detection methods showcasing a minimum improvement of 2% in mAP.