In deep learning-based computing vision for image processing, image segmentation is a prominent issue. There is promising generalisation performance in the medical image segmentation sector for approaches using domain generalisation (DG). Single domain generalisation (SDG) is a more difficult problem than conventional generalisation (DG), which requires numerous source domains to be accessible during network training, as opposed to conventional generalisation (DG). Color medical images may be incorrectly segmented because of the augmentation of the full image in order to increase model generalisation capacity. An arbitrary illumination SDG model for improving generalisation power for colour image segmentation approach for medical images through synthesizing random radiance charts is presented as a first solution to this challenge. Color medical images may be decomposed into reflectivity and illumination maps using retinex-based neural networks (ID-Nets). In order to provide medical colour images under various lighting situations, illumination randomization is used to enhance illumination maps. A new metric, TGCI, called the transfer gradient consistency index was devised to quantify the performance of the breakdown of retinal images by simulating physical lighting. Two of the existing retinal image segmentation tasks are tested extensively in order to assess our suggested system. According to the Dice coefficient, our framework surpasses previous SDGs and image improvement algorithms, outperforming the best SDGs by up to 1.7 per cent.