In the process of automatic grabbing of bridge segment beams, it is crucial to accurately locate and align the corner points of the crane's boom with the beam's lifting holes. This requires the utilization of image processing techniques to precisely detect and locate the corner points of the crane's boom. Existing feature matching methods face challenges such as low detection accuracy and unsuitability for this specific scenario. This article proposes a novel approach for corner point localization by using the intersection points of lines, facilitating the matching of feature points between left and right images. The method consists of three steps: first, a grayscale difference map is constructed by utilizing the R and G channels of the RGB color space. This enhances the bimodal characteristics of the grayscale histograms between the foreground and background, which facilitates the subsequent binarization process. Additionally, opening and closing operations are employed to remove small artifacts from the Canny edge detection results, effectively reducing noise. Second, an adaptive thresholding method based on the mean and variance of Hough transform voting scores is proposed. This method filters out interference lines from the clustering results by selecting appropriate voting scores. Furthermore, an improved centroid calculation method is introduced, which utilizes weighted formulas based on different proportions of voting scores. These weighted formulas replace the original clustering centroids as the basis for line fitting. Finally, the corner coordinates of the crane's boom are computed based on the line fitting results, and the recognition accuracy is compared under different lighting conditions. Experimental results demonstrate that the proposed algorithm exhibits smaller detection errors and higher robustness compared to other corner detection algorithms, particularly when there are numerous interference edge points in the edge detection results. The computed corner coordinates achieve pixel‐level accuracy. The algorithm performs optimally under strong supplementary lighting conditions, with an average detection error percentage of 97.1% within 0–2 pixels and a recognition accuracy of 98.6%. The recognition success rates under different lighting conditions are all above 92.9%. This method is superior to traditional corner detection methods, meets the requirements for automatic grabbing of the boom, and holds practical engineering value. It provides a basis for addressing the accuracy and robustness challenges of crane algorithms influenced by multiple environmental factors.