This study introduces a novel object recognition algorithm employing the CNN-BiLSTM technique. Through the integration of advanced sensory technologies and deep reinforcement learning approaches, the proposed system dynamically adjusts its detection parameters to accommodate diverse object geometries, sizes, and surface properties. The approach entails training a neural network agent to discern, from an image window, which objects within a predetermined region warrant focused attention. To effectively handle high-dimensional and time-series data, a Double DQN framework is utilized. The proposed approach amalgamates object recognition algorithms with sophisticated reinforcement learning techniques, notably Double DQN, enabling robots to dynamically perceive and grasp items in varied environments. The outcomes were obtained by conducting comparative research to evaluate the effectiveness of current approaches in comparison to the proposed method. The Q-network is trained to execute two distinct motions, namely "twist" and "push," with the intended motion provided as an assigned parameter (+1 for "push" and-1 for "twist"). Moreover, the network can generate intermediate movements between the learned motions when the task parameter falls between-1 and +1. The findings reveal notable enhancements in grasping success rates and adaptability across diverse item sizes, shapes, and conditions. Notably, the study reports a significant increase in the task completion rate for grasping, achieving a success rate of 90%.