In recent years, there has been growing interest in autonomous mobile robots equipped with Simultaneous Localization and Mapping (SLAM) technology as a solution to labour shortages in production and distribution settings. SLAM allows these robots to create maps of their environment using devices such as Lidar, radar, and sonar sensors, enabling them to navigate and track routes without prior knowledge of the environment. However, the manual operation of these robots for map construction can be labour-intensive. To address this issue, this research aims to develop a 3D SLAM autonomous mobile robot system that eliminates the need for manual map construction by utilizing existing layout maps. The system includes a PC for self-position estimation, 3DLidar, a camera for verification, a touch panel display, and the mobile robot itself. The proposed SLAM method extracts stable wall point cloud information from 3DLidar, matches it with the wall surface information in the layout map, and uses a particle filter to estimate the robot’s position. The system also includes features such as route creation, tracking, and obstacle detection for autonomous movement. Experiments were conducted to compare the proposed system with conventional 3D SLAM methods. The results showed that the proposed system significantly reduced errors in self-positioning and enabled accurate autonomous movement on specified routes, even in the presence of slight differences in layout maps and obstacles. Ultimately, this research demonstrates the effectiveness of a system that can transport goods without the need for manual environment mapping, addressing labour shortages in such environments.