Autonomous robotic navigation in forested environments is difficult because of the highly variable appearance and geometric properties of the terrain. In most navigation systems, researchers assume a priori knowledge of the terrain appearance properties, geometric properties, or both. In forest environments, vegetation such as trees, shrubs, and bushes has appearance and geometric properties that vary with change of seasons, vegetation age, and vegetation species. In addition, in forested environments the terrain surface is often rough, sloped, and/or covered with a surface layer of grass, vegetation, or snow. The complexity of the forest environment presents difficult challenges for autonomous navigation systems. In this paper, a self-supervised sensing approach is introduced that attempts to robustly identify a drivable terrain surface for robots operating in forested terrain. The sensing system employs both LIDAR and vision sensor data. There are three main stages in the system: feature learning, feature training, and terrain prediction. In the feature learning stage, 3D range points from LIDAR are analyzed to obtain an estimate of the ground surface location. In the feature training stage, the ground surface estimate is used to train a visual classifier to discriminate between ground and nonground regions of the image. In the prediction stage, the ground surface location can be estimated at high frequency solely from vision sensor data. C 2012 Wiley Periodicals, Inc.