2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2015
DOI: 10.1109/cvpr.2015.7298655
|View full text |Cite
|
Sign up to set email alerts
|

SUN RGB-D: A RGB-D scene understanding benchmark suite

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

5
1,163
0
1

Year Published

2016
2016
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 1,580 publications
(1,234 citation statements)
references
References 49 publications
5
1,163
0
1
Order By: Relevance
“…For more comprehensive comparisons, besides these 20 14 , from top left to bottom right is the chronological order from the video. The curves under the images are the accelerometer data at 50 Hz of devices attached to the knife, the mixing spoon, the small spoon, the peeler, the glass, the oil bottle, and the pepper dispenser mentioned datasets above, another 26 extra RGB-D datasets for different applications are also added into the tables: Birmingham University Objects, Category Modeling RGB-D [104], Cornell Activity [47,92], Cornell RGB-D [48], DGait [12], Daily Activities with occlusions [1], Heidelberg University Scenes [63], Microsoft 7-scenes [78], MobileRGBD [96], MPII Multi-Kinect [93], MSR Action3D Dataset [97], MSR 3D Online Action [103], MSRGesture3D [50], DAFT [31], Paper Kinect [70], RGBD-HuDaAct [68], Stanford Scene Object [44], Stanford 3D Scene [105], Sun3D [101], SUN RGB-D [82], TST Fall Detection [28], UTD-MHAD [14], Vienna University Technology Object [2], Willow Garage [99], Workout SU-10 exercise [67] and 3D-Mask [24]. In addition, we name those datasets without original names by means of creation place or applications.…”
Section: Discussionmentioning
confidence: 99%
“…For more comprehensive comparisons, besides these 20 14 , from top left to bottom right is the chronological order from the video. The curves under the images are the accelerometer data at 50 Hz of devices attached to the knife, the mixing spoon, the small spoon, the peeler, the glass, the oil bottle, and the pepper dispenser mentioned datasets above, another 26 extra RGB-D datasets for different applications are also added into the tables: Birmingham University Objects, Category Modeling RGB-D [104], Cornell Activity [47,92], Cornell RGB-D [48], DGait [12], Daily Activities with occlusions [1], Heidelberg University Scenes [63], Microsoft 7-scenes [78], MobileRGBD [96], MPII Multi-Kinect [93], MSR Action3D Dataset [97], MSR 3D Online Action [103], MSRGesture3D [50], DAFT [31], Paper Kinect [70], RGBD-HuDaAct [68], Stanford Scene Object [44], Stanford 3D Scene [105], Sun3D [101], SUN RGB-D [82], TST Fall Detection [28], UTD-MHAD [14], Vienna University Technology Object [2], Willow Garage [99], Workout SU-10 exercise [67] and 3D-Mask [24]. In addition, we name those datasets without original names by means of creation place or applications.…”
Section: Discussionmentioning
confidence: 99%
“…Existing indoor datasets such as the NYU Depth V2 dataset [19] and the SUN RGBD dataset [20] are largely composed by images of cluttered rooms, which are of less interest for our purposes.…”
Section: Implementation and Training A Training Datasetmentioning
confidence: 99%
“…Examples images are shown in Figure 3. It contains 967 images from three sources: 349 images from the SUN RGBD [20] (category "corridor"); 327 images from SUN database [21] (category "corridor") and 291 images from self-collected video taken around the Carnegie Mellon University campus. For the SUN database images, we used annotations where available, and manually annotated an extra 250 images using LabelMe [22].…”
Section: Implementation and Training A Training Datasetmentioning
confidence: 99%
“…We use the SUN RGB-D dataset from Song et al [31]. This dataset contains 10335 RGB images with depth maps, as well as detailed annotations for more than 1000 objects in the form of 2D and 3D bounding boxes.…”
Section: Methodsmentioning
confidence: 99%
“…We circumvent this problem by leveraging the fast RCNN detector of [14] with object proposals generated by Selective Search [36]. In detail, we finetune the ImageNet model from [14] to SUN RGB-D, using the same 19 objects as in [31]. We then run the detector on all images from our training split and keep the proposals with detection scores > 0.7 and a sufficient overlap (measured by the IoU >0.5) with the 2D ground-truth bounding boxes.…”
Section: Monitormentioning
confidence: 99%