{"id":171165,"date":"2013-07-03T01:46:14","date_gmt":"2013-07-03T01:46:14","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/project\/rgb-d-dataset-7-scenes\/"},"modified":"2022-09-07T10:55:36","modified_gmt":"2022-09-07T17:55:36","slug":"rgb-d-dataset-7-scenes","status":"publish","type":"msr-project","link":"https:\/\/www.microsoft.com\/en-us\/research\/project\/rgb-d-dataset-7-scenes\/","title":{"rendered":"RGB-D Dataset 7-Scenes"},"content":{"rendered":"<div class=\"asset-content\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/01\/rgb-d-dataset-7-scenes-1.png\"><img loading=\"lazy\" decoding=\"async\" class=\"size-full wp-image-171166 alignleft\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/01\/rgb-d-dataset-7-scenes-1.png\" alt=\"RGB-D Dataset 7-Scenes\" width=\"200\" height=\"150\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/01\/rgb-d-dataset-7-scenes-1.png 200w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/01\/rgb-d-dataset-7-scenes-1-80x60.png 80w\" sizes=\"auto, (max-width: 200px) 100vw, 200px\" \/><\/a>The 7-Scenes dataset is a collection of tracked RGB-D camera frames. The dataset may be used for evaluation of methods for different applications such as dense tracking and mapping and relocalization techniques.<\/div>\n<p><!-- .asset-content --><\/p>\n<div id=\"en-usprojects7-scenesdefault\" class=\"page-content\">\n<h1><\/h1>\n<h2><\/h2>\n<h2>Overview<\/h2>\n<p>All scenes were recorded from a handheld Kinect RGB-D camera at 640&#215;480 resolution. We use an implementation of the <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/project\/kinectfusion-project-page\/\" target=\"_self\" rel=\"noopener\">KinectFusion<\/a> system to obtain the &#8216;ground truth&#8217; camera tracks, and a dense 3D model. Several sequences were recorded per scene by different users, and split into distinct training and testing sequence sets. Details on how this data can be used for example for the evaluation of relocalization methods can be found in our papers listed under publications.<\/p>\n<p><span id=\"c74bb362-b5d1-4091-a528-12ab2400eabc\" class=\"ImageBlock fn\"><img decoding=\"async\" id=\"Imagec74bb362-b5d1-4091-a528-12ab2400eabc\" title=\"7-Scenes Overview\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/7-scenes-7-scenes-overview.png\" alt=\"7-Scenes Overview\" \/><span id=\"ImageCaptionc74bb362-b5d1-4091-a528-12ab2400eabc\" class=\"ImageCaptionCoreCss ImageCaption\"><\/span><\/span><\/p>\n<h2>Data Description<\/h2>\n<p>For each scene, we provide one zip file which contains several sequences. Each sequence is a continuous stream of tracked RGB-D camera frames. Tracking\u00a0has been performed using ICP and frame-to-model alignment\u00a0with respect to a dense reconstruction\u00a0represented by a\u00a0truncated signed distance\u00a0volume.<\/p>\n<p>Each sequence (seq-XX.zip)\u00a0consists of 500-1000 frames. Each frame consists of three files:<\/p>\n<ul>\n<li>\n<div>Color: frame-XXXXXX.color.png (RGB, 24-bit, PNG)<\/div>\n<\/li>\n<li>\n<div>Depth: frame-XXXXXX.depth.png (depth in millimeters, 16-bit, PNG, invalid depth is set to 65535).<\/div>\n<\/li>\n<li>\n<div>Pose: frame-XXXXXX.pose.txt (camera-to-world, 4&#215;4 matrix in homogeneous coordinates).<\/div>\n<\/li>\n<\/ul>\n<p>For each scene, we further provide:<\/p>\n<ul>\n<li>\n<div>For Evaluation: TrainSplit.txt \/ TestSplit.txt (splits used in the papers listed under publications).<\/div>\n<\/li>\n<li>\n<div>TSDF Volume: The dense reconstruction used for frame-to-model alignment.\u00a0Volumes are stored in\u00a0MetaImage format (2-file-format with a text header and binary data). Binary data stores the signed distances\u00a0for each voxel in 16-bit short. All volumes are of size 512x512x512 with varying element spacings and origin offsets. The spacings and offsets in millimeters are provided in the header file.<\/div>\n<\/li>\n<li>\n<div>A screenshot of the raycasted dense reconstruction.<\/div>\n<\/li>\n<\/ul>\n<p><strong>Please note:<\/strong> The RGB and depth camera have not been calibrated and we can&#8217;t provide calibration parameters at the moment. The recorded frames correspond to the raw, uncalibrated camera images. In the KinectFusion pipeline we used the following default intrinsics for the depth camera: Principle point (320,240), Focal length (585,585).<\/p>\n<h2>License Agreement<\/h2>\n<p>The data is provided for non-commercial use only. By downloading the data, you\u00a0accept\u00a0the <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/7-scenes-msr-la-dataset-7-scenes.rtf\" target=\"_self\" rel=\"noopener\">license agreement<\/a> which can be downloaded <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/7-scenes-msr-la-dataset-7-scenes.rtf\" target=\"_self\" rel=\"noopener\">here<\/a>.<\/p>\n<h2>Citations<\/h2>\n<p>If you report results based on the 7-scenes dataset, please cite at least one of the papers mentioned under publications. You may choose the paper that is more relevant to your own publication.<\/p>\n<h2>Downloads<\/h2>\n<ul>\n<li>\n<h5><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/download.microsoft.com\/download\/2\/8\/5\/28564B23-0828-408F-8631-23B1EFF1DAC8\/chess.zip\" target=\"_blank\" rel=\"noopener noreferrer\">Chess<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/download.microsoft.com\/download\/2\/8\/5\/28564B23-0828-408F-8631-23B1EFF1DAC8\/fire.zip\" target=\"_blank\" rel=\"noopener noreferrer\">Fire<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/download.microsoft.com\/download\/2\/8\/5\/28564B23-0828-408F-8631-23B1EFF1DAC8\/heads.zip\" target=\"_blank\" rel=\"noopener noreferrer\">Heads<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/download.microsoft.com\/download\/2\/8\/5\/28564B23-0828-408F-8631-23B1EFF1DAC8\/office.zip\" target=\"_blank\" rel=\"noopener noreferrer\">Office<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/download.microsoft.com\/download\/2\/8\/5\/28564B23-0828-408F-8631-23B1EFF1DAC8\/pumpkin.zip\" target=\"_blank\" rel=\"noopener noreferrer\">Pumpkin<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/download.microsoft.com\/download\/2\/8\/5\/28564B23-0828-408F-8631-23B1EFF1DAC8\/redkitchen.zip\" target=\"_blank\" rel=\"noopener noreferrer\">RedKitchen<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/download.microsoft.com\/download\/2\/8\/5\/28564B23-0828-408F-8631-23B1EFF1DAC8\/stairs.zip\" target=\"_blank\" rel=\"noopener noreferrer\">Stairs<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/h5>\n<\/li>\n<li>\n<h5><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"http:\/\/download.microsoft.com\/download\/2\/8\/5\/28564B23-0828-408F-8631-23B1EFF1DAC8\/tsdf.zip\">All TSDF Volumes<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/h5>\n<\/li>\n<li>\n<h5><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/7-scenes-msr-la-dataset-7-scenes.rtf\" target=\"_self\" rel=\"noopener\">License Agreement<\/a><\/h5>\n<\/li>\n<\/ul>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>The 7-Scenes dataset is a collection of tracked RGB-D camera frames. The dataset may be used for evaluation of methods for different applications such as dense tracking and mapping and relocalization techniques. Overview All scenes were recorded from a handheld Kinect RGB-D camera at 640&#215;480 resolution. We use an implementation of the KinectFusion system to [&hellip;]<\/p>\n","protected":false},"featured_media":0,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","footnotes":""},"research-area":[13556],"msr-locale":[268875],"msr-impact-theme":[],"msr-pillar":[],"class_list":["post-171165","msr-project","type-msr-project","status-publish","hentry","msr-research-area-artificial-intelligence","msr-locale-en_us","msr-archive-status-active"],"msr_project_start":"2013-01-01","related-publications":[164357,165138],"related-downloads":[],"related-videos":[],"related-groups":[],"related-events":[],"related-opportunities":[],"related-posts":[],"related-articles":[],"tab-content":[],"slides":[],"related-researchers":[{"type":"user_nicename","display_name":"Antonio Criminisi","user_id":41790,"people_section":"Section name 0","alias":"acriminisi"}],"msr_research_lab":[],"msr_impact_theme":[],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project\/171165","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-project"}],"version-history":[{"count":4,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project\/171165\/revisions"}],"predecessor-version":[{"id":875994,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project\/171165\/revisions\/875994"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=171165"}],"wp:term":[{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=171165"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=171165"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=171165"},{"taxonomy":"msr-pillar","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-pillar?post=171165"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}