{"id":612135,"date":"2019-10-03T11:26:39","date_gmt":"2019-10-03T18:26:39","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-project&#038;p=612135"},"modified":"2019-10-03T11:26:39","modified_gmt":"2019-10-03T18:26:39","slug":"holotable","status":"publish","type":"msr-project","link":"https:\/\/www.microsoft.com\/en-us\/research\/project\/holotable\/","title":{"rendered":"HoloTable"},"content":{"rendered":"<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-612546\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2019\/10\/FN5A4098_small.jpg\" alt=\"HoloTable\" width=\"2880\" height=\"1078\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2019\/10\/FN5A4098_small.jpg 2880w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2019\/10\/FN5A4098_small-300x112.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2019\/10\/FN5A4098_small-768x287.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2019\/10\/FN5A4098_small-1024x383.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2019\/10\/FN5A4098_small-1920x720.jpg 1920w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2019\/10\/FN5A4098_small-1600x600.jpg 1600w\" sizes=\"auto, (max-width: 2880px) 100vw, 2880px\" \/><\/p>\n<p>HoloTable explored view-dependent rendering to simulate a 3D experience for visualizing data as a &#8220;holographic&#8221; (&#8220;above-screen&#8221;) or &#8220;volumetric&#8221; (&#8220;below-screen&#8221;) display.<\/p>\n<p>Head-tracking using a Kinect sensor provided depth-cues via parallax. Anaglyph rendering could provide stereo images. Additional depth-cues were provided using screen &#8220;reflections&#8221;.<\/p>\n<p>Interaction was via direct multitouch manipulation on a touchscreen, such as a Perceptive Pixel display.<\/p>\n<div id=\"attachment_612555\" style=\"width: 310px\" class=\"wp-caption alignnone\"><img loading=\"lazy\" decoding=\"async\" aria-describedby=\"caption-attachment-612555\" class=\"size-medium wp-image-612555\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2019\/10\/IMG_1282-300x150.jpg\" alt=\"HoloTable\" width=\"300\" height=\"150\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2019\/10\/IMG_1282-300x150.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2019\/10\/IMG_1282-768x384.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2019\/10\/IMG_1282-1024x512.jpg 1024w\" sizes=\"auto, (max-width: 300px) 100vw, 300px\" \/><p id=\"caption-attachment-612555\" class=\"wp-caption-text\">Content rendered &#8220;above&#8221; display with reflection depth cues<\/p><\/div>\n","protected":false},"excerpt":{"rendered":"<p>HoloTable explored view-dependent rendering to simulate a 3D experience for visualizing data as a &#8220;holographic&#8221; (&#8220;above-screen&#8221;) or &#8220;volumetric&#8221; (&#8220;below-screen&#8221;) display. Head-tracking using a Kinect sensor provided depth-cues via parallax. Anaglyph rendering could provide stereo images. Additional depth-cues were provided using screen &#8220;reflections&#8221;. Interaction was via direct multitouch manipulation on a touchscreen, such as a Perceptive [&hellip;]<\/p>\n","protected":false},"featured_media":0,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","footnotes":""},"research-area":[13563,13554],"msr-locale":[268875],"msr-impact-theme":[],"msr-pillar":[],"class_list":["post-612135","msr-project","type-msr-project","status-publish","hentry","msr-research-area-data-platform-analytics","msr-research-area-human-computer-interaction","msr-locale-en_us","msr-archive-status-active"],"msr_project_start":"2015-07-01","related-publications":[],"related-downloads":[],"related-videos":[],"related-groups":[550641],"related-events":[],"related-opportunities":[],"related-posts":[],"related-articles":[],"tab-content":[],"slides":[],"related-researchers":[{"type":"user_nicename","display_name":"Dave Brown","user_id":31502,"people_section":"Section name 0","alias":"dabrown"}],"msr_research_lab":[],"msr_impact_theme":[],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project\/612135","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-project"}],"version-history":[{"count":2,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project\/612135\/revisions"}],"predecessor-version":[{"id":612558,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project\/612135\/revisions\/612558"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=612135"}],"wp:term":[{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=612135"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=612135"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=612135"},{"taxonomy":"msr-pillar","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-pillar?post=612135"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}