{"id":293411,"date":"2016-09-15T17:53:03","date_gmt":"2016-09-16T00:53:03","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-event&#038;p=293411"},"modified":"2025-08-06T11:59:15","modified_gmt":"2025-08-06T18:59:15","slug":"dialog-state-tracking-challenge","status":"publish","type":"msr-event","link":"https:\/\/www.microsoft.com\/en-us\/research\/event\/dialog-state-tracking-challenge\/","title":{"rendered":"2011 Dialog System Technology Challenge (DSTC)"},"content":{"rendered":"\n\n<p>The 2011 Dialog System Technology Challenge &#8212; formerly the Dialog State Tracking Challenge &#8212; is a series of research community challenge tasks for accurately estimating a user&#8217;s goal in a spoken dialog system.<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<p>The Dialog State Tracking Challenge (DSTC) is an on-going series of research community challenge tasks. Each task released dialog data labeled with dialog state information, such as the user&#8217;s desired restaurant search query given all of the dialog history up to the current turn. The challenge is to create a &#8220;tracker&#8221; that can predict the dialog state for new dialogs. In each challenge, trackers are evaluated using held-out dialog data.<\/p>\n<p>You can find information on the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/sites.google.com\/dstc.community\/dstc8\/\">current DTSC8 competition here<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.<\/p>\n<h3>DSTC1<\/h3>\n<p>DSTC1 used human-computer dialogs in the bus timetable domain. Results were presented in a special session at <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/www.sigdial.org\/workshops\/sigdial2013\/\">SIGDIAL 2013<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. DSTC1 was organized by Jason D. Williams, Alan Black, Deepak Ramachandran, Antoine Raux.<\/p>\n<p><b>For DSTC1 data, and more information, see &#8220;DSTC1&#8221; tabs above.<\/b><\/p>\n<h3>DSTC2 and DSTC3<\/h3>\n<p>DSTC2\/3 used human-computer dialogs in the restaurant information domain. Results were presented in special sessions at <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/www.sigdial.org\/workshops\/conference15\/\">SIGDIAL 2014<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> and <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/www.slt2014.org\/\">IEEE SLT 2014<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. DSTC2 and 3 were organized by Matthew Henderson, Blaise Thomson, and Jason D. Williams.<\/p>\n<p><b>For DSTC2 and DSTC3 data, and more information, see the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/camdial.org\/~mh521\/dstc\/\">DSTC2\/3 website.<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/b><\/p>\n<h3>DSTC4<\/h3>\n<p>DSTC4 used human-human dialogs in the tourist information domain. Results were presented at <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/www.iwsds.org\/\">IWSDS 2015<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. DSTC4 was organized by Seokhwan Kim, Luis F. D&#8217;Haro, Rafael E Banchs, Matthew Henderson, and Jason D. Williams.<\/p>\n<p><b>For more information about DSTC4, see the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/www.colips.org\/workshop\/dstc4\/\">DSTC4 website.<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/b><\/p>\n<h3>DSTC5<\/h3>\n<p>DSTC5 used human-human dialogs in the tourist information domain, where training dialogs were provided in one language, and test dialogs were in a different language. Results were presented in a special session at <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/www.slt2016.org\/\">IEEE SLT 2016<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. DSTC5 was organized by Seokhwan Kim, Luis F. D&#8217;Haro, Rafael E Banchs, Matthew Henderson, Jason D. Williams, and Koichiro Yoshino.<\/p>\n<p><b>For more information about DSTC5, see the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/workshop.colips.org\/dstc5\/tasks.html\">DSTC5 website.<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/b><\/p>\n<h3>DSTC6<\/h3>\n<p>DSTC6 consisted of 3 parallel tracks: End-to-End Goal Oriented Dialog Learning, End-to-End Conversation Modeling, and Dialogue Breakdown Detection. Results will be presented at a workshop immediately after NIPS 2017. DSTC6 is organized by Chiori Hori, Julien Perez, Koichiro Yoshino, and Seokhwan Kim. Tracks were organized by Y-Lan Boureau, Antoine Bordes, Julien Perez, Ryuichi Higashinaka, Chiori Hori, and Takaaki Hori.<\/p>\n<p><b>For more information about DSTC6, see the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/workshop.colips.org\/dstc6\">DSTC6 website.<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/b><span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<p>To join the mailing list, please go to the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/sites.google.com\/dstc.community\/dstc8\/contact\" target=\"_blank\" rel=\"noopener\">current DTSC website<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.<\/p>\n<p>&nbsp;<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<h3>Summary of DSTC1, 2, and 3<\/h3>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/06\/williams2016dstc_overview-1.pdf\">The Dialog State Tracking Challenge Series: A Review<\/a>. Jason D. Williams, Antoine Raux, and Matthew Henderson. Dialogue & Discourse, April 2016.<\/p>\n<h3>DSTC1<\/h3>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/dstc2013.pdf\">The Dialog State Tracking Challenge<\/a>. Jason D. Williams, Antoine Raux, Deepak Ramachadran, and Alan Black. Proceedings of the SIGDIAL 2013 Conference, Metz, France, August 2013.<\/p>\n<h3>DSTC2<\/h3>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/summaryWriteup.pdf\">The Second Dialog State Tracking Challenge<\/a>. Matthew Henderson, Blaise Thomson, and Jason D. Williams. Proceedings of SIGDIAL 2014 Conference, Philadelphia, USA, June 2014.<\/p>\n<h3>DSTC3<\/h3>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/write_up.pdf\">The Third Dialog State Tracking Challenge<\/a>. Matthew Henderson, Blaise Thomson, and Jason D. Williams. Proceedings IEEE Spoken Language Technology Workshop (SLT), South Lake Tahoe, USA, December 2014.<\/p>\n<h3>DSTC4<\/h3>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/06\/dstc4_final-1.pdf\">The Fourth Dialog State Tracking Challenge<\/a>. Seokhwan Kim, Luis F. D&#8217;Haro, Rafael E Banchs, Matthew Henderson, and Jason D. Williams. Proceedings IEEE Spoken Language Technology Workshop (SLT), South Lake Tahoe, USA, December 2014.<\/p>\n<h3>DSTC5<\/h3>\n<p>The Fifth Dialog State Tracking Challenge. Seokhwan Kim, Luis F. D&#8217;Haro, Rafael E Banchs, Matthew Henderson, Jason D. Williams, and Koichiro Yoshino. Proceedings IEEE Spoken Language Technology Workshop (SLT), San Diego, USA, December 2016.<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<p><b>Background and motivation<\/b><\/p>\n<p>In dialog systems, &#8220;state tracking&#8221; \u2013 sometimes also called &#8220;belief tracking&#8221; \u2013 refers to accurately estimating the user\u2019s goal as a dialog progresses. Accurate state tracking is desirable because it provides robustness to errors in speech recognition, and helps reduce ambiguity inherent in language within a temporal process like dialog. Dialog state tracking is an important problem for both traditional uni-modal dialog systems, as well as speech-enabled multi-modal dialog systems on mobile devices, on tablet computers, and in automobiles.<\/p>\n<p>Recently, a host of models have been proposed for dialog state tracking. However, comparisons among models are rare, and different research groups use different data from disparate domains. Moreover, there is currently no common dataset which enables off-line dialog state tracking experiments, so newcomers to the area must first collect dialog data, which is expensive and time-consuming, or resort to simulated dialog data, which can be unreliable. All of these issues hinder advancing the state-of-the-art.<\/p>\n<p><b>Challenge format<\/b><\/p>\n<p>In this challenge, participants will use a provided set of labeled dialogs to develop a dialog state tracking algorithm. Algorithm will then be evaluated on a common set of held-out dialogs, to enable comparisons.<\/p>\n<p>The data for this challenge will be taken from the Spoken Dialog Challenge (Black et al, 2011), which consists of human\/machine spoken dialogs with real users (not usability subjects). Before the start of the challenge, a draft of the labeling guide and evaluation measures will be published, and comments will be invited from the community. The organizers will then perform the labeling.<\/p>\n<p>At the start of the challenge \u2013 the development phase \u2013 participants will be provided with a training set of transcribed and labeled dialogs. Participants will also be given code that implements the evaluation measurements. Participants will then have several months to optimize their algorithms.<\/p>\n<p>At the end of the challenge, participants will be given an untranscribed and unlabeled test set, and a short period to run their algorithm against the test set. Participants will submit their algorithms\u2019 output to the organizers, who will then perform the evaluation. After the challenge, the test set transcriptions and labels will be made public.<\/p>\n<p>Results of the evaluation will be reported to the community. In this challenge, the identities of participants will not be made public in any written results by the organizers or participants, except that participants may identify themselves (only) in their own written results.<\/p>\n<p><b>Schedule<\/b><\/p>\n<table style=\"width: 900px\">\n<tbody>\n<tr>\n<td>1 July 2012<\/td>\n<td>Beginning of comment period on labeling and evaluation metrics<\/td>\n<\/tr>\n<tr>\n<td>4-6 July 2012<\/td>\n<td>YRRSDS and SigDial in Korea: announcement of comment period<\/td>\n<\/tr>\n<tr>\n<td>17 August 2012<\/td>\n<td>End of comment period on labeling and evaluation metrics<\/td>\n<\/tr>\n<tr>\n<td>31 August 2012<\/td>\n<td>Evaluation metrics and labeling guide published; labeling begins<\/td>\n<\/tr>\n<tr>\n<td>10 December 2012<\/td>\n<td>Labeling ends; data available; challenge begins (14 weeks)<\/td>\n<\/tr>\n<tr>\n<td>3-5 December 2012<\/td>\n<td><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/slt2012.org\/\" target=\"_blank\">IEEE SLT 2012<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> (Miami): Information session, 5 Dec 3 PM<\/td>\n<\/tr>\n<tr>\n<td>22 March 2013<\/td>\n<td>Final system due; evaluation begins (1 week)<\/td>\n<\/tr>\n<tr>\n<td>29 March 2013<\/td>\n<td>Evaluation output due to organizers<\/td>\n<\/tr>\n<tr>\n<td>5 April 2013<\/td>\n<td>Results sent to teams<\/td>\n<\/tr>\n<tr>\n<td>3 May 2013<\/td>\n<td>SigDial paper deadline; write up (4 weeks)<\/td>\n<\/tr>\n<tr>\n<td>August 2013<\/td>\n<td>SigDial 2013<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p><b>Organizers<\/b><\/p>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jawillia\/\">Jason D. Williams<\/a>, Microsoft Research, USA (chair)<br \/>\nAlan Black, Carnegie Mellon University, USA<br \/>\nDeepak Ramachandran, Honda Research Institute, USA<br \/>\nAntoine Raux, Honda Research Institute, USA<\/p>\n<p><b>Advisory committee<\/b><\/p>\n<p>Daniel Boies, Microsoft, Canada<br \/>\nPaul Crook, Microsoft, USA<br \/>\nMaxine Eskenazi, Carnegie Mellon University, USA<br \/>\nMilica Gasic, University of Cambridge, UK<br \/>\nDilek Hakkani-Tur, Microsoft, USA<br \/>\nHelen Hastie, Heriot Watt University, UK<br \/>\nKee-Eung Kim, KAIST, Korea<br \/>\nIan Lane, Carnegie Mellon University, USA<br \/>\nSungjin Lee, Carnegie Mellon University, USA<br \/>\nTeruhisa Misu, NICT, Japan<br \/>\nOlivier Pietquin, SUPELEC, France<br \/>\nJoelle Pineau, McGill University, Canada<br \/>\nBlaise Thomson, University of Cambridge, UK<br \/>\nDavid Traum, USC Institute for Creative Technologies, USA<br \/>\nLuke Zettlemoyer, University of Washington, USA<\/p>\n<p><b>More information<\/b><\/p>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/dstc2013.pdf\">The Dialog State Tracking Challenge<\/a>. Jason D. Williams, Antoine Raux, Deepak Ramachadran, and Alan Black. Proceedings of the SIGDIAL 2013 Conference, Metz, France, August 2013.<\/p>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/Dialog20state20tracking20challenge20handbook20V21.pdf\">Dialog state tracking challenge handbook<\/a>. Antoine Raux, Deepak Ramachandran, Alan Black, and Jason D. Williams. Technical Report.<\/p>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/naaclhlt2012.pdf\">A belief tracking challenge task for spoken dialog systems<\/a>. Jason D. Williams. Proceedings NAACL HLT 2012 Workshop on Future directions and needs in the Spoken Dialog Community: Tools and Data, June 2012.<\/p>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/black2011sigdial.pdf\">Spoken Dialog Challenge 2010: Comparison of Live and Control Test Results<\/a>. Alan W Black, Susanne Burger, Alistair Conkie, Helen Hastie, Simon Keizer, Oliver Lemon, Nicolas Merigaud, Gabriel Parent, Gabriel Schubiner, Blaise Thomson, Jason D. Williams, Kai Yu, Steve Young, and Maxine Eskenazi. Proceedings SIGDIAL, Portland, Oregon, USA, 2011.<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<p><b>Training data<\/b><\/p>\n<ul>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/216.98.102.146:8080\/dstc\/train1a.tar.gz\">train1a<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/train1b.tar.gz\">train1b<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> (*)<\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/train1c.tar.gz\">train1c<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> (*)<\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/216.98.102.146:8080\/dstc\/train2.tar.gz\">train2<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/download.microsoft.com\/download\/B\/4\/5\/B4502E67-FBE2-417E-9530-E9A9A990E3F3\/dstc_data_train3_v00.tgz\">train3<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<\/ul>\n<p>(*) Note: train1b and train1c are larger sets of calls with transcriptions but WITHOUT correctness labels.<\/p>\n<p><b>Test data<\/b><\/p>\n<ul>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/test1.tar.gz\">test1<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/test2.tar.gz\">test2<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/download.microsoft.com\/download\/F\/0\/A\/F0AF24D9-34BC-4E17-9B0E-19E985F73DB7\/dstc_data_test3.tgz\">test3<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/test4.tar.gz\">test4<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<\/ul>\n<p><b>Support materials<\/b><\/p>\n<ul>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/download.microsoft.com\/download\/9\/7\/0\/9700B290-4449-4290-AAF0-DEED0C297F94\/dstc_scripts.tgz\">Baseline and evaluation scripts<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/tts.speech.cs.cmu.edu\/awb\/sdc2010_schedules\/\">Bus timetable database active summer 2010<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/Dialog20state20tracking20challenge20handbook20V21.pdf\">Handbook explaining data format<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<\/ul>\n<p><b>DSTC1 results<\/b><\/p>\n<ul>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/dstc-eval-results.zip\">Evaluation of tracker output<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/raw-results.zip\">Complete raw tracker output<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>The entries from team0 are the baselines<\/li>\n<\/ul>\n<p><span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"<p>The 2011 Dialog System Technology Challenge &#8212; formerly the Dialog State Tracking Challenge &#8212; is a series of research community challenge tasks for accurately estimating a user&#8217;s goal in a spoken dialog system.Opens in a new tab The Dialog State Tracking Challenge (DSTC) is an on-going series of research community challenge tasks. Each task released [&hellip;]<\/p>\n","protected":false},"featured_media":0,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr_startdate":"","msr_enddate":"","msr_location":"","msr_expirationdate":"","msr_event_recording_link":"","msr_event_link":"","msr_event_link_redirect":false,"msr_event_time":"","msr_hide_region":false,"msr_private_event":true,"msr_hide_image_in_river":0,"footnotes":""},"research-area":[],"msr-region":[],"msr-event-type":[],"msr-video-type":[],"msr-locale":[268875],"msr-program-audience":[],"msr-post-option":[],"msr-impact-theme":[],"class_list":["post-293411","msr-event","type-msr-event","status-publish","hentry","msr-locale-en_us"],"msr_about":"<!-- wp:msr\/event-details {\"title\":\"2011 Dialog System Technology Challenge (DSTC)\",\"backgroundColor\":\"grey\"} \/-->\n\n<!-- wp:msr\/content-tabs --><!-- wp:msr\/content-tab {\"title\":\"About\"} --><!-- wp:freeform --><p>The 2011 Dialog System Technology Challenge &#8212; formerly the Dialog State Tracking Challenge &#8212; is a series of research community challenge tasks for accurately estimating a user&#8217;s goal in a spoken dialog system.<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<p>The Dialog State Tracking Challenge (DSTC) is an on-going series of research community challenge tasks. Each task released dialog data labeled with dialog state information, such as the user&#8217;s desired restaurant search query given all of the dialog history up to the current turn. The challenge is to create a &#8220;tracker&#8221; that can predict the dialog state for new dialogs. In each challenge, trackers are evaluated using held-out dialog data.<\/p>\n<p>You can find information on the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/sites.google.com\/dstc.community\/dstc8\/\">current DTSC8 competition here<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.<\/p>\n<h3>DSTC1<\/h3>\n<p>DSTC1 used human-computer dialogs in the bus timetable domain. Results were presented in a special session at <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/www.sigdial.org\/workshops\/sigdial2013\/\">SIGDIAL 2013<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. DSTC1 was organized by Jason D. Williams, Alan Black, Deepak Ramachandran, Antoine Raux.<\/p>\n<p><b>For DSTC1 data, and more information, see &#8220;DSTC1&#8221; tabs above.<\/b><\/p>\n<h3>DSTC2 and DSTC3<\/h3>\n<p>DSTC2\/3 used human-computer dialogs in the restaurant information domain. Results were presented in special sessions at <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/www.sigdial.org\/workshops\/conference15\/\">SIGDIAL 2014<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> and <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/www.slt2014.org\/\">IEEE SLT 2014<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. DSTC2 and 3 were organized by Matthew Henderson, Blaise Thomson, and Jason D. Williams.<\/p>\n<p><b>For DSTC2 and DSTC3 data, and more information, see the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/camdial.org\/~mh521\/dstc\/\">DSTC2\/3 website.<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/b><\/p>\n<h3>DSTC4<\/h3>\n<p>DSTC4 used human-human dialogs in the tourist information domain. Results were presented at <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/www.iwsds.org\/\">IWSDS 2015<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. DSTC4 was organized by Seokhwan Kim, Luis F. D&#8217;Haro, Rafael E Banchs, Matthew Henderson, and Jason D. Williams.<\/p>\n<p><b>For more information about DSTC4, see the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/www.colips.org\/workshop\/dstc4\/\">DSTC4 website.<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/b><\/p>\n<h3>DSTC5<\/h3>\n<p>DSTC5 used human-human dialogs in the tourist information domain, where training dialogs were provided in one language, and test dialogs were in a different language. Results were presented in a special session at <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/www.slt2016.org\/\">IEEE SLT 2016<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. DSTC5 was organized by Seokhwan Kim, Luis F. D&#8217;Haro, Rafael E Banchs, Matthew Henderson, Jason D. Williams, and Koichiro Yoshino.<\/p>\n<p><b>For more information about DSTC5, see the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/workshop.colips.org\/dstc5\/tasks.html\">DSTC5 website.<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/b><\/p>\n<h3>DSTC6<\/h3>\n<p>DSTC6 consisted of 3 parallel tracks: End-to-End Goal Oriented Dialog Learning, End-to-End Conversation Modeling, and Dialogue Breakdown Detection. Results will be presented at a workshop immediately after NIPS 2017. DSTC6 is organized by Chiori Hori, Julien Perez, Koichiro Yoshino, and Seokhwan Kim. Tracks were organized by Y-Lan Boureau, Antoine Bordes, Julien Perez, Ryuichi Higashinaka, Chiori Hori, and Takaaki Hori.<\/p>\n<p><b>For more information about DSTC6, see the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/workshop.colips.org\/dstc6\">DSTC6 website.<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/b><span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<!-- \/wp:freeform --><!-- \/wp:msr\/content-tab --><!-- wp:msr\/content-tab {\"title\":\"Mailing list\"} --><!-- wp:freeform --><p>To join the mailing list, please go to the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/sites.google.com\/dstc.community\/dstc8\/contact\" target=\"_blank\" rel=\"noopener\">current DTSC website<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.<\/p>\n<p>&nbsp;<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<!-- \/wp:freeform --><!-- \/wp:msr\/content-tab --><!-- wp:msr\/content-tab {\"title\":\"Overview publications\"} --><!-- wp:freeform --><h3>Summary of DSTC1, 2, and 3<\/h3>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/06\/williams2016dstc_overview-1.pdf\">The Dialog State Tracking Challenge Series: A Review<\/a>. Jason D. Williams, Antoine Raux, and Matthew Henderson. Dialogue &amp; Discourse, April 2016.<\/p>\n<h3>DSTC1<\/h3>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/dstc2013.pdf\">The Dialog State Tracking Challenge<\/a>. Jason D. Williams, Antoine Raux, Deepak Ramachadran, and Alan Black. Proceedings of the SIGDIAL 2013 Conference, Metz, France, August 2013.<\/p>\n<h3>DSTC2<\/h3>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/summaryWriteup.pdf\">The Second Dialog State Tracking Challenge<\/a>. Matthew Henderson, Blaise Thomson, and Jason D. Williams. Proceedings of SIGDIAL 2014 Conference, Philadelphia, USA, June 2014.<\/p>\n<h3>DSTC3<\/h3>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/write_up.pdf\">The Third Dialog State Tracking Challenge<\/a>. Matthew Henderson, Blaise Thomson, and Jason D. Williams. Proceedings IEEE Spoken Language Technology Workshop (SLT), South Lake Tahoe, USA, December 2014.<\/p>\n<h3>DSTC4<\/h3>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/06\/dstc4_final-1.pdf\">The Fourth Dialog State Tracking Challenge<\/a>. Seokhwan Kim, Luis F. D&#8217;Haro, Rafael E Banchs, Matthew Henderson, and Jason D. Williams. Proceedings IEEE Spoken Language Technology Workshop (SLT), South Lake Tahoe, USA, December 2014.<\/p>\n<h3>DSTC5<\/h3>\n<p>The Fifth Dialog State Tracking Challenge. Seokhwan Kim, Luis F. D&#8217;Haro, Rafael E Banchs, Matthew Henderson, Jason D. Williams, and Koichiro Yoshino. Proceedings IEEE Spoken Language Technology Workshop (SLT), San Diego, USA, December 2016.<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<!-- \/wp:freeform --><!-- \/wp:msr\/content-tab --><!-- wp:msr\/content-tab {\"title\":\"DSTC1 information\"} --><!-- wp:freeform --><p><b>Background and motivation<\/b><\/p>\n<p>In dialog systems, &#8220;state tracking&#8221; \u2013 sometimes also called &#8220;belief tracking&#8221; \u2013 refers to accurately estimating the user\u2019s goal as a dialog progresses. Accurate state tracking is desirable because it provides robustness to errors in speech recognition, and helps reduce ambiguity inherent in language within a temporal process like dialog. Dialog state tracking is an important problem for both traditional uni-modal dialog systems, as well as speech-enabled multi-modal dialog systems on mobile devices, on tablet computers, and in automobiles.<\/p>\n<p>Recently, a host of models have been proposed for dialog state tracking. However, comparisons among models are rare, and different research groups use different data from disparate domains. Moreover, there is currently no common dataset which enables off-line dialog state tracking experiments, so newcomers to the area must first collect dialog data, which is expensive and time-consuming, or resort to simulated dialog data, which can be unreliable. All of these issues hinder advancing the state-of-the-art.<\/p>\n<p><b>Challenge format<\/b><\/p>\n<p>In this challenge, participants will use a provided set of labeled dialogs to develop a dialog state tracking algorithm. Algorithm will then be evaluated on a common set of held-out dialogs, to enable comparisons.<\/p>\n<p>The data for this challenge will be taken from the Spoken Dialog Challenge (Black et al, 2011), which consists of human\/machine spoken dialogs with real users (not usability subjects). Before the start of the challenge, a draft of the labeling guide and evaluation measures will be published, and comments will be invited from the community. The organizers will then perform the labeling.<\/p>\n<p>At the start of the challenge \u2013 the development phase \u2013 participants will be provided with a training set of transcribed and labeled dialogs. Participants will also be given code that implements the evaluation measurements. Participants will then have several months to optimize their algorithms.<\/p>\n<p>At the end of the challenge, participants will be given an untranscribed and unlabeled test set, and a short period to run their algorithm against the test set. Participants will submit their algorithms\u2019 output to the organizers, who will then perform the evaluation. After the challenge, the test set transcriptions and labels will be made public.<\/p>\n<p>Results of the evaluation will be reported to the community. In this challenge, the identities of participants will not be made public in any written results by the organizers or participants, except that participants may identify themselves (only) in their own written results.<\/p>\n<p><b>Schedule<\/b><\/p>\n<table style=\"width: 900px\">\n<tbody>\n<tr>\n<td>1 July 2012<\/td>\n<td>Beginning of comment period on labeling and evaluation metrics<\/td>\n<\/tr>\n<tr>\n<td>4-6 July 2012<\/td>\n<td>YRRSDS and SigDial in Korea: announcement of comment period<\/td>\n<\/tr>\n<tr>\n<td>17 August 2012<\/td>\n<td>End of comment period on labeling and evaluation metrics<\/td>\n<\/tr>\n<tr>\n<td>31 August 2012<\/td>\n<td>Evaluation metrics and labeling guide published; labeling begins<\/td>\n<\/tr>\n<tr>\n<td>10 December 2012<\/td>\n<td>Labeling ends; data available; challenge begins (14 weeks)<\/td>\n<\/tr>\n<tr>\n<td>3-5 December 2012<\/td>\n<td><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/slt2012.org\/\" target=\"_blank\">IEEE SLT 2012<\/a> (Miami): Information session, 5 Dec 3 PM<\/td>\n<\/tr>\n<tr>\n<td>22 March 2013<\/td>\n<td>Final system due; evaluation begins (1 week)<\/td>\n<\/tr>\n<tr>\n<td>29 March 2013<\/td>\n<td>Evaluation output due to organizers<\/td>\n<\/tr>\n<tr>\n<td>5 April 2013<\/td>\n<td>Results sent to teams<\/td>\n<\/tr>\n<tr>\n<td>3 May 2013<\/td>\n<td>SigDial paper deadline; write up (4 weeks)<\/td>\n<\/tr>\n<tr>\n<td>August 2013<\/td>\n<td>SigDial 2013<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p><b>Organizers<\/b><\/p>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jawillia\/\">Jason D. Williams<\/a>, Microsoft Research, USA (chair)<br \/>\nAlan Black, Carnegie Mellon University, USA<br \/>\nDeepak Ramachandran, Honda Research Institute, USA<br \/>\nAntoine Raux, Honda Research Institute, USA<\/p>\n<p><b>Advisory committee<\/b><\/p>\n<p>Daniel Boies, Microsoft, Canada<br \/>\nPaul Crook, Microsoft, USA<br \/>\nMaxine Eskenazi, Carnegie Mellon University, USA<br \/>\nMilica Gasic, University of Cambridge, UK<br \/>\nDilek Hakkani-Tur, Microsoft, USA<br \/>\nHelen Hastie, Heriot Watt University, UK<br \/>\nKee-Eung Kim, KAIST, Korea<br \/>\nIan Lane, Carnegie Mellon University, USA<br \/>\nSungjin Lee, Carnegie Mellon University, USA<br \/>\nTeruhisa Misu, NICT, Japan<br \/>\nOlivier Pietquin, SUPELEC, France<br \/>\nJoelle Pineau, McGill University, Canada<br \/>\nBlaise Thomson, University of Cambridge, UK<br \/>\nDavid Traum, USC Institute for Creative Technologies, USA<br \/>\nLuke Zettlemoyer, University of Washington, USA<\/p>\n<p><b>More information<\/b><\/p>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/dstc2013.pdf\">The Dialog State Tracking Challenge<\/a>. Jason D. Williams, Antoine Raux, Deepak Ramachadran, and Alan Black. Proceedings of the SIGDIAL 2013 Conference, Metz, France, August 2013.<\/p>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/Dialog20state20tracking20challenge20handbook20V21.pdf\">Dialog state tracking challenge handbook<\/a>. Antoine Raux, Deepak Ramachandran, Alan Black, and Jason D. Williams. Technical Report.<\/p>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/naaclhlt2012.pdf\">A belief tracking challenge task for spoken dialog systems<\/a>. Jason D. Williams. Proceedings NAACL HLT 2012 Workshop on Future directions and needs in the Spoken Dialog Community: Tools and Data, June 2012.<\/p>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/black2011sigdial.pdf\">Spoken Dialog Challenge 2010: Comparison of Live and Control Test Results<\/a>. Alan W Black, Susanne Burger, Alistair Conkie, Helen Hastie, Simon Keizer, Oliver Lemon, Nicolas Merigaud, Gabriel Parent, Gabriel Schubiner, Blaise Thomson, Jason D. Williams, Kai Yu, Steve Young, and Maxine Eskenazi. Proceedings SIGDIAL, Portland, Oregon, USA, 2011.<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<!-- \/wp:freeform --><!-- \/wp:msr\/content-tab --><!-- wp:msr\/content-tab {\"title\":\"DSTC1 downloads\"} --><!-- wp:freeform --><p><b>Training data<\/b><\/p>\n<ul>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/216.98.102.146:8080\/dstc\/train1a.tar.gz\">train1a<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/train1b.tar.gz\">train1b<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> (*)<\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/train1c.tar.gz\">train1c<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> (*)<\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/216.98.102.146:8080\/dstc\/train2.tar.gz\">train2<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/download.microsoft.com\/download\/B\/4\/5\/B4502E67-FBE2-417E-9530-E9A9A990E3F3\/dstc_data_train3_v00.tgz\">train3<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<\/ul>\n<p>(*) Note: train1b and train1c are larger sets of calls with transcriptions but WITHOUT correctness labels.<\/p>\n<p><b>Test data<\/b><\/p>\n<ul>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/test1.tar.gz\">test1<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/test2.tar.gz\">test2<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/download.microsoft.com\/download\/F\/0\/A\/F0AF24D9-34BC-4E17-9B0E-19E985F73DB7\/dstc_data_test3.tgz\">test3<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/test4.tar.gz\">test4<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<\/ul>\n<p><b>Support materials<\/b><\/p>\n<ul>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/download.microsoft.com\/download\/9\/7\/0\/9700B290-4449-4290-AAF0-DEED0C297F94\/dstc_scripts.tgz\">Baseline and evaluation scripts<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/tts.speech.cs.cmu.edu\/awb\/sdc2010_schedules\/\">Bus timetable database active summer 2010<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/Dialog20state20tracking20challenge20handbook20V21.pdf\">Handbook explaining data format<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<\/ul>\n<p><b>DSTC1 results<\/b><\/p>\n<ul>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/dstc-eval-results.zip\">Evaluation of tracker output<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/raw-results.zip\">Complete raw tracker output<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>The entries from team0 are the baselines<\/li>\n<\/ul>\n<p><span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<!-- \/wp:freeform --><!-- \/wp:msr\/content-tab --><!-- \/wp:msr\/content-tabs -->","tab-content":[{"id":0,"name":"About","content":"The Dialog State Tracking Challenge (DSTC) is an on-going series of research community challenge tasks. Each task released dialog data labeled with dialog state information, such as the user's desired restaurant search query given all of the dialog history up to the current turn. The challenge is to create a \"tracker\" that can predict the dialog state for new dialogs. In each challenge, trackers are evaluated using held-out dialog data.\r\n\r\nYou can find information on the <a href=\"https:\/\/sites.google.com\/dstc.community\/dstc8\/\">current DTSC8 competition here<\/a>.\r\n<h3>DSTC1<\/h3>\r\nDSTC1 used human-computer dialogs in the bus timetable domain. Results were presented in a special session at <a href=\"http:\/\/www.sigdial.org\/workshops\/sigdial2013\/\">SIGDIAL 2013<\/a>. DSTC1 was organized by Jason D. Williams, Alan Black, Deepak Ramachandran, Antoine Raux.\r\n\r\n<b>For DSTC1 data, and more information, see \"DSTC1\" tabs above.<\/b>\r\n<h3>DSTC2 and DSTC3<\/h3>\r\nDSTC2\/3 used human-computer dialogs in the restaurant information domain. Results were presented in special sessions at <a href=\"http:\/\/www.sigdial.org\/workshops\/conference15\/\">SIGDIAL 2014<\/a> and <a href=\"http:\/\/www.slt2014.org\/\">IEEE SLT 2014<\/a>. DSTC2 and 3 were organized by Matthew Henderson, Blaise Thomson, and Jason D. Williams.\r\n\r\n<b>For DSTC2 and DSTC3 data, and more information, see the <a href=\"http:\/\/camdial.org\/~mh521\/dstc\/\">DSTC2\/3 website.<\/a><\/b>\r\n<h3>DSTC4<\/h3>\r\nDSTC4 used human-human dialogs in the tourist information domain. Results were presented at <a href=\"http:\/\/www.iwsds.org\/\">IWSDS 2015<\/a>. DSTC4 was organized by Seokhwan Kim, Luis F. D'Haro, Rafael E Banchs, Matthew Henderson, and Jason D. Williams.\r\n\r\n<b>For more information about DSTC4, see the <a href=\"http:\/\/www.colips.org\/workshop\/dstc4\/\">DSTC4 website.<\/a><\/b>\r\n<h3>DSTC5<\/h3>\r\nDSTC5 used human-human dialogs in the tourist information domain, where training dialogs were provided in one language, and test dialogs were in a different language. Results were presented in a special session at <a href=\"http:\/\/www.slt2016.org\/\">IEEE SLT 2016<\/a>. DSTC5 was organized by Seokhwan Kim, Luis F. D'Haro, Rafael E Banchs, Matthew Henderson, Jason D. Williams, and Koichiro Yoshino.\r\n\r\n<b>For more information about DSTC5, see the <a href=\"http:\/\/workshop.colips.org\/dstc5\/tasks.html\">DSTC5 website.<\/a><\/b>\r\n<h3>DSTC6<\/h3>\r\nDSTC6 consisted of 3 parallel tracks: End-to-End Goal Oriented Dialog Learning, End-to-End Conversation Modeling, and Dialogue Breakdown Detection. Results will be presented at a workshop immediately after NIPS 2017. DSTC6 is organized by Chiori Hori, Julien Perez, Koichiro Yoshino, and Seokhwan Kim. Tracks were organized by Y-Lan Boureau, Antoine Bordes, Julien Perez, Ryuichi Higashinaka, Chiori Hori, and Takaaki Hori.\r\n\r\n<b>For more information about DSTC6, see the <a href=\"http:\/\/workshop.colips.org\/dstc6\">DSTC6 website.<\/a><\/b>"},{"id":1,"name":"Mailing list","content":"To join the mailing list, please go to the <a href=\"https:\/\/sites.google.com\/dstc.community\/dstc8\/contact\" target=\"_blank\" rel=\"noopener\">current DTSC website<\/a>.\r\n\r\n&nbsp;"},{"id":2,"name":"Overview publications","content":"<h3>Summary of DSTC1, 2, and 3<\/h3>\r\n<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/06\/williams2016dstc_overview-1.pdf\">The Dialog State Tracking Challenge Series: A Review<\/a>. Jason D. Williams, Antoine Raux, and Matthew Henderson. Dialogue &amp; Discourse, April 2016.\r\n<h3>DSTC1<\/h3>\r\n<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/dstc2013.pdf\">The Dialog State Tracking Challenge<\/a>. Jason D. Williams, Antoine Raux, Deepak Ramachadran, and Alan Black. Proceedings of the SIGDIAL 2013 Conference, Metz, France, August 2013.\r\n<h3>DSTC2<\/h3>\r\n<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/summaryWriteup.pdf\">The Second Dialog State Tracking Challenge<\/a>. Matthew Henderson, Blaise Thomson, and Jason D. Williams. Proceedings of SIGDIAL 2014 Conference, Philadelphia, USA, June 2014.\r\n<h3>DSTC3<\/h3>\r\n<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/write_up.pdf\">The Third Dialog State Tracking Challenge<\/a>. Matthew Henderson, Blaise Thomson, and Jason D. Williams. Proceedings IEEE Spoken Language Technology Workshop (SLT), South Lake Tahoe, USA, December 2014.\r\n<h3>DSTC4<\/h3>\r\n<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/06\/dstc4_final-1.pdf\">The Fourth Dialog State Tracking Challenge<\/a>. Seokhwan Kim, Luis F. D'Haro, Rafael E Banchs, Matthew Henderson, and Jason D. Williams. Proceedings IEEE Spoken Language Technology Workshop (SLT), South Lake Tahoe, USA, December 2014.\r\n<h3>DSTC5<\/h3>\r\nThe Fifth Dialog State Tracking Challenge. Seokhwan Kim, Luis F. D'Haro, Rafael E Banchs, Matthew Henderson, Jason D. Williams, and Koichiro Yoshino. Proceedings IEEE Spoken Language Technology Workshop (SLT), San Diego, USA, December 2016."},{"id":3,"name":"DSTC1 information","content":"<b>Background and motivation<\/b>\r\n\r\nIn dialog systems, \"state tracking\" \u2013 sometimes also called \"belief tracking\" \u2013 refers to accurately estimating the user\u2019s goal as a dialog progresses. Accurate state tracking is desirable because it provides robustness to errors in speech recognition, and helps reduce ambiguity inherent in language within a temporal process like dialog. Dialog state tracking is an important problem for both traditional uni-modal dialog systems, as well as speech-enabled multi-modal dialog systems on mobile devices, on tablet computers, and in automobiles.\r\n\r\nRecently, a host of models have been proposed for dialog state tracking. However, comparisons among models are rare, and different research groups use different data from disparate domains. Moreover, there is currently no common dataset which enables off-line dialog state tracking experiments, so newcomers to the area must first collect dialog data, which is expensive and time-consuming, or resort to simulated dialog data, which can be unreliable. All of these issues hinder advancing the state-of-the-art.\r\n\r\n<b>Challenge format<\/b>\r\n\r\nIn this challenge, participants will use a provided set of labeled dialogs to develop a dialog state tracking algorithm. Algorithm will then be evaluated on a common set of held-out dialogs, to enable comparisons.\r\n\r\nThe data for this challenge will be taken from the Spoken Dialog Challenge (Black et al, 2011), which consists of human\/machine spoken dialogs with real users (not usability subjects). Before the start of the challenge, a draft of the labeling guide and evaluation measures will be published, and comments will be invited from the community. The organizers will then perform the labeling.\r\n\r\nAt the start of the challenge \u2013 the development phase \u2013 participants will be provided with a training set of transcribed and labeled dialogs. Participants will also be given code that implements the evaluation measurements. Participants will then have several months to optimize their algorithms.\r\n\r\nAt the end of the challenge, participants will be given an untranscribed and unlabeled test set, and a short period to run their algorithm against the test set. Participants will submit their algorithms\u2019 output to the organizers, who will then perform the evaluation. After the challenge, the test set transcriptions and labels will be made public.\r\n\r\nResults of the evaluation will be reported to the community. In this challenge, the identities of participants will not be made public in any written results by the organizers or participants, except that participants may identify themselves (only) in their own written results.\r\n\r\n<b>Schedule<\/b>\r\n<table style=\"width: 900px\">\r\n<tbody>\r\n<tr>\r\n<td>1 July 2012<\/td>\r\n<td>Beginning of comment period on labeling and evaluation metrics<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>4-6 July 2012<\/td>\r\n<td>YRRSDS and SigDial in Korea: announcement of comment period<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>17 August 2012<\/td>\r\n<td>End of comment period on labeling and evaluation metrics<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>31 August 2012<\/td>\r\n<td>Evaluation metrics and labeling guide published; labeling begins<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>10 December 2012<\/td>\r\n<td>Labeling ends; data available; challenge begins (14 weeks)<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>3-5 December 2012<\/td>\r\n<td><a href=\"http:\/\/slt2012.org\/\" target=\"_self\">IEEE SLT 2012<\/a> (Miami): Information session, 5 Dec 3 PM<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>22 March 2013<\/td>\r\n<td>Final system due; evaluation begins (1 week)<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>29 March 2013<\/td>\r\n<td>Evaluation output due to organizers<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>5 April 2013<\/td>\r\n<td>Results sent to teams<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>3 May 2013<\/td>\r\n<td>SigDial paper deadline; write up (4 weeks)<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>August 2013<\/td>\r\n<td>SigDial 2013<\/td>\r\n<\/tr>\r\n<\/tbody>\r\n<\/table>\r\n<b>Organizers<\/b>\r\n\r\n<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jawillia\/\">Jason D. Williams<\/a>, Microsoft Research, USA (chair)\r\nAlan Black, Carnegie Mellon University, USA\r\nDeepak Ramachandran, Honda Research Institute, USA\r\nAntoine Raux, Honda Research Institute, USA\r\n\r\n<b>Advisory committee<\/b>\r\n\r\nDaniel Boies, Microsoft, Canada\r\nPaul Crook, Microsoft, USA\r\nMaxine Eskenazi, Carnegie Mellon University, USA\r\nMilica Gasic, University of Cambridge, UK\r\nDilek Hakkani-Tur, Microsoft, USA\r\nHelen Hastie, Heriot Watt University, UK\r\nKee-Eung Kim, KAIST, Korea\r\nIan Lane, Carnegie Mellon University, USA\r\nSungjin Lee, Carnegie Mellon University, USA\r\nTeruhisa Misu, NICT, Japan\r\nOlivier Pietquin, SUPELEC, France\r\nJoelle Pineau, McGill University, Canada\r\nBlaise Thomson, University of Cambridge, UK\r\nDavid Traum, USC Institute for Creative Technologies, USA\r\nLuke Zettlemoyer, University of Washington, USA\r\n\r\n<b>More information<\/b>\r\n\r\n<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/dstc2013.pdf\">The Dialog State Tracking Challenge<\/a>. Jason D. Williams, Antoine Raux, Deepak Ramachadran, and Alan Black. Proceedings of the SIGDIAL 2013 Conference, Metz, France, August 2013.\r\n\r\n<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/Dialog20state20tracking20challenge20handbook20V21.pdf\">Dialog state tracking challenge handbook<\/a>. Antoine Raux, Deepak Ramachandran, Alan Black, and Jason D. Williams. Technical Report.\r\n\r\n<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/naaclhlt2012.pdf\">A belief tracking challenge task for spoken dialog systems<\/a>. Jason D. Williams. Proceedings NAACL HLT 2012 Workshop on Future directions and needs in the Spoken Dialog Community: Tools and Data, June 2012.\r\n\r\n<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/black2011sigdial.pdf\">Spoken Dialog Challenge 2010: Comparison of Live and Control Test Results<\/a>. Alan W Black, Susanne Burger, Alistair Conkie, Helen Hastie, Simon Keizer, Oliver Lemon, Nicolas Merigaud, Gabriel Parent, Gabriel Schubiner, Blaise Thomson, Jason D. Williams, Kai Yu, Steve Young, and Maxine Eskenazi. Proceedings SIGDIAL, Portland, Oregon, USA, 2011."},{"id":4,"name":"DSTC1 downloads","content":"<b>Training data<\/b>\r\n<ul>\r\n \t<li><a href=\"http:\/\/216.98.102.146:8080\/dstc\/train1a.tar.gz\">train1a<\/a><\/li>\r\n \t<li><a href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/train1b.tar.gz\">train1b<\/a> (*)<\/li>\r\n \t<li><a href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/train1c.tar.gz\">train1c<\/a> (*)<\/li>\r\n \t<li><a href=\"http:\/\/216.98.102.146:8080\/dstc\/train2.tar.gz\">train2<\/a><\/li>\r\n \t<li><a href=\"https:\/\/download.microsoft.com\/download\/B\/4\/5\/B4502E67-FBE2-417E-9530-E9A9A990E3F3\/dstc_data_train3_v00.tgz\">train3<\/a><\/li>\r\n<\/ul>\r\n(*) Note: train1b and train1c are larger sets of calls with transcriptions but WITHOUT correctness labels.\r\n\r\n<b>Test data<\/b>\r\n<ul>\r\n \t<li><a href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/test1.tar.gz\">test1<\/a><\/li>\r\n \t<li><a href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/test2.tar.gz\">test2<\/a><\/li>\r\n \t<li><a href=\"https:\/\/download.microsoft.com\/download\/F\/0\/A\/F0AF24D9-34BC-4E17-9B0E-19E985F73DB7\/dstc_data_test3.tgz\">test3<\/a><\/li>\r\n \t<li><a href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/test4.tar.gz\">test4<\/a><\/li>\r\n<\/ul>\r\n<b>Support materials<\/b>\r\n<ul>\r\n \t<li><a href=\"https:\/\/download.microsoft.com\/download\/9\/7\/0\/9700B290-4449-4290-AAF0-DEED0C297F94\/dstc_scripts.tgz\">Baseline and evaluation scripts<\/a><\/li>\r\n \t<li><a href=\"http:\/\/tts.speech.cs.cmu.edu\/awb\/sdc2010_schedules\/\">Bus timetable database active summer 2010<\/a><\/li>\r\n \t<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2016\/02\/Dialog20state20tracking20challenge20handbook20V21.pdf\">Handbook explaining data format<\/a><\/li>\r\n<\/ul>\r\n<b>DSTC1 results<\/b>\r\n<ul>\r\n \t<li><a href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/dstc-eval-results.zip\">Evaluation of tracker output<\/a><\/li>\r\n \t<li><a href=\"http:\/\/dogmatix.honda-ri.com:8080\/dstc\/raw-results.zip\">Complete raw tracker output<\/a>The entries from team0 are the baselines<\/li>\r\n<\/ul>"}],"msr_startdate":"","msr_enddate":"","msr_event_time":"","msr_location":"","msr_event_link":"","msr_event_recording_link":"","msr_startdate_formatted":"","msr_register_text":"Register now","msr_cta_link":"","msr_cta_text":"","msr_cta_bi_name":"","featured_image_thumbnail":null,"event_excerpt":"The Dialog State Tracking Challenge (DSTC) is an on-going series of research community challenge tasks. Each task released dialog data labeled with dialog state information, such as the user's desired restaurant search query given all of the dialog history up to the current turn. The challenge is to create a \"tracker\" that can predict the dialog state for new dialogs. In each challenge, trackers are evaluated using held-out dialog data. You can find information on&hellip;","msr_research_lab":[],"related-researchers":[],"msr_impact_theme":[],"related-academic-programs":[],"related-groups":[390593],"related-projects":[],"related-opportunities":[],"related-publications":[],"related-videos":[],"related-posts":[],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/293411","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-event"}],"version-history":[{"count":5,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/293411\/revisions"}],"predecessor-version":[{"id":1147253,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/293411\/revisions\/1147253"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=293411"}],"wp:term":[{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=293411"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=293411"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=293411"},{"taxonomy":"msr-video-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-video-type?post=293411"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=293411"},{"taxonomy":"msr-program-audience","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-program-audience?post=293411"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=293411"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=293411"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}