Microsoft Research Blog

Artificial intelligence

  1. AL: autogenerating supervised learning programs 

    October 9, 2019 | José Cambronero and Martin C. Rinard

    We present AL, a novel automated machine learning system that learns to generate new supervised learning pipelines from an existing corpus of supervised learning programs. In contrast to existing automated machine learning tools, which typically implement a search over manually selected machine learning functions and…

  2. We Know What You Will Ask: A Dialogue System for Multi-intent Switch and Prediction 

    October 9, 2019

    Existing task-oriented dialogue systems seldom emphasize multi-intent scenarios, which makes them hard to track complex intent switch in a multi-turn dialogue, and even harder to make proactive reactions for the user’s next potential intent. In this paper, we formalize the multi-intent tracking task and introduce…

  3. Characterizing Developer Use of Automatically Generated Patches 

    September 30, 2019

    We present a study that characterizes the way developers use automatically generated patches when fixing software defects. Our study tasked two groups of developers with repairing defects in C programs. Both groups were provided with the defective line of code. One was also provided with…

  4. Laplacian using Abstract State Transition Graphs: A Framework for Skill Acquisition 

    September 30, 2019 | Matheus R. F. Mendonça, Artur Ziviani, and André da Motta Salles Barreto

    Automatic definition of macro-actions for Reinforcement Learning (RL) is a way of breaking a large problem into smaller sub-problems. Macro-actions are known to boost the agent's learning process, leading to a better performance. One recent approach, called Laplacian Framework, uses the Proto-Value Functions of the…

  5. Beyond Word for Word: Fact Guided Training for Neural Data-to-Document Generation 

    September 30, 2019

    Recent end-to-end encoder-decoder neural models for data-to-text generation can produce fluent and seemingly informative texts despite these models disregard the traditional content selection and surface realization architecture. However, texts generated by such neural models are often missing important facts and contradict the input data, particularly…

  6. CodeSearchNet Challenge: Evaluating the State of Semantic Code Search. 

    September 20, 2019

    Semantic code search is the task of retrieving relevant code given a natural language query. While related to other information retrieval tasks, it requires bridging the gap between the language used in code (often abbreviated and highly technical) and natural language more suitable to describe…

  7. A Coarse-to-Fine Framework for Learned Color Enhancement with Non-Local Attention 

    September 1, 2019 | Chaowei Shan, Zhizheng Zhang, and Zhibo Chen

    Automatic color enhancement are aimed to automaticly and adaptively adjust photos to expected styles and tones. For current learned methods in this field, global harmonious perception and local details are hard to be well-considered in a single model simultaneously. To address this problem, we propose…

  8. Learning How to Mutate Source Code from Bug-Fixes 

    August 31, 2019

    Mutation testing has been widely accepted as an approach to guide test case generation or to assess the effectiveness of test suites. Empirical studies have shown that mutants are representative of real faults; yet they also indicated a clear need for better, possibly customized, mutation…

  9. Patient Knowledge Distillation for BERT Model Compression 

    August 24, 2019 | Siqi Sun, Yu Cheng, Zhe Gan, and Jingjing Liu

    Pre-trained language models such as BERT have proven to be highly effective for natural language processing (NLP) tasks. However, the high demand for computing resources in training such models hinders their application in practice. In order to alleviate this resource hunger in large-scale model training,…

  10. Parameter-free Sentence Embedding via Orthogonal Basis. 

    August 14, 2019 | Ziyi Yang, Chenguang Zhu, and Weizhu Chen

    We propose a simple and robust non-parameterized approach for building sentence representations. Inspired by the Gram-Schmidt Process in geometric theory, we build an orthogonal basis of the subspace spanned by a word and its surrounding context in a sentence. We model the semantic meaning of…