Microsoft Research Blog

Artificial intelligence

  1. End-to-End NLP Knowledge Graph Construction 

    June 1, 2021 | Ishani Mondal, Yufang Hou, and Charles Jochim

    This paper studies the end-to-end construction of an NLP Knowledge Graph (KG) from scientific papers. We focus on extracting four types of relations: evaluatedOn between tasks and datasets, evaluatedBy between tasks and evaluation metrics, as well as coreferent and related relations between the same type…

  2. Hi-Transformer: Hierarchical Interactive Transformer for Efficient and Effective Long Document Modeling 

    June 1, 2021 | Chuhan Wu, Fangzhao Wu, Tao Qi, and Yongfeng Huang

    Transformer is important for text modeling. However, it has difficulty in handling long documents due to the quadratic complexity with input text length. In order to handle this problem, we propose a hierarchical interactive Transformer (Hi-Transformer) for efficient and effective long document modeling. Hi-Transformer models…

  3. SocAoG: Incremental Graph Parsing for Social Relation Inference in Dialogues 

    June 1, 2021

    Inferring social relations from dialogues is vital for building emotionally intelligent robots to interpret human language better and act accordingly. We model the social network as an And-or Graph, named SocAoG, for the consistency of relations among a group and leveraging attributes as inference cues.…

  4. Addressing the Exorbitant Cost of Labeling Medical Images with Active Learning 

    June 1, 2021 | Saba Rahimi, Ozan Oktay, Javier Alvarez-Valle, and Sujeeth Bharadwaj

    Successful application of deep learning in medical image analysis necessitates unprecedented amounts of labeled training data. Unlike conventional 2D applications, radiological images can be three dimensional (e.g. CT, MRI) consisting of many instances within each image. The problem is exacerbated when expert annotations are required…

  5. Structure-Grounded Pretraining for Text-to-SQL 

    June 1, 2021

    Learning to capture text-table alignment is essential for table related tasks like text-to-SQL. The model needs to correctly recognize natural language references to columns and values and to ground them in the given database schema. In this paper, we present a novel weakly supervised Structure-Grounded…

  6. Verbal Focus-of-Attention System for Learning-from-Observation 

    May 29, 2021 | Naoki Wake, Iori Yanokura, Kazuhiro Sasabuchi, and Katsushi Ikeuchi

    The learning-from-observation (LfO) framework aims to map human demonstrations to a robot to reduce programming effort. To this end, an LfO system encodes a human demonstration into a series of execution units for a robot, which are referred to as task models. Although previous research…

  7. SAMF: a Self-adaptive Protein Modeling Framework 

    May 27, 2021

    MOTIVATION Gradient descent-based protein modeling is a popular protein structure prediction approach that takes as input the predicted inter-residue distances and other necessary constraints and folds protein structures by minimizing protein-specific energy potentials. The constraints from multiple predicted protein properties provide redundant and sometime conflicting…

  8. Self-Supervised Bug Detection and Repair 

    May 26, 2021 | Miltos Allamanis, Henry Jackson-Flux, and Marc Brockschmidt

    Machine learning-based program analyses have recently shown the promise of integrating formal and probabilistic reasoning towards aiding software development. However, in the absence of large annotated corpora, training these analyses is challenging. Towards addressing this, we present BugLab, an approach for self-supervised learning of bug…

  9. How Does Distilled Data Complexity Impact the Quality and Confidence of Non-Autoregressive Machine Translation? 

    May 26, 2021 | Weijia Xu, Shuming Ma, Dongdong Zhang, and Marine Carpuat

    While non-autoregressive (NAR) models are showing great promise for machine translation, their use is limited by their dependence on knowledge distillation from autoregressive models. To address this issue, we seek to understand why distillation is so effective. Prior work suggests that distilled training data is…

  10. Compositional Processing Emerges in Neural Networks Solving Math Problems 

    May 18, 2021

    A longstanding question in cognitive science concerns the learning mechanisms underlying compositionality in human cognition. Humans can infer the structured relationships (e.g., grammatical rules) implicit in their sensory observations (e.g., auditory speech), and use this knowledge to guide the composition of simpler meanings into complex…