Microsoft Research Blog

Artificial intelligence

  1. Estimating GPU Memory Consumption of Deep Learning Models 

    November 7, 2020

    Deep learning (DL) has been increasingly adopted by a variety of software-intensive systems. Developers mainly use GPUs to accelerate the training, testing, and deployment of DL models. However, the GPU memory consumed by a DL model is often unknown to them before a DL job…

  2. Identifying linked incidents in large-scale online service systems 

    November 7, 2020

    In large-scale online service systems, incidents occur frequently due to a variety of causes, from updates of software and hardware to changes in operation environment. These incidents could significantly degrade system’s availability and customers’ satisfaction. Some incidents are linked because they are duplicate or inter-related.…

  3. A Simple Approach to Learning Unsupervised Multilingual Embeddings 

    November 1, 2020 | Pratik Jawanpuria, Mayank Meghwanshi, and Bamdev Mishra

    Recent progress on unsupervised cross-lingual embeddings in the bilingual setting has given the impetus to learning a shared embedding space for several languages. A popular framework to solve the latter problem is to solve the following two sub-problems jointly: 1) learning unsupervised word alignment between…

  4. Few-Shot Induction of Generalized Logical Concepts via Human Guidance 

    November 1, 2020 | Mayukh Das, Nandini Ramanan, Janardhan Rao Doppa, and Sriraam Natarajan

    We consider the problem of learning generalized first-order representations of concepts from a small number of examples. We augment an inductive logic programming learner with two novel contributions. First, we define a distance measure between candidate concept representations that improves the efficiency of search for…

  5. Routing Enforced Generative Model for Recipe Generation. 

    November 1, 2020 | Zhiwei Yu, Hongyu Zang, and Xiaojun Wan

    One of the most challenging part of recipe generation is to deal with the complex restrictions among the input ingredients. Previous researches simplify the problem by treating the inputs independently and generating recipes containing as much information as possible. In this work, we propose a…

  6. Homophonic Pun Generation with Lexically Constrained Rewriting. 

    November 1, 2020 | Zhiwei Yu, Hongyu Zang, and Xiaojun Wan

    Punning is a creative way to make conversation enjoyable and literary writing elegant. In this paper, we focus on the task of generating a pun sentence given a pair of homophones. We first find the constraint words supporting the semantic incongruity for a sentence. Then…

  7. Long Document Ranking with Query-Directed Sparse Transformer 

    October 31, 2020 | Jyun-Yu Jiang, Chenyan Xiong, Chia-Jung Lee, and Wei Wang

    The computing cost of transformer self-attention often necessitates breaking long documents to fit in pretrained models in document ranking tasks. In this paper, we design Query-Directed Sparse attention that induces IR-axiomatic structures in transformer self-attention. Our model, QDS-Transformer, enforces the principle properties desired in ranking:…

  8. Depth Completion Using a View-constrained Deep Prior 

    October 31, 2020

    Recent work has shown that the structure of convolutional neural networks (CNNs) induces a strong prior that favors natural images. This prior, known as a deep image prior (DIP), is an effective regularizer in inverse problems such as image denoising and inpainting. We extend the…