Microsoft Research Blog

Artificial intelligence

  1. Label super-resolution networks 

    September 26, 2018

    We present a deep learning-based method for super-resolving coarse (low-resolution) labels assigned to groups of image pixels into pixel-level (high-resolution) labels, given the joint distribution between those low- and high-resolution labels. This method involves a novel loss function that minimizes the distance between a distribution…

  2. Algorithmically Generated Domain Detection and Malware Family Classification 

    September 18, 2018

    In this paper, we compare the performance of several machine learning based approaches for the tasks of detecting algorithmically generated malicious domains and the categorization of domains according to their malware family. The datasets used for model comparison were provided by the shared task on…

  3. Autofocus Layer for Semantic Segmentation 

    September 15, 2018

    We propose the autofocus convolutional layer for semantic segmentation with the objective of enhancing the capabilities of neural networks for multi-scale processing. Autofocus layers adaptively change the size of the effective receptive field based on the processed context to generate more powerful features. This is…

  4. Operation-guided Neural Networks for High Fidelity Data-To-Text Generation 

    September 8, 2018

    Recent neural models for data-to-text generation are mostly based on data-driven end-to-end training over encoder-decoder networks. Even though the generated texts are mostly fluent and informative, they often generate descriptions that are not consistent with the input structured data. This is a critical issue especially…

  5. Integral Human Pose Regression 

    September 7, 2018

    State-of-the-art human pose estimation methods are based on heat map representation. In spite of the good performance, the representation has a few issues in nature, such as non-differentiable post-processing and quantization error. This work shows that a simple integral operation relates and unifies the heat…

  6. An Argument in Favor of Strong Scaling for Deep Neural Networks with Small Datasets 

    August 31, 2018 | Renato L. de F. Cunha, Eduardo Rocha Rodrigues, Matheus Palhares Viana, and Dario Augusto Borges Oliveira

    In recent years, with the popularization of deep learning frameworks and large datasets, researchers have started parallelizing their models in order to train faster. This is crucially important, because they typically explore many hyperparameters in order to find the best ones for their applications. This…

  7. DeepDownscale: A Deep Learning Strategy for High-Resolution Weather Forecast 

    August 14, 2018 | Eduardo Rocha Rodrigues, Igor Oliveira, Renato L. de F. Cunha, and Marco A.S. Netto

    Running high-resolution physical models is computationally expensive and essential for many disciplines. Agriculture, transportation, and energy are sectors that depend on high-resolution weather models, which typically consume many hours of large High Performance Computing (HPC) systems to deliver timely results. Many users cannot afford to…

  8. Deep exemplar-based colorization 

    August 9, 2018

    We propose the first deep learning approach for exemplar-based local colorization. Given a reference color image, our convolutional neural network directly maps a grayscale image to an output colorized image. Rather than using hand-crafted rules as in traditional exemplar-based methods, our end-to-end colorization network learns…

  9. Deep exemplar-based colorization 

    August 9, 2018

    We propose the first deep learning approach for exemplar-based local colorization. Given a reference color image, our convolutional neural network directly maps a grayscale image to an output colorized image. Rather than using hand-crafted rules as in traditional exemplar-based methods, our end-to-end colorization network learns…

  10. A Survey of Machine Learning for Big Code and Naturalness 

    July 30, 2018 | Miltos Allamanis, Earl T. Barr, Premkumar Devanbu, and Charles Sutton

    Research at the intersection of machine learning, programming languages, and software engineering has recently taken important steps in proposing learnable probabilistic models of source code that exploit the abundance of patterns of code. In this article, we survey this work. We contrast programming languages against…

  11. Adaptive Neural Trees 

    July 17, 2018

    Deep neural networks and decision trees operate on largely separate paradigms; typically, the former performs representation learning with pre-specified architectures, while the latter is characterised by learning hierarchies over pre-specified features with data-driven architectures. We unite the two via adaptive neural trees (ANTs) that incorporates…