Microsoft Research Blog

Artificial intelligence

  1. On the challenges of learning with inference networks on sparse, high-dimensional data 

    October 16, 2017 | Rahul Gopalkrishnan, Dawen Liang, and Matthew D. Hoffman

    We study parameter estimation in Nonlinear Factor Analysis (NFA) where the generative model is parameterized by a deep neural network. Recent work has focused on learning such models using inference (or recognition) networks; we identify a crucial problem when modeling large, sparse, high-dimensional datasets --…

  2. Scale-out acceleration for machine learning 

    October 13, 2017

    The growing scale and complexity of Machine Learning (ML) algorithms has resulted in prevalent use of distributed general-purpose systems. In a rather disjoint effort, the community is focusing mostly on high performance single-node accelerators for learning. This work bridges these two paradigms and offers CoSMIC,…

  3. Coherent Online Video Style Transfer 

    September 30, 2017

    Training a feed-forward network for the fast neural style transfer of images has proven successful, but the naive extension of processing videos frame by frame is prone to producing flickering results. We propose the first end-toend network for online video style transfer, which generates temporally…

  4. Flow-Guided Feature Aggregation for Video Object Detection 

    September 30, 2017

    Extending state-of-the-art object detectors from image to video is challenging. The accuracy of detection suffers from degenerated object appearances in videos, e.g., motion blur, video defocus, rare poses, etc. Existing work attempts to exploit temporal information on box level, but such methods are not trained…

  5. Learning Intrinsic Sparse Structures within Long Short-Term Memory 

    September 14, 2017

    Model compression is significant for the wide adoption of Recurrent Neural Networks (RNNs) in both user devices possessing limited resources and business clusters requiring quick responses to large-scale service requests. This work aims to learn structurally-sparse Long Short-Term Memory (LSTM) by reducing the sizes of…

  6. Fine-tuned convolutional neural nets for cardiac MRI acquisition plane recognition 

    September 3, 2017

    In this paper, we propose a convolutional neural network-based method to automatically retrieve missing or noisy cardiac acquisition plane information from magnetic resonance imaging and predict the five most common cardiac views. We fine-tune a convolutional neural network (CNN) initially trained on a large natural…

  7. Learning Fine-Grained Expressions to Solve Math Word Problems 

    September 1, 2017 | Danqing Huang, Shuming Shi, Chin-Yew Lin, and Jian Yin

    This paper presents a novel template-based method to solve math word problems. This method learns the mappings between math concept phrases in math word problems and their math expressions from training data. For each equation template, we automatically construct a rich template sketch by aggregating…

  8. Going deeper in the automated identification of Herbarium specimens. 

    August 10, 2017

    Hundreds of herbarium collections have accumulated a valuable heritage and knowledge of plants over several centuries. Recent initiatives started ambitious preservation plans to digitize this information and make it available to botanists and the general public through web portals. However, thousands of sheets are still…

  9. Consistency Analysis for Binary Classification Revisited 

    July 17, 2017 | Krzysztof Dembczyński, Wojciech Kotłowski, Oluwasanmi Koyejo, and Nagarajan Natarajan

    Statistical learning theory is at an inflection point enabled by recent advances in understanding and optimizing a wide range of metrics. Of particular interest are non-decomposable metrics such as the F-measure and the Jaccard measure which cannot be represented as a simple average over examples.…

  10. Developing Bug-Free Machine Learning Systems with Formal Mathematics 

    July 16, 2017 | Daniel Selsam, Percy Liang, and David L. Dill

    Noisy data, non-convex objectives, model misspecification, and numerical instability can all cause undesired behaviors in machine learning systems. As a result, detecting actual implementation errors can be extremely difficult. We demonstrate a methodology in which developers use an interactive proof assistant to both implement their…

  11. Deep Feature Flow for Video Recognition 

    June 30, 2017

    Deep convolutional neutral networks have achieved great success on image recognition tasks. Yet, it is non-trivial to transfer the state-of-the-art image recognition networks to videos as per-frame evaluation is too slow and unaffordable. We present deep feature flow, a fast and accurate framework for video…