{"id":4171,"date":"2015-12-04T10:00:00","date_gmt":"2015-12-04T10:00:00","guid":{"rendered":"https:\/\/blogs.technet.microsoft.com\/inside_microsoft_research\/2015\/12\/04\/deep-learning-machine-learning-advancements-highlight-microsofts-research-at-nips-2015\/"},"modified":"2016-08-05T21:14:06","modified_gmt":"2016-08-06T04:14:06","slug":"deep-learning-machine-learning-advancements-highlight-microsofts-research-at-nips-2015","status":"publish","type":"post","link":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/blog\/deep-learning-machine-learning-advancements-highlight-microsofts-research-at-nips-2015\/","title":{"rendered":"Deep learning, machine learning advancements highlight Microsoft&#8217;s research at NIPS 2015"},"content":{"rendered":"<p><em>Posted by <span class=\"author\">George Thomas Jr.<\/em><br \/>\n<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" href=\"https:\/\/nips.cc\/Conferences\/2015\" target=\"_blank\"><img decoding=\"async\" style=\"float: right; margin: 8px 10px;\" title=\"NIPS 2015\" src=\"https:\/\/msdnshared.blob.core.windows.net\/media\/TNBlogsFS\/prod.evol.blogs.technet.com\/CommunityServer.Blogs.Components.WeblogFiles\/00\/00\/00\/90\/35\/nips-2015-logo.PNG\" alt=\"NIPS 2015\" \/><span class=\"sr-only\"> (opens in new tab)<\/span><\/a>Deep learning is all around us.<\/p>\n<p>As part of the broad discipline that is <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"machine learning\" href=\"https:\/\/en.wikipedia.org\/wiki\/Machine_learning\" target=\"_blank\">machine learning<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"deep learning\" href=\"https:\/\/en.wikipedia.org\/wiki\/Deep_learning\" target=\"_blank\">deep learning<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> increasingly is embedded in our daily lives.<\/p>\n<p><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Skype Translator\" href=\"http:\/\/blogs.skype.com\/2014\/12\/15\/skype-translator-how-it-works\/\" target=\"_blank\">Skype Translator<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> is learning how to translate spoken words into more and more languages in near real-time, allowing for face-to-face communication once exclusive to Star Trek&#8217;s fictional universal translator.<\/p>\n<p><a title=\"Cortana\" href=\"https:\/\/newed.any0.dpdns.org\/en-us\/mobile\/experiences\/cortana\/\" target=\"_blank\">Cortana<\/a>, the personal digital assistant that debuted on Windows Phones and is now widely available on Windows 10 and for Android and iPhone users, learns from your interactions and helps you do things like manage your calendar, track packages and even chat with you and tell jokes \u2014 all the while customizing the experience for truly personal interaction.<\/p>\n<p><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Clutter\" href=\"https:\/\/blogs.office.com\/2014\/11\/11\/de-clutter-inbox-office-365\/\" target=\"_blank\">Clutter<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, a feature of Microsoft Office 2016, learns which emails are most important to you and automatically redirects less important mail to a separate folder, helping keep your inbox clean.<\/p>\n<p>And that&#8217;s just the tip of the iceberg.<\/p>\n<h2>A new computational era<\/h2>\n<p>&#8220;We&#8217;re just at the very, very beginning of a computational era \u2014 computation will touch every aspect of our lives,&#8221; says <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Cynthia Dwork\" href=\"http:\/\/research.microsoft.com\/en-us\/people\/dwork\/\" target=\"_blank\">Cynthia Dwork<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, a cryptographer and distinguished scientist at Microsoft Research.<\/p>\n<p>Dwork is among a bevy of Microsoft researchers and engineers whose work \u2014 more than 20 accepted papers &#8212; will be presented next week at the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"2015 Conference and Workshop on Neural Information Processing Systems\" href=\"https:\/\/nips.cc\/Conferences\/2015\" target=\"_blank\">2015 Conference and Workshop on Neural Information Processing Systems<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> (NIPS). It is the premiere conference on machine learning, but the pervasive nature of machine learning has seen it nearly double in size, growing to more than 4,000 attendees compared to 2,500 last year.<\/p>\n<p>&#8220;People see machine learning as more and more important \u2014 and deep learning is increasingly central to business,&#8221; says Li Deng, partner research manager in the Redmond, Wash. research lab.<\/p>\n<p>Deng, recently selected for the <a title=\"2015 IEEE Signal Processing Society Technical Achievement Award\" href=\"\/b\/inside_microsoft_research\/archive\/2015\/12\/03\/deng-receives-prestigious-ieee-technical-achievement-award.aspx\" target=\"_blank\">2015 IEEE Signal Processing Society Technical Achievement Award<\/a> for outstanding contributions to deep learning and to automatic speech recognition, pioneered Microsoft&#8217;s deep-learning speech recognition research. Its implications translate into any number of speech-enabled applications.<\/p>\n<p>Microsoft&#8217;s strength, in fact, is how it is integrating deep learning methodologies across many of the company&#8217;s products, including Office products, Bing search, Bing ads, and a number of others.<\/p>\n<h2>Deep learning at Microsoft<\/h2>\n<table style=\"width: 285px;\" border=\"0\" cellspacing=\"5\" align=\"right\">\n<tbody>\n<tr>\n<td><img decoding=\"async\" src=\"https:\/\/msdnshared.blob.core.windows.net\/media\/TNBlogsFS\/prod.evol.blogs.technet.com\/CommunityServer.Blogs.Components.WeblogFiles\/00\/00\/00\/90\/35\/6675.li-deng.png\" alt=\" \" border=\"0\" \/><\/td>\n<\/tr>\n<tr>\n<td><strong>Li Deng says deep learning is a major area of focus at Microsoft and already accounts for numerous successes.<\/strong><\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p>&#8220;In short, there is massive work with success achieved by deep learning within Microsoft,&#8221; Deng says.<\/p>\n<p>Deng co-authored one of the papers, with his colleagues Jianshu Chen, <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Lin Xiao\" href=\"http:\/\/research.microsoft.com\/en-us\/people\/lixiao\/\" target=\"_blank\">Lin Xiao<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Xiaodong He\" href=\"http:\/\/research.microsoft.com\/en-us\/people\/xiaohe\/\" target=\"_blank\">Xiaodong He<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Jianfeng Gao\" href=\"http:\/\/research.microsoft.com\/en-us\/um\/people\/jfgao\/\" target=\"_blank\">Jianfeng Gao<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> and Ji He, accepted into NIPS: <em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"End-to-end Learning of LDA by Mirror-Descent Back Propagation over a Deep Architecture\" href=\"http:\/\/arxiv.org\/pdf\/1508.03398.pdf\" target=\"_blank\">End-to-end Learning of LDA by Mirror-Descent Back Propagation over a Deep Architecture<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em>. &#8220;It&#8217;s a new way of thinking about deep learning,&#8221; he notes, as it stems from a unique approach emanating from business needs.<\/p>\n<p>Previously, standard discriminative deep neural networks were used by Deng&#8217;s team to predict, with high accuracy, a set of outcomes of high business value. However, the system&#8217;s users also demand high interpretability: they want to know why and how the high accuracy is achieved and what kinds of raw features account for the success, because these are critical to their business decisions.<\/p>\n<p>To obtain both interpretability and high prediction accuracy, Deng and colleagues combined the strengths of deep generative and discriminative models. The novel method described in their NIPS paper exploits the power of discriminative learning (via <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"backpropagation\" href=\"https:\/\/en.wikipedia.org\/wiki\/Backpropagation\" target=\"_blank\">backpropagation<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>) to train the underlying generative topic model called <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Latent Dirichlet Allocation\" href=\"https:\/\/en.wikipedia.org\/wiki\/Latent_Dirichlet_allocation\" target=\"_blank\">Latent Dirichlet Allocation<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> (LDA).<\/p>\n<p>This is akin to the earlier popular method of training (shallow) generative speech models using discriminative criteria before deep neural nets took over the field. But in contrast, the new model architecture described in the paper is very deep, where the data flow in the intermediate layers is designed according to the natural inference steps associated with the original generative LDA model.<\/p>\n<h2>A paradigm shift in adaptive data analysis<\/h2>\n<table style=\"width: 285px;\" border=\"0\" cellspacing=\"5\" align=\"right\">\n<tbody>\n<tr>\n<td><img decoding=\"async\" src=\"https:\/\/msdnshared.blob.core.windows.net\/media\/TNBlogsFS\/prod.evol.blogs.technet.com\/CommunityServer.Blogs.Components.WeblogFiles\/00\/00\/00\/90\/35\/cynthia-dwork.png\" alt=\" \" border=\"0\" \/><\/td>\n<\/tr>\n<tr>\n<td><strong>Cynthia Dwork and her colleagues have created new algorithms that represent a paradigm shift in machine learning and data analytics.<\/strong><\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p>Dwork, too, is presenting a new approach to machine learning. Her paper, <em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Generalization in Adaptive Data Analysis and Holdout Reuse\" href=\"http:\/\/arxiv.org\/pdf\/1506.02629.pdf\" target=\"_blank\">Generalization in Adaptive Data Analysis and Holdout Reuse<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em>, in fact represents a true paradigm shift in adaptive data analysis by solving the problem known as overfitting, comically referred to as &#8220;the bane of data analysts&#8221; in the paper&#8217;s abstract.<\/p>\n<p>Analysts rarely run just a single machine learning algorithm on data; analysis is more often an &#8220;adaptive&#8221; process, in which each new step may depend on the outcomes of previous steps. But this can result in &#8220;overfitting&#8221; &#8212; learning things about the data set that do not apply to the population from which the data were drawn &#8212; and, Dwork notes, is among the reasons for inaccuracies in some scientific research.<\/p>\n<p>But Dwork and her colleagues have shown that accessing the data through a &#8220;differentially private&#8221; algorithm &#8212; a concept developed at Microsoft Research and the subject of over a thousand scholarly articles &#8212; prevents overfitting even in adaptive analysis. Informed by this insight, they have provided algorithms allowing analysts to work as they are accustomed to do with a training set, and to check their results using differential privacy, on a holdout set. By accessing the holdout set in a differentially private manner it can be re-used a great many times, repeatedly playing the role of &#8220;fresh&#8221; data.<\/p>\n<p>&#8220;Having a reusable holdout set is key,&#8221; Dwork says, adding others already are building on her work for a range of applications.<\/p>\n<h2>Computer vision and crowdsourcing<\/h2>\n<p>Computer vision also is well represented at NIPS. <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Pushmeet Kohli\" href=\"http:\/\/research.microsoft.com\/en-us\/people\/pkohli\/\" target=\"_blank\">Pushmeet Kohli<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> from Microsoft&#8217;s research lab in Cambridge, U.K., contributed to the paper <em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Deep Convolutional Inverse Graphics Network\" href=\"http:\/\/arxiv.org\/pdf\/1503.03167.pdf\" target=\"_blank\">Deep Convolutional Inverse Graphics Network<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em>, which introduces a 3D rendering engine that enables a computer to render an object in 3D that it&#8217;s never seen before. For example, you can show it an image of a chair and ask what the chair would look like, rotated.<\/p>\n<p>Another paper coauthored by senior researcher <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Dengyong Zhou\" href=\"http:\/\/research.microsoft.com\/en-us\/people\/denzho\/\" target=\"_blank\">Dengyong Zhou<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> helps solve a key machine learning problem involving crowdsourcing. <em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Double or Nothing: Multiplicative Incentive Mechanisms for Crowdsourcing\" href=\"http:\/\/arxiv.org\/pdf\/1408.1387.pdf\" target=\"_blank\">Double or Nothing: Multiplicative Incentive Mechanisms for Crowdsourcing<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em> introduces a method that in initial tests has led to a three-fold decrease in error rates in crowdsourced data labeling.<\/p>\n<p>Data in machine learning models must have labels to be meaningful \u2014 like a photo of a car that is labeled &#8220;car.&#8221; The labeled data then can be compared to unlabeled data, and the computer can guess how to classify that data based on previously labeled similar data.<\/p>\n<p>In the past, data labeling was left to experts, but the limited number of experts would limit the size of data sets. This led to crowdsourcing the data labeling task \u2014 recruiting non-experts via the Internet \u2014 which in turn resulted in some poor quality labeling.<\/p>\n<p>Zhou and his colleague, <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Nihar B. Shah\" href=\"http:\/\/www.eecs.berkeley.edu\/~nihar\/\" target=\"_blank\">Nihar B. Shah<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> of the University of California-Berkeley, introduce a simple and unique incentive that allows the crowdsourced &#8220;judges&#8221; to skip items and estimate their confidence in each label, thereby significantly increasing the quality of labeling.<\/p>\n<h2>Reducing\u00a0bottlenecks in machine learning<\/h2>\n<p>But reducing the bottleneck in properly labeled data isn&#8217;t the only one affecting the industry.<\/p>\n<p>&#8220;Machine learning used to be all about generalization,&#8221; says <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Patrice Simard\" href=\"http:\/\/research.microsoft.com\/en-us\/people\/patrice\/\" target=\"_blank\">Patrice Simard<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, a distinguished engineer also based in Microsoft&#8217;s Redmond lab. &#8220;Then people realized that for some problems, the bottleneck is not the algorithm, it is the teacher.&#8221;<\/p>\n<p>The &#8220;some problems,&#8221; he says, applies to problems for which very little data is available. &#8220;In these cases, the risk of overfitting is pervasive and the need for human supervision is dire&#8221; in the form of labels or features.<\/p>\n<p>For example, one such problem could be to recognize a command given to an oven, a car, or a personal assistant like Cortana. Each command, in each language, for each device, service, or application requires building a custom model that is robust to all the creative ways humans can say the same command (or not). A custom system that correctly interprets &#8220;lights on first floor off&#8221; or &#8220;patio, on&#8221; in the right context can be built with a few hundred labels and the right features. With the right tools, this can be done in a couple hours of teaching time with little machine learning expertise. It does not require deep learning or solving artificial intelligence.<\/p>\n<table style=\"width: 285px;\" border=\"0\" cellspacing=\"5\" align=\"right\">\n<tbody>\n<tr>\n<td><img decoding=\"async\" src=\"https:\/\/msdnshared.blob.core.windows.net\/media\/TNBlogsFS\/prod.evol.blogs.technet.com\/CommunityServer.Blogs.Components.WeblogFiles\/00\/00\/00\/90\/35\/machine-teaching-simard.png\" alt=\" \" border=\"0\" \/><\/td>\n<\/tr>\n<tr>\n<td><strong>Patrice Simard says the future of machine learning is machine teaching.<\/strong><\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p>Simard is at the forefront of a <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"machine teaching project\" href=\"http:\/\/blogs.microsoft.com\/next\/2015\/07\/10\/the-next-evolution-of-machine-learning-machine-teaching\/\" target=\"_blank\">machine teaching project<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> developing tools that would allow anyone to teach a computer how to do machine learning tasks even if that person lacks expertise in data analysis or computer science.<\/p>\n<p>An example of this is the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Language Understanding Intelligent Service (LUIS)\" href=\"https:\/\/www.luis.ai\/Home\/About\" target=\"_blank\">Language Understanding Intelligent Service (LUIS)<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, which offers for free a fast and effective way of adding language understanding to applications. Recently released to beta, LUIS is part of <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Project Oxford\" href=\"https:\/\/www.projectoxford.ai\/\" target=\"_blank\">Project Oxford<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> and provides world-class, pre-built models from Bing and Cortana and guides users through the process of quickly building specialized models.<\/p>\n<p>&#8220;Right now most of the machine learning community is interested in algorithms, and I believe it will evolve to care far more about productivity,&#8221; Simard says, with machine teaching helping to make machine learning easier for non-experts.<\/p>\n<h2>A world of algorithms<\/h2>\n<p>In the meantime, algorithms certainly aren&#8217;t going away.<\/p>\n<p>&#8220;Algorithms are increasingly affecting our world,&#8221; says Dwork. &#8220;Data sets are growing and becoming of increasing interest to social scientists, politicians \u2014 we&#8217;re about to witness a huge increase in computational social sciences.&#8221;<\/p>\n<p>While computational neuroscience remains an aspect of NIPS, recent proceedings have been dominated by papers on machine learning, artificial intelligence and statistics, according to the NIPS website. And this year a dedicated tutorial and a dedicated symposium on deep learning at NIPS hints at the future \u2014 one Microsoft is anticipating.<\/p>\n<p>&#8220;Right now we have over 100 people across the company dedicated to deep learning,&#8221; Deng says. And given he may not be aware of all related activity, he says the number likely is higher.<\/p>\n<p>&#8220;In speech recognition, visual object recognition, and a few other areas of AI, if you&#8217;re not in deep learning, you&#8217;re outside the mainstream today,&#8221; he says.<\/p>\n<p><strong>Related:<\/strong><\/p>\n<ul>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Machine learning and artificial intelligence research at Microsoft\" href=\"http:\/\/research.microsoft.com\/en-us\/research-areas\/machine-learning-ai.aspx\" target=\"_blank\">Machine learning and artificial intelligence research at Microsoft<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"The quest to create technology that understands speech as well as a human\" href=\"http:\/\/blogs.microsoft.com\/next\/2015\/12\/03\/the-quest-to-create-technology-that-understands-speech-as-well-as-a-human\/\" target=\"_blank\">The quest to create technology that understands speech as well as a human<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Privacy and accuracy: How Cynthia Dwork is improving data analysis\" href=\"http:\/\/blogs.microsoft.com\/next\/2015\/08\/07\/privacy-and-accuracy-how-cynthia-dwork-is-making-data-analysis-better\/\" target=\"_blank\">Privacy and accuracy: How Cynthia Dwork is improving data analysis<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"The next evolution of machine learning: Machine teaching\" href=\"http:\/\/blogs.microsoft.com\/next\/2015\/07\/10\/the-next-evolution-of-machine-learning-machine-teaching\/\" target=\"_blank\">The next evolution of machine learning: Machine teaching<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<\/ul>\n<h2>Other Microsoft research papers at NIPS 2015<\/h2>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Efficient Non-greedy Optimization of Decision Trees and Forests\" href=\"http:\/\/arxiv.org\/abs\/1511.04056\" target=\"_blank\">Efficient Non-greedy Optimization of Decision Trees and Forests<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributors: Matthew Johnson, Pushmeet Kohli<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Logarithmic Time Online Multiclass prediction\" href=\"http:\/\/arxiv.org\/pdf\/1406.1822)\" target=\"_blank\">Logarithmic Time Online Multiclass prediction<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributor: John Langford<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Efficient and Parsimonious Agnostic Active Learning\" href=\"http:\/\/arxiv.org\/pdf\/1506.08669.pdf\" target=\"_blank\">Efficient and Parsimonious Agnostic Active Learning<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributors: Tzu-Kuo Huang, Alekh Agarwal, John Langford, Robert Schapire<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Private Graphon Estimation for Sparse Graphs\" href=\"http:\/\/arxiv.org\/pdf\/1506.06162.pdf\" target=\"_blank\">Private Graphon Estimation for Sparse Graphs<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributors: Christian Borgs, Jennifer Chayes<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Visalogy: Answering Visual Analogy Questions\" href=\"http:\/\/arxiv.org\/abs\/1510.08973\" target=\"_blank\">Visalogy: Answering Visual Analogy Questions<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributors: Ross Girshick, Larry Zitnick<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" title=\"Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks\" href=\"http:\/\/arxiv.org\/pdf\/1506.01497.pdf\">Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributors: Kaiming He, Ross Girshick, Jian Sun<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Robust Regression via Hard Thresholding\" href=\"http:\/\/arxiv.org\/pdf\/1506.02428.pdf\" target=\"_blank\">Robust Regression via Hard Thresholding<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributors: Purushottam Kar, Prateek Jain, Kush Bhatia<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Locally Non-linear Embeddings for Extreme Multi-label Learning\" href=\"http:\/\/arxiv.org\/pdf\/1507.02743.pdf\" target=\"_blank\">Locally Non-linear Embeddings for Extreme Multi-label Learning<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributors: Purushottam Kar, Prateek Jain, Manik Varma, Kush Bhatia<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Convergence Rates of Active Learning for Maximum Likelihood Estimation\" href=\"http:\/\/arxiv.org\/pdf\/1506.02348.pdf\" target=\"_blank\">Convergence Rates of Active Learning for Maximum Likelihood Estimation<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributor: Praneeth Netrapalli<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Multi-class SVMs: From Tighter Data-Dependent Generalization Bounds to Novel Algorithms\" href=\"http:\/\/arxiv.org\/pdf\/1506.04359.pdf\" target=\"_blank\">Multi-class SVMs: From Tighter Data-Dependent Generalization Bounds to Novel Algorithms<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributor: Urun Dogan<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Bandit Smooth Convex Optimization: Improving the Bias-Variance Tradeoff\" href=\"http:\/\/arxiv.org\/pdf\/1506.08669.pdf\" target=\"_blank\">Bandit Smooth Convex Optimization: Improving the Bias-Variance Tradeoff<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributor: Ofer Dekel<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Fast Convergence of Regularized Learning in Games\" href=\"http:\/\/arxiv.org\/pdf\/1507.00407.pdf\" target=\"_blank\">Fast Convergence of Regularized Learning in Games<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributors: Vasilis Syrgkanis, Alekh Agarwal, Robert Schapire<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"No-Regret Learning in Repeated Bayesian Games\" href=\"http:\/\/arxiv.org\/pdf\/1507.00418.pdf\" target=\"_blank\">No-Regret Learning in Repeated Bayesian Games<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributor: Vasilis Syrgkanis<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"On Elicitation Complexity and Conditional Elicitation\" href=\"http:\/\/arxiv.org\/pdf\/1506.07212.pdf\" target=\"_blank\">On Elicitation Complexity and Conditional Elicitation<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributor: Ian Kash<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Streaming Min-max Hypergraph Partitioning\" href=\"http:\/\/research.microsoft.com\/apps\/pubs\/default.aspx?id=232516\" target=\"_blank\">Streaming Min-max Hypergraph Partitioning<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributors: Dan Alistarh, Milan Vojnovic<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Alternating Minimization for Regression Problems with Vector-valued Outputs\" href=\"http:\/\/dept.stat.lsa.umich.edu\/~tewaria\/research\/jain15alternating.pdf\" target=\"_blank\">Alternating Minimization for Regression Problems with Vector-valued Outputs<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributor: Prateek Jain<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Predtron: A Family of Online Algorithms for General Prediction Problems\" href=\"http:\/\/dept.stat.lsa.umich.edu\/~tewaria\/research\/jain15predtron.pdf\" target=\"_blank\">Predtron: A Family of Online Algorithms for General Prediction Problems<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributor: Prateek Jain<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Stochastic Online Greedy Learning with Semi-bandit Feedbacks\" href=\"http:\/\/research.microsoft.com\/en-US\/people\/weic\/nips2015_full_og.pdf\" target=\"_blank\">Stochastic Online Greedy Learning with Semi-bandit Feedbacks<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributor: Wei Chen<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Fast and Memory Optimal Low-Rank Matrix Approximation\" href=\"http:\/\/papers.nips.cc\/paper\/5929-fast-and-memory-optimal-low-rank-matrix-approximation.pdf\" target=\"_blank\">Fast and Memory Optimal Low-Rank Matrix Approximation<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributor: Seyoung Yun<\/p>\n<p><em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" title=\"Finite-Time Analysis of Projected Langevin Monte Carlo\" href=\"http:\/\/papers.nips.cc\/paper\/4225-finite-time-analysis-of-stratified-sampling-for-monte-carlo.pdf\" target=\"_blank\">Finite-Time Analysis of Projected Langevin Monte Carlo<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/em><br \/>\nMicrosoft contributor: Sebastien Bubeck<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Posted by George Thomas Jr. Deep learning is all around us. As part of the broad discipline that is machine learning, deep learning increasingly is embedded in our daily lives. Skype Translator is learning how to translate spoken words into more and more languages in near real-time, allowing for face-to-face communication once exclusive to Star [&hellip;]<\/p>\n","protected":false},"author":30766,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr-author-ordering":[],"msr_hide_image_in_river":0,"footnotes":""},"categories":[194467,194455],"tags":[200265,186834,186897,201165,186925,202433,186418,203065,203221],"research-area":[13556,13562],"msr-region":[],"msr-event-type":[],"msr-locale":[268875],"msr-post-option":[],"msr-impact-theme":[],"msr-promo-type":[],"msr-podcast-series":[],"class_list":["post-4171","post","type-post","status-publish","format-standard","hentry","category-artifical-intelligence","category-machine-learning","tag-adaptive-data-analysis","tag-algorithms","tag-computer-vision","tag-cynthia-dwork","tag-deep-learning","tag-li-deng","tag-machine-learning","tag-nips-2015","tag-patrice-simard","msr-research-area-artificial-intelligence","msr-research-area-computer-vision","msr-locale-en_us"],"msr_event_details":{"start":"","end":"","location":""},"podcast_url":"","podcast_episode":"","msr_research_lab":[],"msr_impact_theme":[],"related-publications":[],"related-downloads":[],"related-videos":[],"related-academic-programs":[],"related-groups":[],"related-projects":[],"related-events":[],"related-researchers":[],"msr_type":"Post","byline":"","formattedDate":"December 4, 2015","formattedExcerpt":"Posted by George Thomas Jr. Deep learning is all around us. As part of the broad discipline that is machine learning, deep learning increasingly is embedded in our daily lives. Skype Translator is learning how to translate spoken words into more and more languages in&hellip;","locale":{"slug":"en_us","name":"English","native":"","english":"English"},"_links":{"self":[{"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/posts\/4171","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/users\/30766"}],"replies":[{"embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/comments?post=4171"}],"version-history":[{"count":3,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/posts\/4171\/revisions"}],"predecessor-version":[{"id":272460,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/posts\/4171\/revisions\/272460"}],"wp:attachment":[{"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/media?parent=4171"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/categories?post=4171"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/tags?post=4171"},{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=4171"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=4171"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=4171"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=4171"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=4171"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=4171"},{"taxonomy":"msr-promo-type","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-promo-type?post=4171"},{"taxonomy":"msr-podcast-series","embeddable":true,"href":"https:\/\/newed.any0.dpdns.org\/en-us\/research\/wp-json\/wp\/v2\/msr-podcast-series?post=4171"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}