Putting the cloud under the sea with Ben Cutler
Episode 40, September 5, 2018 – In today’s podcast we find out a bit about what else the Special Projects team is up to, and then we hear all about Project Natick and how Ben…
Towards a Theory for Sample-efficient Reinforcement Learning with Rich Observations
How can we tractably solve sequential decision making problems where the learning agent receives rich observations? We begin with a new model called Contextual Decision Processes (CDPs) for studying such problems, and show that it…
Super-Human AI for Strategic Reasoning
Poker has been a challenge problem in game theory, operations research, and artificial intelligence for decades. As a game of imperfect information, it involves obstacles not present in games like chess and go, and requires…
Fully Online Matching
We introduce a fully online model of maximum cardinality matching in which all vertices arrive online. On the arrival of a vertex, its incident edges to previously-arrived vertices are revealed. Each vertex has a deadline…
LSH-Sampling Breaks the Computation Chicken-and-Egg Loop in Adaptive Stochastic Gradient Estimation
Stochastic Gradient Descent or SGD is the most popular algorithm for large-scale optimization. In SGD, the gradient is estimated by uniform sampling with sample size one. There have been several results that show better gradient…