Time | Title | Speaker | |
---|---|---|---|
9:15 – 9:45
|
Registration & Welcome tea/coffee
|
||
9:45 – 10:05
|
Learning to Code: Machine Learning for Program Induction
|
Alex Gaunt
|
|
10:05 – 10:25
|
Learning Program Representations: Symbols to Vectors to Semantics
|
||
10:25 – 10:45
|
TBD
|
||
10:45 – 11:05
|
From GANs to Variational Divergence Minimization
|
||
11:05 – 11:30
|
Coffee break
|
||
11:30 – 11:50
|
Automatic Discovery of the Statistical Types of Variables in a Dataset
|
Isabel Valera
|
|
11:50 – 12:10
|
Approximate Inference with Amortised MCMC
|
Yingzhen Li
|
|
12:10 – 12:30
|
The Automatic Statistician: a project update
|
Zoubin Ghahramani
|
|
12:30 – 12:50
|
The Supervised Word Mover’s Distance
|
Matt Kusner
|
|
1:00 – 2:00
|
Lunch
|
||
2:00 – 2:20
|
Learning and Policy Search in Stochastic Dynamical Systems with Bayesian Neural Networks
|
Jose Miguel Hernandez Lobato
|
|
2:20 – 2:40
|
Bayesian optimisation in many dimensions with bespoke models
|
Adrian Weller
|
|
2:40 – 3:00
|
Grammar Variational Autoencoder
|
Brooks Paige
|
|
3:00 – 3:20
|
Invertible Transformations for Bayesian Neural Network Inference
|
Amar Shah
|
|
3:20 – 4:00
|
Coffee break
|
||
4:00 – 4:20
|
AI for Healthcare
|
||
4:20 – 4:40
|
Project Alexandria: a web scale probabilistic program for unsupervised knowledge base construction
|
||
4:40 – 5:00
|
The Malmo Collaborative AI Challenge
|