Time discretization invariance in Machine Learning, applications to reinforcement learning...

Subscribers:
344,000
Published on ● Video Link: https://www.youtube.com/watch?v=XkDuJwWAUqQ



Duration: 57:19
1,625 views
33


While computers are well equipped to deal with discrete flows of data, the real world often provides intrisically continuous time data sequences, e.g. visual, sensory streams, time series, or state variables in continuous control environments. Most algorithms and notably machine learning approaches require discretization of time continuous data flows, introducing a notion of processing discretization timestep. Using smaller discretization timesteps usually provides more information to the processing algorithm, and should normally be associated with better algorithm performance. However, many commonly used algorithms fail to follow this trend, their performance decreases with smaller discretization timesteps, and dramatically drops when the discretization timestep approaches 0. In this talk, I will focus on the design of time discretization invariant algorithms, i.e. algorithms that work for any given time discretization, and notably remain viable for very small time discretizations. Such algorithms often rely on the design of a theoretical, inherently time continuous, but untractable algorithm, that is then discretized. Algorithms that don't scale to small time discretization typically don't admit such a time continuous limit algorithm. The talk will focus on two specific applications, namely the design of Q-learning approaches robust to time discretization, and the analysis of time discretization invariant architectures for recurrent neural networks. Beside the practical benefits, I will show that time discretization invariant designs provide interesting theoretical insights, and, for instance, lead to rethinking some widely spread exploration strategies, or shed new light on the use of gating mechanisms in recurrent networks.

See more at https://www.microsoft.com/en-us/research/video/time-discretization-invariance-in-machine-learning-applications-to-reinforcement-learning-and-recurrent-neural-networks/




Other Videos By Microsoft Research


2019-09-13Fly-through in the AirSim simulation Team Explorer created for the DARPA SubT Challenge
2019-09-12Compacting the Uncompactable: The Mesh Compacting Memory Allocator
2019-09-12Privacy-Preserving Statistical Learning and Testing
2019-09-12Program Synthesis meets Notebooks
2019-09-11Inside the Microsoft AI Residency Program with Dr. Brian Broll [Podcast]
2019-09-10Real-time Single-channel Speech Enhancement with Recurrent Neural Networks
2019-09-06AI Institute Geometry of Deep Learning 2019 [Workshop] Day 2 | Session 1]
2019-09-05AI Institute "Geometry of Deep Learning" 2019 [Workshop] Day 3 | Session 1
2019-09-05AI Institute "Geometry of Deep Learning" 2019 [Workshop] Day 3 | Session 2
2019-09-05EzPC (Easy Secure Multi-party Computation)
2019-09-04Time discretization invariance in Machine Learning, applications to reinforcement learning...
2019-09-04Antennas for light and their applications in classical optics, Dr Rupert Oulton, Imperial College
2019-09-04Photonics for Computing: from Optical Interconnects to Neuromorphic Architectures
2019-09-04Inclusive design for all, or ICT4D and 4U! with Dr. Ed Cutrell [Podcast]
2019-09-03AI Institute "Geometry of Deep Learning" 2019 [Workshop] Day 2 | Session 4
2019-09-03AI Institute "Geometry of Deep Learning" 2019 [Workshop] Day 2 | Session 3
2019-09-03AI Institute "Geometry of Deep Learning" 2019 [Workshop] Day 2 | Session 2
2019-09-03AI Institute "Geometry of Deep Learning" 2019 [Workshop] Day 1 | Session 4
2019-09-03AI Institute "Geometry of Deep Learning" 2019 [Workshop] Day 1 | Session 3
2019-09-03AI Institute "Geometry of Deep Learning" 2019 [Workshop] Day 1 | Session 1
2019-09-03AI Institute "Geometry of Deep Learning" 2019 [Workshop] Day 1 | Session 2



Tags:
microsoft research
machine learning