Download Open Datasets on 1000s of Projects + Share Projects on One Platform. Explore Popular Topics Like Government, Sports, Medicine, Fintech, Food, More. Flexible Data Ingestion.
In this blog post we will begin to look at Monte Carlo methods and how they can be used. These form the backbone of (essentially) all statistical computer modelling.
Any fundamental discovery involves a significant degree of risk. If an idea is guaranteed to work then it moves from the realm of research to engineering. Unfortunately, this also means that most…
While deep learning has achieved remarkable success in supervised and reinforcement learning problems, such as image classification, speech recognition, and game playing, these models are, to a large degree, specialized for the single task they are trained for. This course will cover the setting where there are multiple tasks to be solved, and study how the structure arising from multiple tasks can be leveraged to learn more efficiently or effectively. This includes:
- goal-conditioned reinforcement learning techniques that leverage the structure of the provided goal space to learn many tasks significantly faster
- meta-learning methods that aim to learn efficient learning algorithms that can learn new tasks quickly
- curriculum and lifelong learning, where the problem requires learning a sequence of tasks, leveraging their shared structure to enable knowledge transfer
This is a graduate-level course. By the end of the course, students will be able to understand and implement the state-of-the-art multi-task learning and meta-learning algorithms and be ready to conduct research on these topics.
- Aug. 19 – Aug. 28, 2020
- Nike Sun (Massachusetts Institute of Technology; chair), Jian Ding (University of Pennsylvania), Ronen Eldan (Weizmann Institute), Elchanan Mossel (Massachusetts Institute of Technology), Joe Neeman (University of Texas at Austin), Jelani Nelson (UC Berkeley), Tselil Schramm (Stanford University; Microsoft Research Fellow)