Meta-learning Research Overview and Paper Group
Incomplete and in progress. Feel free to comment paper recommendations below.
Radical Meta-Learning Paper Group
FAQ
What is the purpose of this reading group?
There is a trend for algorithms to move away from handcrafted parameters and toward minimizing the human-specified inductive bias as much as possible. Meta-learning is a continuation of this trend.
In this reading group, we want to gain a deep understanding of “radical” meta-learning, i.e., algorithms that learn how to learn how to learn – and on till the nth meta-level.
Why “radical”?
“Radical” is a word taken from Jurgen Schmidhuber to distinguish this branch of meta-learning from current transfer learning techniques (e.g. MAML). For example, this difference could be between learning gradient descent itself (radical meta-learning) and optimizing gradient descent across many data distributions (a type of meta-learning / transfer learning). We want to study the first camp: learning as much of the algorithm as possible from scratch.
How do I join?
Shoot a quick email to Sonia Joseph at smjoseph@alumni.princeton.edu
What is the format?
One person will sign up to present a paper, which everyone will read. We’ll collectively take notes on a Google doc before and during the presentation. The presenter will go through the paper, and we’ll liberally discuss along the way. Some of the papers are long, so please feel free to come anyway even if you haven’t finished it.
Phase 1:
Sample reading list [please put your name next to what you would like to present]
- AIXI. Link.
- Evolutionary principles in self-referential learning. (On learning how to learn: The meta-meta-… hook.). 1987. Link. Presenter: Sonia Joseph
- Ultimate Cognition a la Godel. 2009. Link. Presenter: Sid
- Reinforcement Learning with Self-Modifying Policies. 1997. Link.
- Optimally Ordered Problem Solver. 2004. Link.
- AI-GAs: AI-generating algorithms, an alternate paradigm for producing general artificial intelligence. Link. Presenter: Acyr.
- Improving Generalization in Meta Reinforcement Learning using Learned Objectives. 2019. Link. Presenter: Rob/Louis Kirsch (?)
- Discovering Reinforcement Learning Algorithms. 2020. Link. Presenter: Rob
- Meta-Gradient Reinforcement Learning with an Objective Discovered Online. 2020. Link. Presenter: Rob
Schedule [tentative]
We will have four meetings.This is rough— we’ll finalize the schedule with a whenisgood.
- Seminar 1. Sunday Jan 29, 6:00 pm GMT
- Whenisgood for first meeting: https://whenisgood.net/78wbi7k
- Paper: Evolutionary principles in self-referential learning. (On learning how to learn: The meta-meta-… hook.). 1987. Link.
- Presenter: Sonia Joseph
- Notes: https://docs.google.com/document/d/1HBzO4YEm2L8drf82UPxCRDlAqiRzMJcJW77ccyCoLW4/edit?usp=sharing
- Seminar 2. Feb 14th, 6:00 pm GMT
- Paper: Ultimate Cognition a la Godel.
- Presenter: Sid
- Notes: https://docs.google.com/document/d/1Wg4Aoy9qhhaBpPrUxp9KNhBf2f3-HwiagCN4arQhJ5E/edit#
- Seminar 3. Feb 26 weekend.
- Paper: AI-GAs: AI-generating algorithms, an alternate paradigm for producing general artificial intelligence. Link.
- Presenter: Acyr.
- Notes: Coming soon.
- Seminar 4. March 12th weekend.
- Paper: Meta-Policy Gradients - Papers: 6,7,8.
- Presenter: Rob.
- Notes: Checkout this fantastic blog post 👨🔧
Phase 2
Reading list [AI-GA related]
- Generative teaching networks: learning to teach by generating synthetic training data. Link
- Differentiable plasticity: training plastic neural networks with backpropagation. Link
- Paired open-ended trailblazer (poet): Endlessly generating increasingly complex and diverse learning environments and their solutions. Link
Schedule [tentative]
- Seminar 1. April 11th, 6:00 pm GMT
- POET by Anushan.
- Seminar 2. April 25th, 6:00 pm GMT
- Differentiable Plasticity by Harsha
Research Papers by Concept
Proof-Search
Transfer Learning
Reinforcement Learning
- Meta-Gradient Reinforcement Learning with an Online Objective Discovered Online (2020)
- Discovering Reinforcement Learning Algorithms (2021)
- Improving Generalization in Meta Reinforcement Learning using Learned Objectives (2019)
Leave a comment