Modular Meta Learning
CoRL 2018
Motivation
Previous approaches to meta-learning have focused on finding distributions or initial values of parameters.
Our objective is similar, but rather than focusing on transferring information about parameter values, we focus on finding a set of reusable modules that can form components of a solution to a new task, possibly with a small amount of tuning.
The authors provide an algorithm, called BounceGrad, which learns a set of modules and then combines them appropriately for a new task.
Objective
Given the specification of a composition rule and a basis set of modules, (C,F,Θ)represents a set of possible functional input-output mappings that will serve as the hypothesis space for the meta-test task.
F is a basis set of modules, which are functions f1​,f2​,…,fk​
Each function has a parametric form y=fi​(x;θi​)where θi​is a fixed-dimensional vector of parameters.
In this work, all the fi​are neural networks, potentially with different architectures, and the parameters Θ=(θ1​,…,θk​) are the weights of the neural networks, which differ among the modules. Some examples of modules:
Single module h(x)=fi​(x)
A fixed compositional structure: h(x)=fi​(x)+fj​(x)
A weighted ensemble.

S: the set of possible structures and S∈S is a particular structure generated by C . This approach has two phases: an off-line meta-learning phase and an on-line meta-test learning phase.
Meta-learning phase: we take training and validation data sets for tasks 1,…,k as input and generate a parametrization Θfor each module. The objective is to construct modules that will work together as good building blocks for future tasks.
At meta-learning time, S is specified, and the objective is to find parameter values Θ that constitute a set of modules that can be recombined to effectively solve each of the training tasks.
Validation set is used for the meta-training tasks to avoid choosing Θ in a way that overfits.
The training objective is to find Θthat minimizing the average generalization performance of the hypotheses SΘ∗​using parameter set Θ : (See above figure)
Meta-test learning phase: we take a training data set for the meta-test task as input, as well as S and Θ; the output is a compositional form S∈Swhich includes a selection of modules f1​,…,fm​ to be used in that form. SinceΘis already specified, the choice of S completely determines a mapping from inputs to outputs.
It looks that this is a bi-level optimization problem, and use iterative optimization method to solve it.
Notes:
Simulated Annealing
iteration optimization vs bilevel optimization
Reference
Last updated