Editable Neural Networks
ICLR 2020 8-22-2020
Motivation
it is crucially important to correct model mistakes quickly as they appear during training the neural network. In this work, we investigate the problem of neural network editing — how one can efficiently patch a mistake of the model on a particular sample, without influencing the model behavior on other samples. Namely, we propose Editable Training, a model-agnostic training technique that encourages fast editing of the trained model.
Editing Neural Networks
: a neural network
: task-specific objective loss function
The goal is to change model's predictions on a subset of inputs, corresponding to misclassified objects, without affecting other inputs, by changing the model parameters .
An editor function could be used to formalized this: , with a constraint:
is the changed parameters
For example, multi-class classification.
where is the desired label.
if under the constraint: ,
The constraint: is satisfied iff
So the goal is how to design the editor neural network.
Reference:
Last updated