Editable Neural Networks
ICLR 2020 8-22-2020
Last updated
Was this helpful?
ICLR 2020 8-22-2020
Last updated
Was this helpful?
it is crucially important to correct model mistakes quickly as they appear during training the neural network. In this work, we investigate the problem of neural network editing — how one can efficiently patch a mistake of the model on a particular sample, without influencing the model behavior on other samples. Namely, we propose Editable Training, a model-agnostic training technique that encourages fast editing of the trained model.
: a neural network
: task-specific objective loss function
The goal is to change model's predictions on a subset of inputs, corresponding to misclassified objects, without affecting other inputs, by changing the model parameters .
An editor function could be used to formalized this: , with a constraint:
is the changed parameters
For example, multi-class classification.
where is the desired label.
if under the constraint: ,
The constraint: is satisfied iff
So the goal is how to design the editor neural network.