ICML2020 9-25-2020
Task-imbalanced meta learning: different tasks need different numbers of gradient steps for adaptation
Deep learning based algorithms usually have a fixed number of iterations in the architecture.
could we learn to stop automatically?
predictive model Fθ\mathcal{F}_\thetaFθ​ : transforms the input x to generate a path of states x1,x2,...,xTx_1, x_2, ...,x_Tx1​,x2​,...,xT​
stop policy πϕ\pi_\phiπϕ​ : sequentially observes the states xtx_txt​ and determines the probability of stop at layer ttt
variational stop time distribution qϕq_\phiqϕ​ : stop time distribution induced by stopping policy πϕ\pi_\phiπϕ​
https://icml.cc/Conferences/2020/ScheduleMultitrack?event=6279arrow-up-right
https://icml.cc/media/Slides/icml/2020/virtual(no-parent)-14-17-00UTC-6279-learning_to_sto.pdfarrow-up-right
https://www.youtube.com/watch?v=3VqzhP44Eicarrow-up-right
Last updated 5 years ago