导航
学术报告|
报告题目: Nonlinear Meta-learning Can Guarantee Faster Rates
报 告 人: Dr. Zhu Li
报告人所在单位: University College London
报告日期: 2024-05-13
报告时间: 14:30-15:30
报告地点: 光华楼东主楼2001室
   
报告摘要:

Many recent theoretical works on meta-learning aim to achieve guarantees in leveraging similar representational structures from related tasks towards simplifying a target task. The main aim of theoretical guarantees on the subject is to establish the extent to which convergence rates---in learning a common representation---may scale with the number N of tasks (as well as the number of samples per task). First steps in this setting demonstrate this property when both the shared representation amongst tasks, and task-specific regression functions, are linear. This linear setting readily reveals the benefits of aggregating tasks, e.g., via averaging arguments. In practice, however, the representation is often highly nonlinear, introducing nontrivial biases in each task that cannot easily be averaged out as in the linear case.

In the present work, we derive theoretical guarantees for meta-learning with nonlinear representations. In particular, assuming the shared nonlinearity maps to an infinite-dimensional reproducing kernel Hilbert spaces, we show that additional biases can be mitigated with careful regularization that leverages the smoothness of task-specific regression functions, yielding improved rates that scale with the number of tasks as desired.

学术海报.pdf

   
本年度学院报告总序号: 1204

版权所有 Copyright © beat·365(中国)官方网站-最新版2022 沪ICP备042465