论文标题
元学习的分布依赖性分析
A Distribution-Dependent Analysis of Meta-Learning
论文作者
论文摘要
元学习理论中的一个关键问题是了解任务分布如何影响转移风险,这是元学习者在未知任务分布中绘制的新任务上的预期错误。在本文中,专注于使用高斯噪声和高斯任务(或参数)分布的固定设计线性回归,我们在任何算法的转移风险上给出了分布依赖的下限,而我们还表明,所谓的偏置偏置的正则调节方法的新颖,加权版本能够匹配这些下限的固定固定恒定因子。值得注意的是,加权源自高斯任务分布的协方差。总的来说,我们的结果提供了在这种高斯环境中元学习难度的精确表征。尽管此问题设置看起来很简单,但我们表明它足够丰富,可以统一元学习的“参数共享”和“表示学习”流。特别是,当任务分布的协方差矩阵未知时,获得表示为特殊情况。对于这种情况,我们建议采用EM方法,该方法在我们的情况下被证明可以享受有效的更新。该论文是由EM的经验研究完成的。特别是,我们的实验结果表明,随着任务数量的增加,EM算法可以达到下限,而在表示在表示中,该算法也成功地与其替代方案竞争。
A key problem in the theory of meta-learning is to understand how the task distributions influence transfer risk, the expected error of a meta-learner on a new task drawn from the unknown task distribution. In this paper, focusing on fixed design linear regression with Gaussian noise and a Gaussian task (or parameter) distribution, we give distribution-dependent lower bounds on the transfer risk of any algorithm, while we also show that a novel, weighted version of the so-called biased regularized regression method is able to match these lower bounds up to a fixed constant factor. Notably, the weighting is derived from the covariance of the Gaussian task distribution. Altogether, our results provide a precise characterization of the difficulty of meta-learning in this Gaussian setting. While this problem setting may appear simple, we show that it is rich enough to unify the "parameter sharing" and "representation learning" streams of meta-learning; in particular, representation learning is obtained as the special case when the covariance matrix of the task distribution is unknown. For this case we propose to adopt the EM method, which is shown to enjoy efficient updates in our case. The paper is completed by an empirical study of EM. In particular, our experimental results show that the EM algorithm can attain the lower bound as the number of tasks grows, while the algorithm is also successful in competing with its alternatives when used in a representation learning context.