论文标题

多任务学习的大量分析和改进

Large Dimensional Analysis and Improvement of Multi Task Learning

论文作者

Tiomoko, Malik, Couillet, Romain, Tiomoko, Hafiz

论文摘要

多任务学习(MTL)有效地利用了多个相关任务中包含的有用信息,以帮助改善所有任务的概括性能。本文对简单的简单分析进行了较大的维度分析,但正如我们将看到的,当经过精心调整时,MTL的最低方向支持向量机(LSSVM)版本非常强大,在该政权中,数据的尺寸$ p $及其数字$ n $以相同的速度生长较大。 在对输入数据的轻度假设下,对MTL-LSSVM算法的理论分析首先揭示了该算法所利用的“足够统计数据”及其工作中的相互作用。这些结果表明,令人惊讶的是,MTL-LSSVM的标准方法在很大程度上是次优的,可能会导致负转移的严重影响,但是这些障碍很容易纠正。这些校正变成了改进的MTL-LSSVM算法,该算法只能从其他数据中受益,并且还可以分析其理论性能。 正如最近的许多著作中所证明的和理论上维持的那样,这些较大的维度结果对于广泛的数据分布范围是可靠的,我们目前的实验证实了这一点。具体而言,本文报告了流行数据集对理论和经验性能之间的系统紧密行为,这强烈暗示了所提出的精心调整的MTL-LSSVM方法对真实数据的适用性。这种微调完全基于理论分析,并且不需要任何交叉验证程序。此外,实际数据集上报道的表演几乎超过了更精致,更直观的最直觉的多任务和转移学习方法。

Multi Task Learning (MTL) efficiently leverages useful information contained in multiple related tasks to help improve the generalization performance of all tasks. This article conducts a large dimensional analysis of a simple but, as we shall see, extremely powerful when carefully tuned, Least Square Support Vector Machine (LSSVM) version of MTL, in the regime where the dimension $p$ of the data and their number $n$ grow large at the same rate. Under mild assumptions on the input data, the theoretical analysis of the MTL-LSSVM algorithm first reveals the "sufficient statistics" exploited by the algorithm and their interaction at work. These results demonstrate, as a striking consequence, that the standard approach to MTL-LSSVM is largely suboptimal, can lead to severe effects of negative transfer but that these impairments are easily corrected. These corrections are turned into an improved MTL-LSSVM algorithm which can only benefit from additional data, and the theoretical performance of which is also analyzed. As evidenced and theoretically sustained in numerous recent works, these large dimensional results are robust to broad ranges of data distributions, which our present experiments corroborate. Specifically, the article reports a systematically close behavior between theoretical and empirical performances on popular datasets, which is strongly suggestive of the applicability of the proposed carefully tuned MTL-LSSVM method to real data. This fine-tuning is fully based on the theoretical analysis and does not in particular require any cross validation procedure. Besides, the reported performances on real datasets almost systematically outperform much more elaborate and less intuitive state-of-the-art multi-task and transfer learning methods.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源