论文标题

顶级调整:关于转移学习的研究,可提供有效的替代方法,用于通过快速内核方法进行图像分类的微调

Top-Tuning: a study on transfer learning for an efficient alternative to fine tuning for image classification with fast kernel methods

论文作者

Alfano, Paolo Didier, Pastore, Vito Paolo, Rosasco, Lorenzo, Odone, Francesca

论文摘要

深度学习体系结构的令人印象深刻的性能与模型复杂性的巨大增加有关。需要对数百万个参数进行调整,并通过训练时间和推理时间缩放以及能耗。但是,大规模的微调总是必要的吗?在本文中,着眼于图像分类,我们考虑了一种简单的传输学习方法利用预训练的卷积特征作为快速训练核心方法的输入。我们将此方法称为\ textIt {顶级调整},因为仅在目标数据集中对内核分类器进行了训练。在我们的研究中,我们执行3000多个培训过程,重点是32个中小型目标数据集,这是需要转移学习的典型情况。我们表明,最高的调整方法在微调方面提供了可比的准确性,训练时间较小。这些结果表明,顶级调整是在中小型数据集中进行微调的有效替代方法,当训练时间效率和节省计算资源至关重要时,特别有用。

The impressive performance of deep learning architectures is associated with a massive increase in model complexity. Millions of parameters need to be tuned, with training and inference time scaling accordingly, together with energy consumption. But is massive fine-tuning always necessary? In this paper, focusing on image classification, we consider a simple transfer learning approach exploiting pre-trained convolutional features as input for a fast-to-train kernel method. We refer to this approach as \textit{top-tuning} since only the kernel classifier is trained on the target dataset. In our study, we perform more than 3000 training processes focusing on 32 small to medium-sized target datasets, a typical situation where transfer learning is necessary. We show that the top-tuning approach provides comparable accuracy with respect to fine-tuning, with a training time between one and two orders of magnitude smaller. These results suggest that top-tuning is an effective alternative to fine-tuning in small/medium datasets, being especially useful when training time efficiency and computational resources saving are crucial.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源