Transfer learning (TL) has emerged as a powerful tool for improving estimation and prediction performance by leveraging information from related datasets, with the offset TL (O-TL) being a prevailing implementation. In this paper, we adapt the control-variates (CVS) method for TL and develop CVS-based estimators for scalar-on-function regression, one of the most fundamental models in functional data analysis. These estimators rely exclusively on dataset-specific summary statistics, thereby avoiding the pooling of subject-level data and remaining applicable in privacy-restricted or decentralized settings. We establish, for the first time, a theoretical connection between O-TL and CVS-based TL, showing that these two seemingly distinct TL strategies adjust local estimators in fundamentally similar ways. We further derive convergence rates that explicitly account for the unavoidable but typically overlooked smoothing error arising from discretely observed functional predictors, and clarify how similarity among covariance functions across datasets governs the performance of TL. Numerical studies support the theoretical findings and demonstrate that the proposed methods achieve competitive estimation and prediction performance compared with existing alternatives.
翻译:迁移学习(TL)已成为一种通过利用相关数据集信息来提升估计与预测性能的强大工具,其中偏移迁移学习(O-TL)是一种主流的实现方式。本文采用控制变量(CVS)方法进行迁移学习,并针对函数数据分析中最基础的模型之一——标量对函数回归,开发了基于CVS的估计量。这些估计量完全依赖于数据集特定的汇总统计量,从而避免了主体级数据的汇集,适用于隐私受限或去中心化的场景。我们首次建立了O-TL与基于CVS的TL之间的理论联系,表明这两种看似不同的TL策略在本质上以相似的方式调整局部估计量。我们进一步推导了收敛速率,其中明确考虑了由离散观测的函数预测变量所产生、通常被忽略但不可避免的平滑误差,并阐明了数据集间协方差函数的相似性如何主导TL的性能。数值研究支持了理论发现,并证明所提出的方法相较于现有替代方案,在估计和预测性能上具有竞争力。