The conditional average treatment effect (CATE) is widely used in personalized medicine to inform therapeutic decisions. However, state-of-the-art methods for CATE estimation (so-called meta-learners) often perform poorly in the presence of low overlap. In this work, we introduce a new approach to tackle this issue and improve the performance of existing meta-learners in the low-overlap regions. Specifically, we introduce Overlap-Adaptive Regularization (OAR) that regularizes target models proportionally to overlap weights so that, informally, the regularization is higher in regions with low overlap. To the best of our knowledge, our OAR is the first approach to leverage overlap weights in the regularization terms of the meta-learners. Our OAR approach is flexible and works with any existing CATE meta-learner: we demonstrate how OAR can be applied to both parametric and non-parametric second-stage models. Furthermore, we propose debiased versions of our OAR that preserve the Neyman-orthogonality of existing meta-learners and thus ensure more robust inference. Through a series of (semi-)synthetic experiments, we demonstrate that our OAR significantly improves CATE estimation in low-overlap settings in comparison to constant regularization.
翻译:条件平均处理效应(CATE)在个性化医疗中被广泛用于指导治疗决策。然而,当前最先进的CATE估计方法(即元学习器)在数据重叠度较低时往往表现不佳。本研究提出了一种新方法来解决这一问题,并提升现有元学习器在低重叠区域的性能。具体而言,我们提出了重叠度自适应正则化(OAR)方法,该方法根据重叠权重按比例正则化目标模型,使得在低重叠区域施加更强的正则化约束。据我们所知,我们的OAR是首个在元学习器正则化项中利用重叠权重的方法。OAR具有灵活性,可与任何现有CATE元学习器结合使用:我们展示了如何将OAR应用于参数化和非参数化的第二阶段模型。此外,我们提出了去偏版本的OAR,该版本保持了现有元学习器的Neyman正交性,从而确保更稳健的统计推断。通过一系列(半)合成实验,我们证明相较于恒定正则化方法,我们的OAR能显著提升低重叠场景下的CATE估计精度。