The conditional average treatment effect (CATE) is widely used in personalized medicine to inform therapeutic decisions. However, state-of-the-art methods for CATE estimation (so-called meta-learners) often perform poorly in the presence of low overlap. In this work, we introduce a new approach to tackle this issue and improve the performance of existing meta-learners in the low-overlap regions. Specifically, we introduce Overlap-Adaptive Regularization (OAR) that regularizes target models proportionally to overlap weights so that, informally, the regularization is higher in regions with low overlap. To the best of our knowledge, our OAR is the first approach to leverage overlap weights in the regularization terms of the meta-learners. Our OAR approach is flexible and works with any existing CATE meta-learner: we demonstrate how OAR can be applied to both parametric and non-parametric second-stage models. Furthermore, we propose debiased versions of our OAR that preserve the Neyman-orthogonality of existing meta-learners and thus ensure more robust inference. Through a series of (semi-)synthetic experiments, we demonstrate that our OAR significantly improves CATE estimation in low-overlap settings in comparison to constant regularization.
翻译:条件平均处理效应(CATE)在个性化医疗中被广泛用于指导治疗决策。然而,当前最先进的CATE估计方法(即所谓的元学习器)在数据重叠度较低的情况下往往表现不佳。本研究提出了一种新方法来解决这一问题,并提升现有元学习器在低重叠区域的性能。具体而言,我们引入了重叠自适应正则化(OAR),该方法根据重叠权重按比例对目标模型进行正则化,从而非正式地讲,在重叠度较低的区域施加更强的正则化。据我们所知,我们的OAR是首个在元学习器的正则化项中利用重叠权重的方法。OAR方法具有灵活性,可与任何现有的CATE元学习器结合使用:我们展示了OAR如何应用于参数化和非参数化的第二阶段模型。此外,我们提出了OAR的去偏版本,这些版本保持了现有元学习器的Neyman正交性,从而确保了更稳健的统计推断。通过一系列(半)合成实验,我们证明与恒定正则化相比,我们的OAR在低重叠设置下显著改善了CATE估计。