Outlying observations can be challenging to handle and adversely affect subsequent analyses, particularly, in complex high-dimensional datasets. Although outliers are not always undesired anomalies in the data and may possess valuable insights, only methods that are robust to outliers are able to accurately identify them and resist their influence. In this paper, we propose a method that generates an ensemble of sparse and diverse predictive models that are resistant to outliers. We show that the ensembles generally outperform single-model sparse and robust methods in high-dimensional prediction tasks. Cross-validation is used to tune model parameters to control levels of sparsity, diversity and resistance to outliers. We establish the finitesample breakdown point of the ensembles and the models that comprise them, and we develop a tailored computing algorithm to learn the ensembles by leveraging recent developments in L0 optimization. Our extensive numerical experiments on synthetic and artificially contaminated real datasets from bioinformatics and cheminformatics demonstrate the competitive advantage of our method over state-of-the-art single-model methods.
翻译:异常观测值可能难以处理,并对后续分析产生不利影响,尤其是在复杂的高维数据集中。尽管异常值并非总是数据中不受欢迎的异常现象,且可能蕴含宝贵信息,但只有对异常值具有鲁棒性的方法才能准确识别它们并抵御其影响。本文提出一种方法,能够生成一组稀疏且多样化的预测模型,这些模型对异常值具有抵抗力。我们证明,在高维预测任务中,集成模型通常优于单模型稀疏鲁棒方法。通过交叉验证调整模型参数,以控制稀疏性、多样性和对异常值的抵抗水平。我们建立了集成模型及其组成模型的有限样本崩溃点,并开发了一种定制计算算法,通过利用L0优化领域的最新进展来学习这些集成模型。我们在合成数据以及来自生物信息学和化学信息学的人工污染真实数据集上进行的大量数值实验表明,与最先进的单模型方法相比,我们的方法具有竞争优势。