Current machine learning methods for medical image analysis primarily focus on developing models tailored for their specific tasks, utilizing data within their target domain. These specialized models tend to be data-hungry and often exhibit limitations in generalizing to out-of-distribution samples. In this work, we show that employing models that incorporate multiple domains instead of specialized ones significantly alleviates the limitations observed in specialized models. We refer to this approach as multi-domain model and compare its performance to that of specialized models. For this, we introduce the incorporation of diverse medical image domains, including different imaging modalities like X-ray, MRI, CT, and ultrasound images, as well as various viewpoints such as axial, coronal, and sagittal views. Our findings underscore the superior generalization capabilities of multi-domain models, particularly in scenarios characterized by limited data availability and out-of-distribution, frequently encountered in healthcare applications. The integration of diverse data allows multi-domain models to utilize information across domains, enhancing the overall outcomes substantially. To illustrate, for organ recognition, multi-domain model can enhance accuracy by up to 8% compared to conventional specialized models.
翻译:当前医学影像分析的机器学习方法主要集中于针对特定任务开发模型,并利用其目标领域内的数据。这些专用模型往往需要大量数据,且在泛化至分布外样本时通常表现出局限性。本研究表明,采用融合多领域信息的模型而非专用模型,能显著缓解专用模型所存在的局限性。我们将此方法称为多领域模型,并将其性能与专用模型进行比较。为此,我们引入了多样化的医学影像领域,包括不同的成像模态(如X射线、MRI、CT和超声图像)以及多种视图(如轴向、冠状和矢状面)。我们的研究结果突显了多领域模型卓越的泛化能力,尤其是在医疗应用中常见的数据可用性有限和分布外场景下。多样化数据的整合使多领域模型能够跨领域利用信息,从而大幅提升整体性能。例如,在器官识别任务中,多领域模型相较于传统专用模型可将准确率提升高达8%。