Stochastic diffusion processes are pervasive in nature, from the seemingly erratic Brownian motion to the complex interactions of synaptically-coupled spiking neurons. Recently, drawing inspiration from Langevin dynamics, neuromorphic diffusion models were proposed and have become one of the major breakthroughs in the field of generative artificial intelligence. Unlike discriminative models that have been well developed to tackle classification or regression tasks, diffusion models as well as other generative models such as ChatGPT aim at creating content based upon contexts learned. However, the more complex algorithms of these models result in high computational costs using today's technologies, creating a bottleneck in their efficiency, and impeding further development. Here, we develop a spintronic voltage-controlled magnetoelectric memory hardware for the neuromorphic diffusion process. The in-memory computing capability of our spintronic devices goes beyond current Von Neumann architecture, where memory and computing units are separated. Together with the non-volatility of magnetic memory, we can achieve high-speed and low-cost computing, which is desirable for the increasing scale of generative models in the current era. We experimentally demonstrate that the hardware-based true random diffusion process can be implemented for image generation and achieve comparable image quality to software-based training as measured by the Frechet inception distance (FID) score, achieving ~10^3 better energy-per-bit-per-area over traditional hardware.
翻译:随机扩散过程在自然界中无处不在,从看似无规律的布朗运动到突触耦合的脉冲神经元之间的复杂相互作用。近期,受朗之万动力学启发,神经形态扩散模型被提出,并已成为生成式人工智能领域的主要突破之一。与已充分发展用于处理分类或回归任务的判别式模型不同,扩散模型以及诸如ChatGPT等其他生成式模型,旨在基于学习到的上下文创造内容。然而,这些模型更复杂的算法导致使用现有技术时计算成本高昂,造成了效率瓶颈,并阻碍了进一步发展。在此,我们为神经形态扩散过程开发了一种自旋电子电压控制磁电存储器硬件。我们自旋电子器件的存内计算能力超越了当前内存与计算单元分离的冯·诺依曼架构。结合磁存储器的非易失性,我们可以实现高速、低成本的计算,这对于当前时代生成式模型日益增长的规模而言是理想的。我们通过实验证明,基于硬件的真随机扩散过程可用于图像生成,并且根据弗雷歇起始距离(FID)分数评估,其图像质量可与基于软件的训练相媲美,同时每比特每面积能耗比传统硬件提升了约10^3倍。