Neural networks are widely deployed models across many scientific disciplines and commercial endeavors ranging from edge computing and sensing to large-scale signal processing in data centers. The most efficient and well-entrenched method to train such networks is backpropagation, or reverse-mode automatic differentiation. To counter an exponentially increasing energy budget in the artificial intelligence sector, there has been recent interest in analog implementations of neural networks, specifically nanophotonic neural networks for which no analog backpropagation demonstration exists. We design mass-manufacturable silicon photonic neural networks that alternately cascade our custom designed "photonic mesh" accelerator with digitally implemented nonlinearities. These reconfigurable photonic meshes program computationally intensive arbitrary matrix multiplication by setting physical voltages that tune the interference of optically encoded input data propagating through integrated Mach-Zehnder interferometer networks. Here, using our packaged photonic chip, we demonstrate in situ backpropagation for the first time to solve classification tasks and evaluate a new protocol to keep the entire gradient measurement and update of physical device voltages in the analog domain, improving on past theoretical proposals. Our method is made possible by introducing three changes to typical photonic meshes: (1) measurements at optical "grating tap" monitors, (2) bidirectional optical signal propagation automated by fiber switch, and (3) universal generation and readout of optical amplitude and phase. After training, our classification achieves accuracies similar to digital equivalents even in presence of systematic error. Our findings suggest a new training paradigm for photonics-accelerated artificial intelligence based entirely on a physical analog of the popular backpropagation technique.
翻译:神经网络是许多科学学科和商业努力中广泛部署的模型,从边缘计算和遥感到数据中心大规模信号处理。培训这类网络最高效和精密的方法是反向反向反向演化或反向模式自动区分。为了应对人工智能部门急剧增加的能源预算,最近人们有兴趣模拟实施神经网络,特别是纳米光谱神经网络,没有模拟反向演化演示。我们设计了从边缘计算和感测到大规模信号处理的大规模可制造的硅光谱神经网络,从而交替更新了我们定制的“光电流网”和大型信号处理器。最高效和精密的方法是用数字式电流处理的加速器。这些可重新配置的光线性线性线性线性线性线性线性线性线性线性网程序通过设置物理电流调节光编码输入数据的干扰,特别是纳米神经神经神经网络的模拟测试。在这里,我们使用包装的光谱芯片,首次在现场进行反向反向反向分析,并评价新的协议,以保持整个等值的等值后性培训结构模拟模拟模拟模拟智能智能智能智能数据显示并更新系统模拟模拟模拟模拟模拟模拟模拟数据测量测量。