The best performing learning algorithms devised for event cameras work by first converting events into dense representations that are then processed using standard CNNs. However, these steps discard both the sparsity and high temporal resolution of events, leading to high computational burden and latency. For this reason, recent works have adopted Graph Neural Networks (GNNs), which process events as ``static" spatio-temporal graphs, which are inherently "sparse". We take this trend one step further by introducing Asynchronous, Event-based Graph Neural Networks (AEGNNs), a novel event-processing paradigm that generalizes standard GNNs to process events as ``evolving" spatio-temporal graphs. AEGNNs follow efficient update rules that restrict recomputation of network activations only to the nodes affected by each new event, thereby significantly reducing both computation and latency for event-by-event processing. AEGNNs are easily trained on synchronous inputs and can be converted to efficient, "asynchronous" networks at test time. We thoroughly validate our method on object classification and detection tasks, where we show an up to a 11-fold reduction in computational complexity (FLOPs), with similar or even better performance than state-of-the-art asynchronous methods. This reduction in computation directly translates to an 8-fold reduction in computational latency when compared to standard GNNs, which opens the door to low-latency event-based processing.
翻译:为事件摄影机工作设计的最佳学习算法是首先将事件转换成密集的演示,然后使用标准CNN进行处理。 但是,这些步骤抛弃了事件的紧张性和高时间分辨率,导致计算负担和时空的高度。 因此,最近的工作采用了图形神经网络(GNNS),将事件处理为“静态”的时空图,这在本质上是“偏差”的。我们将这一趋势进一步向前迈出一步,引入Asynchronous,以事件为基础的图形神经网络(AEGNNS),这是一个新的事件处理模式,将标准GNNS普遍化为事件处理“动态”的时空图。一个GNNS遵循有效的更新规则,将网络的再化限制到每个新事件所影响的节点,从而大大降低对事件逐项处理的计算和耐久性。一个EGNNNS很容易被训练为同步的低级输入,可以转换为高效的,“不稳性”网络,在测试时间里程中将GNNNNS转换为“动态”的系统。我们彻底地验证了将运行方式的计算方法,在运行中将运行到更精细的计算方法,在运行中将运行中,在运行中将运行中,在运行中将运行中进行更精确的递减为更精确的计算中,在运行中,在运行中将运行到更精确的计算。