Deep Hypercomplex Networks for Spatiotemporal Data Processing: Parameter efficiency and superior performance [Hypercomplex Signal and Image Processing]

Research output: Contribution to journalArticlepeer-review

4 Scopus citations

Abstract

Hypercomplex numbers, such as quaternions and octonions, have recently gained attention because of their advantageous properties over real numbers, e.g., in the development of parameter-efficient neural networks. For instance, the 16-component sedenion has the capacity to reduce the number of network parameters by a factor of 16. Moreover, hypercomplex neural networks offer advantages in the processing of spatiotemporal data as they are able to represent variable temporal data divisions through the hypercomplex components. Similarly, they support multimodal learning, with each component representing an individual modality. In this article, the key components of deep learning in the hypercomplex domain are introduced, encompassing concatenation, activation functions, convolution, and batch normalization. The use of the backpropagation algorithm for training hypercomplex networks is discussed in the context of hypercomplex algebra. These concepts are brought together in the design of a ResNet backbone using hypercomplex convolution, which is integrated within a U-Net configuration and applied in weather and traffic forecasting problems. The results demonstrate the superior performance of hypercomplex networks compared to their real-valued counterparts, given a fixed parameter budget, highlighting their potential in spatiotemporal data processing.

Original languageBritish English
Pages (from-to)101-112
Number of pages12
JournalIEEE Signal Processing Magazine
Volume41
Issue number3
DOIs
StatePublished - 2024

Fingerprint

Dive into the research topics of 'Deep Hypercomplex Networks for Spatiotemporal Data Processing: Parameter efficiency and superior performance [Hypercomplex Signal and Image Processing]'. Together they form a unique fingerprint.

Cite this