Speaker
Description
This presentation will cover the foundational principles of deep learning, emphasizing representation learning and inductive biases in real-valued neural networks, before extending to hypercomplex neural network architectures. We present theoretical and experimental results demonstrating the enhanced representational capabilities of hypercomplex networks for complex-valued data and multidimensional signal processing. The theoretical framework encompasses proper activation function design, Wirtinger calculus applications, and its extensions to hypercomplex domains. Various methodological approaches to hypercomplex network architectures are examined. The work provides a comprehensive foundation for understanding and implementing hypercomplex neural networks across different application domains, such as RF spectroscopy.