Title
Deep Complex Networks.
Abstract
At present, the vast majority of building blocks, techniques, and architectures for deep learning are based on real-valued operations and representations. However, recent work on recurrent neural networks and older fundamental theoretical analysis suggests that complex numbers could have a richer representational capacity and could also facilitate noise-robust memory retrieval mechanisms. Despite their attractive properties and potential for opening up entirely new neural architectures, complex-valued deep neural networks have been marginalized due to the absence of the building blocks required to design such models. In this work, we provide the key atomic components for complex-valued deep neural networks and apply them to convolutional feed-forward networks and convolutional LSTMs. More precisely, we rely on complex convolutions and present algorithms for complex batch-normalization, complex weight initialization strategies for complex-valued neural nets and we use them in experiments with end-to-end training schemes. We demonstrate that such complex-valued models are competitive with their real-valued counterparts. We test deep complex models on several computer vision tasks, on music transcription using the MusicNet dataset and on Speech Spectrum Prediction using the TIMIT dataset. We achieve state-of-the-art performance on these audio-related tasks.
Year
Venue
DocType
2018
ICLR
Conference
Volume
Citations 
PageRank 
abs/1705.09792
0
0.34
References 
Authors
0
9
Name
Order
Citations
PageRank
Chiheb Trabelsi1272.13
olexa bilaniuk231.48
Dmitriy Serdyuk300.34
Sandeep Subramanian439216.75
João Felipe Santos5708.21
Soroush Mehri6122.56
Negar Rostamzadeh702.70
Yoshua Bengio8426773039.83
Chris Pal92140106.53