Approximation by Fully Complex Multilayer Perceptrons

Author:

Kim Taehwan1,Adalı Tülay2

Affiliation:

1. MITRE Corporation, McLean, Virginia 22102 U.S.A.,

2. Department of Computer Science and Electrical Engineering, University of Maryland Baltimore County, Baltimore, Maryland 21250 U.S.A.,

Abstract

We investigate the approximation ability of a multi layer perceptron (MLP) network when it is extended to the complex domain. The main challenge for processing complex data with neural networks has been the lack of bounded and analytic complex nonlinear activation functions in the complex domain, as stated by Liouville's theorem. To avoid the conflict between the boundedness and the analyticity of a nonlinear complex function in the complex domain, a number of ad hoc MLPs that include using two real-valued MLPs, one processing the real part and the other processing the imaginary part, have been traditionally employed. However, since nonanalytic functions do not meet the Cauchy-Riemann conditions, they render themselves into degenerative backpropagation algorithms that compromise the efficiency of nonlinear approximation and learning in the complex vector field. A number of elementary transcendental functions (ETFs) derivable from the entire exponential functionez that are analytic are defined as fully complex activation functions and are shown to provide a parsimonious structure for processing data in the complex domain and address most of the shortcomings of the traditional approach. The introduction of ETFs, however, raises a new question in the approximation capability of this fully complex MLP. In this letter, three proofs of the approximation capability of the fully complex MLP are provided based on the characteristics of singularity among ETFs. First, the fully complex MLPs with continuous ETFs over a compact set in the complex vector field are shown to be the universal approximator of any continuous complex mappings. The complex universal approximation theorem extends to bounded measurable ETFs possessing a removable singularity. Finally, it is shown that the output of complex MLPs using ETFs with isolated and essential singularities uniformly converges to any nonlinear mapping in the deleted annulus of singularity nearest to the origin.

Publisher

MIT Press - Journals

Subject

Cognitive Neuroscience,Arts and Humanities (miscellaneous)

Cited by 143 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Machine Learning Methods for Geotechnical Site Characterization and Scour Assessment;Transportation Research Record: Journal of the Transportation Research Board;2024-07-25

2. A hybrid complex spectral conjugate gradient learning algorithm for complex-valued data processing;Engineering Applications of Artificial Intelligence;2024-07

3. GNSS Anomaly Detection with Complex-Valued LSTM Networks;2024 International Conference on Localization and GNSS (ICL-GNSS);2024-06-25

4. Wave-Function Network Description and Kolmogorov Complexity of Quantum Many-Body Systems;Physical Review X;2024-05-21

5. Complex-Valued Convolutional Gated Recurrent Neural Network for Ultrasound Beamforming;IEEE Transactions on Neural Networks and Learning Systems;2024

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3