Noisy Linear Networks
نویسنده
چکیده
During the last few years, (bottleneck) linear neural network architectures have received considerable amounts of interest due to their importance in feature extraction. Suppose that an n-dimensional random vector x is to be compressed into a p-dimensional vector y = c(x), where p < n, in a way that relative to some performance criterion, y contains as much information about x as possible. If the mean square error of the best linear estimate of x from y (the \linear reconstruction error") is used as criterion and x is centered and square integrable, extraction of the rst p principal components of the input covari-ance matrix = Cov(x), a linear compression method, is optimal (Bourlard & Kamp, 1988). I.e., the error function I Ejx?Bc(x)j 2 is minimized for B = U p T ?1 and c(x) = TU 0 p x, where the p columns of U p are mutually perpendicular unit length eigenvectors of associated with the p largest eigenvalues of , and T is an invertible p p matrix. If in addition the data is gaussian, these choices also minimize the mutual information between x and y, cf. e.g. Baldi & Hornik (1992). In this paper, we are concerned with situations where a linear feature extraction process is contaminated by noise. More precisely, suppose that upon presentation of an input x, the network actually computes y = Ax + e, where e is the processing noise. Such noise might e.g. be caused by intrinsic unreliability of the network units or external noise during transmission of the signals. We assume that e is centered with covariance matrix Cov(e) = R and uncorrelated with x. Here, is the noise level, and R describes the noise structure (if e.g. the output units are physically close, it may be unrealistic to assume that the noise components are uncorrelated). Clearly, this is not the only possible noise model. For ease of exposition, we also assume that both and R are strictly positive. In this case, the optimal linear reconstruction By of x from y = Ax + e is obtained with B = B (A) = A 0 (AA 0 +R) + , and the conditional covariance
منابع مشابه
A method to solve the problem of missing data, outlier data and noisy data in order to improve the performance of human and information interaction
Abstract Purpose: Errors in data collection and failure to pay attention to data that are noisy in the collection process for any reason cause problems in data-based analysis and, as a result, wrong decision-making. Therefore, solving the problem of missing or noisy data before processing and analysis is of vital importance in analytical systems. The purpose of this paper is to provide a metho...
متن کاملComparative Study of Particle Swarm Optimization and Genetic Algorithm Applied for Noisy Non-Linear Optimization Problems
Optimization of noisy non-linear problems plays a key role in engineering and design problems. These optimization problems can't be solved effectively by using conventional optimization methods. However, metaheuristic algorithms such as Genetic Algorithm (GA) and Particle Swarm Optimization (PSO) seem very efficient to approach in these problems and became very popular. The efficiency of these ...
متن کاملOn the effect of low-quality node observation on learning over incremental adaptive networks
In this paper, we study the impact of low-quality node on the performance of incremental least mean square (ILMS) adaptive networks. Adaptive networks involve many nodes with adaptation and learning capabilities. Low-quality mode in the performance of a node in a practical sensor network is modeled by the observation of pure noise (its observation noise) that leads to an unreliable measurement....
متن کاملSegmentation of Medical Ultrasound Images Using Convolutional Neural Networks with Noisy Activating Functions
The attempts to segment medical ultrasound images have had limited success than the attempts to segment images from other medical imaging modalities. In this project, we attempt to segment medical ultrasound images using convolutional neural networks (CNNs) with a group of noisy activation functions which have recently been demonstrated to improve the performance of neural networks. We report o...
متن کاملExtension of Cube Attack with Probabilistic Equations and its Application on Cryptanalysis of KATAN Cipher
Cube Attack is a successful case of Algebraic Attack. Cube Attack consists of two phases, linear equation extraction and solving the extracted equation system. Due to the high complexity of equation extraction phase in finding linear equations, we can extract nonlinear ones that could be approximated to linear equations with high probability. The probabilistic equations could be considered as l...
متن کاملIeee Information Theory Society Newsletter President's Column from the Editor It Society Member Honored Scholar One Website for Ieee Transactions on Information Theory Has Gone Live Throughput and Capacity Regions Coding for Noisy Networks
Network information theory aims to establish the limits on information fl ow in noisy networks and the coding schemes that achieve these limits. Although a general theory is yet to be developed, several coding schemes and bounds on the capacity have been developed. This article reviews these results with emphasis on two recently developed coding schemes—noisy network coding and compute–forward....
متن کامل