Lyapunov functions for convergence of principal component algorithms
نویسنده
چکیده
Recent theoretical analyses of a class of unsupervized Hebbian principal component algorithms have identified its local stability conditions. The only locally stable solution for the subspace P extracted by the network is the principal component subspace P∗. In this paper we use the Lyapunov function approach to discover the global stability characteristics of this class of algorithms. The subspace projection error, least mean squared projection error, and mutual information I are all Lyapunov functions for convergence to the principal subspace, although the various domains of convergence indicated by these Lyapunov functions leave some of P-space uncovered. A modification to I yields a ‘principal subspace information’ Lyapunov function I ′ with a domain of convergence which covers almost all of P-space. This shows that this class of algorithms converges to the principal subspace from almost everywhere.
منابع مشابه
On convergence of sample and population Hilbertian functional principal components
In this article we consider the sequences of sample and population covariance operators for a sequence of arrays of Hilbertian random elements. Then under the assumptions that sequences of the covariance operators norm are uniformly bounded and the sequences of the principal component scores are uniformly sumable, we prove that the convergence of the sequences of covariance operators would impl...
متن کاملGradient algorithms for finding common Lyapunov functions
This paper is concerned with the problem of finding a quadratic common Lyapunov function for a large family of stable linear systems. We present gradient iteration algorithms which give deterministic convergence for finite system families and probabilistic convergence for infinite families.
متن کاملTHE EFFECTS OF INITIAL SAMPLING AND PENALTY FUNCTIONS IN OPTIMAL DESIGN OF TRUSSES USING METAHEURISTIC ALGORITHMS
Although Genetic algorithm (GA), Ant colony (AC) and Particle swarm optimization algorithm (PSO) have already been extended to various types of engineering problems, the effects of initial sampling beside constraints in the efficiency of algorithms, is still an interesting field. In this paper we show that, initial sampling with a special series of constraints play an important role in the conv...
متن کاملSparse Structured Principal Component Analysis and Model Learning for Classification and Quality Detection of Rice Grains
In scientific and commercial fields associated with modern agriculture, the categorization of different rice types and determination of its quality is very important. Various image processing algorithms are applied in recent years to detect different agricultural products. The problem of rice classification and quality detection in this paper is presented based on model learning concepts includ...
متن کاملA Proof of Convergence of the Concave-Convex Procedure Using Zangwill's Theory
The concave-convex procedure (CCCP) is an iterative algorithm that solves d.c. (difference of convex functions) programs as a sequence of convex programs. In machine learning, CCCP is extensively used in many learning algorithms, including sparse support vector machines (SVMs), transductive SVMs, and sparse principal component analysis. Though CCCP is widely used in many applications, its conve...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Neural Networks
دوره 8 شماره
صفحات -
تاریخ انتشار 1995