Multivariate skewness and kurtosis measures with an application in ICA

Similar documents
Multivariate Skewness: Measures, Properties and Applications

1/42. Wirtschaftsuniversität Wien, Nicola Loperfido, Urbino University

On the Distribution and Its Properties of the Sum of a Normal and a Doubly Truncated Normal

On the Distribution of Kurtosis Test for Multivariate Normality

A New Multivariate Kurtosis and Its Asymptotic Distribution

Estimating the Parameters of Closed Skew-Normal Distribution Under LINEX Loss Function

An Information Based Methodology for the Change Point Problem Under the Non-central Skew t Distribution with Applications.

Multivariate measures of skewness for the scale mixtures of skew-normal distributions

ON JARQUE-BERA TESTS FOR ASSESSING MULTIVARIATE NORMALITY

ROM Simulation with Exact Means, Covariances, and Multivariate Skewness

Package MultiSkew. June 24, 2017

Modelling strategies for bivariate circular data

On the Distribution of Multivariate Sample Skewness for Assessing Multivariate Normality

Lecture Note 9 of Bus 41914, Spring Multivariate Volatility Models ChicagoBooth

Noureddine Kouaissah, Sergio Ortobelli, Tomas Tichy University of Bergamo, Italy and VŠB-Technical University of Ostrava, Czech Republic

Quasi-Monte Carlo for Finance

A Family of Kurtosis Orderings for Multivariate Distributions

Asymptotic Distribution Free Interval Estimation

Australian Journal of Basic and Applied Sciences. Conditional Maximum Likelihood Estimation For Survival Function Using Cox Model

A Skewed Truncated Cauchy Logistic. Distribution and its Moments

Course information FN3142 Quantitative finance

Approximating a multifactor di usion on a tree.

discussion Papers Some Flexible Parametric Models for Partially Adaptive Estimators of Econometric Models

A Convenient Way of Generating Normal Random Variables Using Generalized Exponential Distribution

2.1 Properties of PDFs

ROM SIMULATION Exact Moment Simulation using Random Orthogonal Matrices

Technical Note: An Improved Range Chart for Normal and Long-Tailed Symmetrical Distributions

arxiv: v1 [math.st] 18 Sep 2018

Analyzing Oil Futures with a Dynamic Nelson-Siegel Model

Solution of the problem of the identified minimum for the tri-variate normal

Fitting financial time series returns distributions: a mixture normality approach

Two-term Edgeworth expansions of the distributions of fit indexes under fixed alternatives in covariance structure models

A Skewed Truncated Cauchy Uniform Distribution and Its Moments

Chapter 14. The Multi-Underlying Black-Scholes Model and Correlation

The rth moment of a real-valued random variable X with density f(x) is. x r f(x) dx

Probability and Random Variables A FINANCIAL TIMES COMPANY

Random Variables and Probability Distributions

Empirical Analysis of the US Swap Curve Gough, O., Juneja, J.A., Nowman, K.B. and Van Dellen, S.

Calibration approach estimators in stratified sampling

Market Risk Analysis Volume I

STOCHASTIC CALCULUS AND BLACK-SCHOLES MODEL

Optimal weights for the MSCI North America index. Optimal weights for the MSCI Europe index

Multivariate longitudinal data analysis for actuarial applications

A Markov Chain Approach. To Multi-Risk Strata Mortality Modeling. Dale Borowiak. Department of Statistics University of Akron Akron, Ohio 44325

A Correlated Sampling Method for Multivariate Normal and Log-normal Distributions

KURTOSIS OF THE LOGISTIC-EXPONENTIAL SURVIVAL DISTRIBUTION

Dynamic Portfolio Execution Detailed Proofs

GENERATION OF STANDARD NORMAL RANDOM NUMBERS. Naveen Kumar Boiroju and M. Krishna Reddy

ESTIMATION OF MODIFIED MEASURE OF SKEWNESS. Elsayed Ali Habib *

2. Copula Methods Background

Lindner, Szimayer: A Limit Theorem for Copulas

Modeling Credit Risk of Loan Portfolios in the Presence of Autocorrelation (Part 2)

On Describing Multivariate Skewed Distributions: A Directional Approach

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

A New Test for Correlation on Bivariate Nonnormal Distributions

A RIDGE REGRESSION ESTIMATION APPROACH WHEN MULTICOLLINEARITY IS PRESENT

A THREE-FACTOR CONVERGENCE MODEL OF INTEREST RATES

Optimal stopping problems for a Brownian motion with a disorder on a finite interval

The Economic and Social BOOTSTRAPPING Review, Vol. 31, No. THE 4, R/S October, STATISTIC 2000, pp

Geoffrey J. McLachlan, Sharon X. Lee Department of Mathematics, University of Queensland, St. Lucia, Brisbane, Australia

MODELLING OF INCOME AND WAGE DISTRIBUTION USING THE METHOD OF L-MOMENTS OF PARAMETER ESTIMATION

Introduction Recently the importance of modelling dependent insurance and reinsurance risks has attracted the attention of actuarial practitioners and

The mean-variance portfolio choice framework and its generalizations

Portfolio Choice. := δi j, the basis is orthonormal. Expressed in terms of the natural basis, x = j. x j x j,

Comparison of MINQUE and Simple Estimate of the Error Variance in the General Linear Models

A Numerical Approach to the Estimation of Search Effort in a Search for a Moving Object

Equivalence between Semimartingales and Itô Processes

Modified ratio estimators of population mean using linear combination of co-efficient of skewness and quartile deviation

Applied Mathematics Letters

The University of Chicago, Booth School of Business Business 41202, Spring Quarter 2009, Mr. Ruey S. Tsay. Solutions to Final Exam

Bivariate Birnbaum-Saunders Distribution

General structural model Part 2: Nonnormality. Psychology 588: Covariance structure and factor models

The Two Sample T-test with One Variance Unknown

BROWNIAN MOTION Antonella Basso, Martina Nardon

Maximum Likelihood Estimation

Edgeworth Binomial Trees

An Improved Skewness Measure

Probability Distributions II

A NEW NOTION OF TRANSITIVE RELATIVE RETURN RATE AND ITS APPLICATIONS USING STOCHASTIC DIFFERENTIAL EQUATIONS. Burhaneddin İZGİ

Empirical Test of Affine Stochastic Discount Factor Model of Currency Pricing. Abstract

Dynamic Replication of Non-Maturing Assets and Liabilities

STRATEGIC PAYOFFS OF NORMAL DISTRIBUTIONBUMP INTO NASH EQUILIBRIUMIN 2 2 GAME

1102 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 51, NO. 3, MARCH Genyuan Wang and Xiang-Gen Xia, Senior Member, IEEE

A Comparison Between Skew-logistic and Skew-normal Distributions

symmys.com 3.2 Projection of the invariants to the investment horizon

SAMPLE STANDARD DEVIATION(s) CHART UNDER THE ASSUMPTION OF MODERATENESS AND ITS PERFORMANCE ANALYSIS

SELECTION OF INDEPENDENT FACTOR MODEL IN FINANCE. Lai-Wan Chan and Siu-Ming Cha

Exercises * on Independent Component Analysis

Simulation of Moment, Cumulant, Kurtosis and the Characteristics Function of Dagum Distribution

An Improved Saddlepoint Approximation Based on the Negative Binomial Distribution for the General Birth Process

A Test of the Normality Assumption in the Ordered Probit Model *

Testing the significance of the RV coefficient

Mossin s Theorem for Upper-Limit Insurance Policies

Log-linear Modeling Under Generalized Inverse Sampling Scheme

EX-POST VERIFICATION OF PREDICTION MODELS OF WAGE DISTRIBUTIONS

Multi-dimensional Term Structure Models

A No-Arbitrage Theorem for Uncertain Stock Model

Developmental Math An Open Program Unit 12 Factoring First Edition

Highly Persistent Finite-State Markov Chains with Non-Zero Skewness and Excess Kurtosis

Information Processing and Limited Liability

Transcription:

Journal of Multivariate Analysis 99 (2008) 2328 2338 www.elsevier.com/locate/jmva Multivariate skewness and kurtosis measures with an application in ICA Tõnu Kollo Institute of Mathematical Statistics, University of Tartu, J. Liivi 2, 50409 Tartu, Estonia Received 24 May 2006 Available online 10 March 2008 Abstract In this paper skewness and kurtosis characteristics of a multivariate p-dimensional distribution are introduced. The skewness measure is defined as a p-vector while the kurtosis is characterized by a p p- matrix. The introduced notions are extensions of the corresponding measures of Mardia [K.V. Mardia, Measures of multivariate skewness and kurtosis with applications, Biometrika 57 (1970) 519 530] and Móri, Rohatgi & Székely [T.F. Móri, V.K. Rohatgi, G.J. Székely, On multivariate skewness and kurtosis, Theory Probab. Appl. 38 (1993) 547 551]. Basic properties of the characteristics are examined and compared with both the above-mentioned results in the literature. Expressions for the measures of skewness and kurtosis are derived for the multivariate Laplace distribution. The kurtosis matrix is used in Independent Component Analysis (ICA) where the solution of an eigenvalue problem of the kurtosis matrix determines the transformation matrix of interest [A. Hyvärinen, J. Karhunen, E. Oja, Independent Component Analysis, Wiley, New York, 2001]. c 2008 Elsevier Inc. All rights reserved. AMS 2000 subject classifications: 62E20; 62H10; 65C60 Keywords: Independent component analysis; Multivariate cumulants; Multivariate kurtosis; Multivariate moments; Multivariate skewness 1. Introduction, motivation Since Mardia [1] introduced measures of multivariate skewness β 1,p and kurtosis β 2,p they have become somewhat standard characteristics of a multivariate distribution. Let X 1 and X 2 be E-mail address: tonu.kollo@ut.ee. 0047-259X/$ - see front matter c 2008 Elsevier Inc. All rights reserved. doi:10.1016/j.jmva.2008.02.033

T. Kollo / Journal of Multivariate Analysis 99 (2008) 2328 2338 2329 Fig. 1. Bivariate Laplace distributions. independent identically distributed copies of a random p-vector X with expectation EX = µ and dispersion matrix DX =. The population measures of p-variate skewness and kurtosis are respectively (Mardia [1,3]) β 1,p (X) = E[(X 1 µ) 1 (X 2 µ)] 3 ; (1.1) β 2,p (X) = E[(X µ) 1 (X µ)] 2. (1.2) Shortcomings of Mardia s characteristic (1.1) are carefully examined in Gutjahr et al. [4]. They point out that Mardia s skewness measure β 1,p equals zero not only in the case of multivariate normality, but also within the much wider class of elliptically symmetric distributions. Therefore using it in normality-tests must be done with precaution. A need for a different skewness measure is motivated by the following circumstances. In recent years several new distribution families have been introduced for modelling skewed data. The multivariate skew normal distribution was introduced in [5], different skew elliptical distributions are presented in the collective monograph [6]. The asymmetric multivariate Laplace distribution is examined with applications in [7] while the multivariate skew t-distribution is under consideration in [8]. Typically these families are characterized by three parameters: one vector as the shift parameter, another vector-parameter controlling location and skewness and a matrix scale parameter. Unfortunately parameter estimation creates problems for all these families. The maximum likelihood method can give wrong answers (the skew symmetric normal distribution) or cannot be applied as an explicit density expression is missing (the multivariate asymmetric Laplace distribution). The moment method can be applied but at least the first three moments are needed. It would be good to have a characteristic of skewness as a p-vector to determine one vector-parameter. Univariate Mardia characteristics (1.1) and (1.2) do not help much in this context: these characteristics can have the same numerical values for distributions with different shapes. This is demonstrated in Fig. 1 where density estimates of two bivariate Laplace distributions with the same Mardia characteristics are sketched. For both distributions β 1,p = 5.63 and β 2,p = 20. Several suggestions and generalizations for modifying multivariate skewness and kurtosis measures have been made. For recent treatments we refer to Klar [9], who gives a thorough overview of the problems, introduces a modified skewness measure and examines asymptotic distributions of different multivariate skewness and kurtosis characteristics. Malkovich and Afifi [10] suggest examining univariate skewness and kurtosis of linear combinations of the

2330 T. Kollo / Journal of Multivariate Analysis 99 (2008) 2328 2338 coordinates of the initial vector X. A different approach is suggested in Móri, Rohatgi and Székely [2] who define multivariate skewness as a p-vector: s(x) = E( Y 2 Y) and multivariate kurtosis as a p p-matrix (1.3) where K(X) = E(YY YY ) (p + 2)I p, (1.4) Y = 1/2 (X µ) and 1/2 is any symmetric square root of. The kurtosis matrix K(X) has been used in Independent Component Analysis [11]. We shall examine some properties of these characteristics later on. One disadvantage of Mardia s kurtosis β 2,p was pointed out by Koziol [12]. He noticed that not all fourth order mixed central moments are taken into account in (1.2) and suggested a new characteristic β 2,p (X) = { i, j,k,l (1.5) E[(Y i Y j Y k Y l )] 2 } 1/2. (1.6) We are going to introduce multivariate skewness and kurtosis characteristics analogous to (1.3) and (1.4). A kurtosis characteristic as a p p-matrix can be applied for at least two purposes. The first is the aforementioned application in ICA. The second concerns parameter estimation. There exist several p-dimensional distribution families with two p p parameter matrices (see [6], Chapter 2, for example). One of them can be estimated using the sample covariance matrix, for another one we could use a p p-matrix constructed from the fourth order mixed moments. In our definitions all mixed third and fourth order moments will be taken into account. At the same time, compressing mixed moments is necessary as the huge matrices of the third and fourth order mixed moments are not perspicuous and do not help in solving the aforementioned estimation problems without transforming the matrices. The introduced notions will be compared with the existing ones. In Section 2 we shall give necessary notation and notions. In Section 3 we shall deal with skewness measures, and Section 4 is devoted to the kurtosis. Finally in Section 5 we apply the new kurtosis measure to determine the ICA transformation matrix. 2. Notation and notions We need some basic notions from matrix algebra and probability theory. The Kronecker product A B of A : m n and B : p q is defined as a partitioned matrix A B = [a i j B], i = 1,..., m; j = 1,... n, vec A denotes an mn-vector, obtained from an m n-matrix A by stacking its columns in the natural order. K p,q stands for the pq pq commutation matrix consisting of q p- blocks, where the i jth element in the jith block equals 1 and all the other elements are zeros, i = 1,..., m; j = 1,... n. For the properties of the vec-operator, the commutation matrix, the Kronecker product and related matrix algebra the interested reader is referred to Harville [13], Schott [14], Magnus and Neudecker [15] or Kollo and von Rosen [16].

T. Kollo / Journal of Multivariate Analysis 99 (2008) 2328 2338 2331 In Sections 3 and 4 we define skewness and kurtosis measures with the help of the star product of matrices. The star product was introduced in [17], where some basic properties were proved. Definition 1. Let A be an m n-matrix and B an mr ns partitioned matrix consisting of r s- blocks B i j, i = 1,..., m; j = 1,... n. The star product A B of A and B is an r s-matrix A B = m i=1 j=1 n a i j B i j. In the following bold capital letters X, Y, Z stand for random vectors and lower-case bold letters x, y, z for their realizations. Lower-case letters t, u, v,... are used for arbitrary constant vectors. Constant matrices are denoted by capital letters A, B, C,.... If X is a continuous random p-vector then the density function of X is denoted by f X (x). Let X be a continuous random p-vector. The characteristic function ϕ X (t) of X is defined as an expectation: ϕ X (t) = E exp(it T X), t R p and the cumulant function ψ X (t): ψ X (t) = ln ϕ X (t). Moments and cumulants of X can be obtained by differentiation of ϕ X (t) and ψ X (t) respectively (see [18] or [16], for example). The third and the fourth moments of X can be presented in the form m 3 (X) = E(X X X), (2.1) m 4 (X) = E(X X X X ) = E[(XX ) (XX )]. (2.2) The corresponding central moments are m 3 (X) = E[(X µ) (X µ) (X µ)], (2.3) m 4 (X) = E[((X µ)(x µ) ) ((X µ)(x µ) )]. (2.4) The third and the fourth cumulants can be expressed through the corresponding central moments (2.3) and (2.4): c 3 (X) = m 3 (X), (2.5) c 4 (X) = m 4 (X) (I p 2 + K p,p )( ) vec vec. (2.6) It is worth noticing that if Z N p (µ, Σ), then c 4 (X) = m 4 (X) m 4 (Z). (2.7) 3. Skewness In Mardia [3] the skewness measure (1.1) is also presented in an alternative way: β 1,p (X) = i, j,k{e[y i Y j Y k ]} 2,

2332 T. Kollo / Journal of Multivariate Analysis 99 (2008) 2328 2338 where Y = (Y 1,..., Y p ) is defined in (1.5). Kollo and Srivastava [19] have shown that β 1,p can be represented via the third order multivariate moments or cumulants given by (2.1) and (2.5) respectively: β 1,p (X) = tr[m 3 (Y)m 3 (Y)] = tr[c 3 (Y)c 3 (Y)]. This means that all third order central moments of X have been used for finding β 1,p. Let us write out the measure s(x) in (1.3) in terms of the coordinates of Y: s(x) = E( Y 2 Y) = E[(Y Y)Y] = = E ( Y 2 i=1 i Y 1,..., i=1 i=1 Y 2 i Y p). E[Y 2 i Y] One can conclude that not all third order mixed moments appear in the expression of s(x). Now we shall define a skewness characteristic as a p-vector which includes all mixed moments of the third order. Definition 2. Let X be a random p-vector and Y be defined by (1.5). Then a p-vector b(x) is called a skewness vector of X, if b(x) = 1 p p m 3 (Y). (3.1) In terms of the coordinates of Y we have the following representation: [ ] b(x) = E (Y i Y j )Y. i, j Definition 2 is in good agreement with Mardia s skewness measure β 1,p : b(x) 2 = i Y j Y k )] i, j,k[e(y 2 = β 1,p (X). Notice that a similar relation does not hold for s(x). A sample estimate b(x) of the skewness vector (3.1) has the following form: where b(x) = 1 n 1 p p n (y i y i y i), i=1 y i = S 1/2 (x i x), where, in turn, x and S are the sample mean and the sample covariance matrix of the initial sample (x 1,..., x n ). 4. Kurtosis In [19] it is shown that Mardia s kurtosis measure β 2,p (X) can be represented as the trace of the fourth moment of the vector Y given by (2.2) or through the fourth cumulant (formulae (2.6)

T. Kollo / Journal of Multivariate Analysis 99 (2008) 2328 2338 2333 and (2.7)): β 2,p (X) = tr[e(yy YY )]. Rewritten as a sum of expectations this expression becomes β 2,p (X) = E(Yi 2 Y j 2 ). This means that not all mixed fourth order moments are taken into account in the definition of Mardia s kurtosis (1.2). This was pointed out by Koziol [12] who suggested the use of a modified characteristic β 2,p (X) given in (1.6) instead of Mardia s measure β 2,p. Paper [2] introduced a p p-matrix K(X) as a characteristic of kurtosis of a p-vector X: K(X) = E(YY YY ) (p + 2)I p. As was shown by Móri et al. [2], the trace of K(X) can be expressed with the aid of Mardia s β 2,p : β 2,p (X) = tr[k(x) + (p + 2)I p ]. In order to have all mixed fourth moments included into the notion of kurtosis we define it in the following way. Definition 3. Let X be a random p-vector and Y = 1/2 (X µ), where E(X) = µ; D(X) =. Then the kurtosis matrix of X is a p p-matrix B(X) = 1 p p m 4 (Y), where 1 p p is a p p-matrix consisting of ones. A sample estimate B(X) of the kurtosis matrix then is: B(X) = 1 n n 1 p p (y i y i y iy i ), where, again, i=1 y i = S 1/2 (x i x), and x and S are the sample mean and the sample covariance matrix of the initial sample (x 1,..., x n ). It is straightforward to check that the [2] kurtosis matrix can be represented as a star product in the following way: K(X) = I p m 4 (Y) (p + 2)I p. From (2.6) and (2.7) it follows that the matrix K(X) has a representation through cumulants: K(X) = I p c 4 (Y). Applying the definition of the star product we get a representation of the matrix B(X) as a sum: B(X) = E(Y i Y j YY ). (4.1)

2334 T. Kollo / Journal of Multivariate Analysis 99 (2008) 2328 2338 Now the trace function can be easily calculated: tr(b) = i, j,k=1 E(Y i Y j Y 2 k ). Clearly this characteristic does not include all fourth order moments. It turns out that the norm of the matrix B(X) equals Koziol s measure of kurtosis (1.6): B(X) = tr[b (X)B(X)] = vec (B(X))vec(B(X)) [ ] 1/2 = (E(Y i Y j Y k Y l )) 2. i, j,k,l=1 Example 1. Let Z N p (µ, Σ). It follows from (2.7) that m 4 (Z) = (I p 2 + K p,p )( ) + vec vec. Denote Z 0 N p (0, I p ). The vector Z 0 is an analog of the vector Y for X. Then m 4 (Z 0 ) = m 4 (Z 0 ) = (I p 2 + K p,p ) + vec I p vec I p. Let us find the kurtosis matrix B(Z). By Definition 3 B(Z) = 1 p p m 4 (Z 0 ). Applying the star product to the three terms one by one we get the following sum: B(Z) = pi p + 21 p 1 p. For the cumulants we introduce a p p-matrix C(X) analogous to the matrix B(X): C(X) = 1 p p c 4 (Y). (4.2) The analog of C(X) in the approach of Móri et al. [2] is the matrix K(X) in (4.1). As matrices of the fourth order moments and cumulants are simple functions of each other (see (2.6) and (2.7)), we can rewrite the kurtosis matrix B(X) in terms of C(X) using the expression of B(Z) from Example 1: B(X) = C(X) + pi p + 21 p 1 p. (4.3) Let us examine in an example how the different notions vary in the case of Laplace distributions. Example 2. Consider the multivariate Laplace distribution X M L p (θ, ) in the following parametrization [19]. A p-vector X has multivariate Laplace distribution with parameters θ and if the characteristic function of X is of the form: ϕ X (t) = 1 1 it θ 1 2 (t θ) 2 + 1 2 t t. In this parametrization E(X) = θ; D(X) =.

T. Kollo / Journal of Multivariate Analysis 99 (2008) 2328 2338 2335 Kollo and Srivastava [19] have shown that β 1,p (X) = a(a 2 6a + 3(p + 2)); (4.4) β 2,p (X) = 2(p + a)(p + 2) 3a 2, (4.5) where a = θ 1 θ. The measures of skewness s(x) and kurtosis K(X) (formulae (1.3) and (1.4)) of Móri et al. [2] can be found after some algebra: s(x) = (p + 2 a)c; (4.6) K(X) = (2p + 4 + a)i p + (p + 4 3a)cc, (4.7) where a is the same as in (4.4) and (4.5) and c = 1/2 θ. The new characteristics of skewness (3.1) and kurtosis (4.3) are of the form b(x) = (p tr(d))c; (4.8) B(X) = (2p + tr(d))i p + 41 p p + 2(D + D ) + (p 3tr(D))cc, (4.9) where a and c are defined as above and D = 1 p p cc. To see the differences between the different characteristics let us consider the bivariate distribution X M L 2 (θ, ). We first take a symmetric distribution with parameters ( 0 θ = 0) and = ( 0.5 0.5 0.5 1 The different skewness and kurtosis measures have the following values: Mardia ((4.4) and (4.5)): β 12 (X) = 0, β 22 (X) = 16; ). Móri, Rohatgi and Székely ((4.6) and (4.7)): ( ( ) 0 8 0 s(x) = and K(X) = ; 0) 0 8 new characteristics ((4.8) and (4.9)): ( ( ) 0 8 4 b(x) = and B(X) =. 0) 4 8 Next consider an asymmetric distribution with parameters ( ) ( ) 1 1.5 1.5 θ = and =. 1 1.5 3 Then the same characteristics are the following: Mardia ((4.4) and (4.5)): β 12 (X) = 5.63, β 22 (X) = 20; Móri, Rohatgi & Székely ((4.6) and (4.7)): ( ) 2.43 s(x) = and K(X) = 1.22 ( 10.8 1.07 1.07 9.2 ) ;

2336 T. Kollo / Journal of Multivariate Analysis 99 (2008) 2328 2338 new characteristics ((4.8) and (4.9)): ( ) 2.78 b(x) = and B(X) = 2.48 ( ) 11.55 5.97. 5.97 10.59 5. Application to ICA Independent Component Analysis (ICA) is a method for finding initial components from multivariate statistical data [11]. In the basic form the ICA model can be presented in the following way. Denote an unknown p-dimensional multivariate signal by S, an unknown mixing p p-matrix by A and a known centred p-vector by X. It is assumed that the coordinates of the initial vector S are independent and non-normally distributed. The aim is to find the unknown S and A. To simplify the problem we assume that our data is whitened before analysis, i.e. instead of the p-vector X we have the p-vector Z with uncorrelated coordinates Z i, DZ i = 1: Z = VX. Through our initial vector S we have the representation Z = VAS = W S, where W is orthogonal. To solve the problem we have to find the orthogonal transformation matrix W. Then the initial signal will be detected up to the order of coordinates, as S = WZ. In [11] several methods are suggested for finding the orthogonal matrix W. One of them is based on the tensor of the fourth order cumulants (Hyvärinen, et al. [11], Chapter 11). In matrix representation the method can be reduced to the following eigenvalue problem: K(Z)W = W ; W W = I p, where K(Z) is the [2] kurtosis matrix in (1.4). As explained by Hyvärinen, et al. [11] this simple method does not always give a satisfactory solution. One important situation when the method does not work is the case of identically distributed random variables S i. Then (p 1) eigenvalues in are equal and it is not possible to identify the original signal S by the orthogonal matrix W of the eigenvectors of K(Z). We guess that a complication may arise because not all mixed fourth order cumulants have been taken into account in the matrix K(Z). Our aim is to work out an analog of the method based on the eigenvalue problem described above. Instead of the matrix K(Z) we start from the matrix C(Z) which includes all fourth order mixed cumulants. Fortunately in this set-up we are able to find the desired transformation in the case of identically distributed signals although we end up with a different eigenvalue problem. Summarizing, we have the following starting point: Z = W S, where W is orthogonal, the coordinates of S are independent and E(Z) = 0, D(Z) = I p. The fourth order moment of Z equals: m 4 (Z) = E(ZZ ZZ ) = E[(W SS W) (W SS W)]

T. Kollo / Journal of Multivariate Analysis 99 (2008) 2328 2338 2337 = E[(W W) (SS SS )(W W)] = (W W) m 4 (S)(W W). A similar equality holds for the fourth order cumulant matrix, defined in (2.6): c 4 (Z) = (W W) c 4 (S)(W W). Because of independence of the coordinates of S we have c 4 (S) in a very simple form: the p p- matrix c 4 (S) is block-diagonal with the ith diagonal block of the form c 4 (S i )(e i ) d, i = 1,..., p (the notation (e i ) d is explained before the next lemma). Applying equality (4.2) we get the matrix C(Z) = 1 p p c 4 (Z) = = E(Z i Z j ZZ ) (W w i ) c 4 (S)(w j W). The expression thus obtained can be simplified with some matrix algebra. For that we need some additional notation. Let A d stand for the diagonalized matrix A and let a d denote a diagonal matrix with the components of a on the main diagonal. The notation above is used in the next lemma. Lemma. Let M, N, U and V be p p-matrices and a, b be p-vectors. Then (M a )K d (U V)K d (b N) = Ma d (U V)b d N, (5.1) where denotes the elementwise (Hadamard) product of matrices and K d = (K p,p ) d. The statement follows from the fact that the corresponding elements of the matrices on the left- and right-hand sides are equal, as can be checked in a straightforward manner. We are going to apply the lemma to the expression of C(X). Note first that the diagonal matrix c 4 (S) can be expressed as a product: c 4 (S 1 ) 0 0 c 4 (S) = K d I p K d. (5.2) 0 0 c 4 (S p ) Denote c 4 (S 1 ) 0 0 (c 4 (S)) d =. 0 0 c 4 (S p ) Then using (5.1) and (5.2) we get C(Z) = W (w i ) d (c 4 (S)) d (w j ) d W = = W DW, W (w i ) d (w j ) d (c 4 (S)) d W

2338 T. Kollo / Journal of Multivariate Analysis 99 (2008) 2328 2338 where Let (( ) D = (c 4 (S)) d w i i=1 d V = W D 1/2. ( ) ) w i i=1 d Then C(Z) = VV with V V = D. Therefore the vectors v i and v j are orthogonal if i j and v i v i = d i. This means that v i is an eigenvector of C(Z) with length d i corresponding to the eigenvalue d i. The matrix V can be found as a solution of an eigenvalue problem. Our transformation matrix of interest W can be expressed as the product W = D 1/2 V. Acknowledgments The author is grateful to the Ph.D. student Helle Kilgi who made the calculations for the examples and to the Referees and the Associate Editor for valuable remarks which considerably improved the presentation of the paper. The author is grateful to the Estonian Science Foundation for support through the grants ETF5686 and ETF7435. References [1] K.V. Mardia, Measures of multivariate skewness and kurtosis with applications, Biometrika 57 (1970) 519 530. [2] T.F. Móri, V.K. Rohatgi, G.J. Székely, On multivariate skewness and kurtosis, Theory Probab. Appl. 38 (1993) 547 551. [3] K.V. Mardia, Applications of some measures of multivariate skewness and kurtosis in testing normality and robustness studies, Sankhya Ser. B 36 (1974) 115 128. [4] S. Gutjahr, N. Henze, M. Folkers, Shortcomings of generalized affine invariant skewness measures, J. Multivariate Anal. 71 (1999) 1 23. [5] A. Azzalini, A. Dalla Valle, The multivariate skew normal distribution, Biometrika 83 (1996) 715 726. [6] M.G. Genton (Ed.), Skew-Elliptical Distributions and Their Applications. A Journey Beyond Normality, Chapman & Hall/CRC, Boca Raton, 2004. [7] S. Kotz, T.J. Kozubowski, K. Podgorski, The Laplace Distribution, and Generalizations: A Revisit with Applications to Communications, Economics, Engineering and Finance, Birkhäuser, Boston, 2001. [8] S. Kotz, S. Nadarajah, Multivariate t Distributions and Their Applications, Cambridge University Press, Cambridge, 2004. [9] B. Klar, A treatment of multivariate skewness, kurtosis, and related statistics, J. Multivariate Anal. 83 (2002) 141 165. [10] J.F. Malkovich, A.A. Afifi, On tests for multivariate normality, J. Amer. Statist. Assoc. 68 (1973) 176 179. [11] A. Hyvärinen, J. Karhunen, E. Oja, Independent Component Analysis, Wiley, New York, 2001. [12] J.A. Koziol, A note on measures of multivariate kurtosis, Biometrical J. 31 (1989) 619 624. [13] D.A. Harville, Matrix Algebra from a Statistician s Perspective, Springer, New York, 1997. [14] J.R. Schott, Matrix Analysis for Statistics, Wiley, New York, 1997. [15] J.R. Magnus, H. Neudecker, Matrix Differential Calculus with Applications in Statistics and Econometrics, second edition, Wiley, Chichester, 1999. [16] T. Kollo, D. von Rosen, Advanced Multivariate Statistics with Matrices, Springer, Dordrecht, 2005. [17] E.C. MacRae, Matrix derivatives with an application to an adaptive linear decision problem, Ann. Statist. 2 (1974) 337 346. [18] T. Kollo, Matrix Derivative for Multivariate Statistics, Tartu University Press, Tartu, 1991 (in Russian). [19] T. Kollo, M.S. Srivastava, Estimation and testing of parameters in multivariate Laplace distribution, Comm. Statist. 33 (2004) 2363 2687..