Signal Processing 104 (2014) 319–324
Contents lists available at ScienceDirect
Signal Processing journal homepage: www.elsevier.com/locate/sigpro
Fast communication
New blind source separation method of independent/ dependent sources A. Keziou a,n, H. Fenniri b, A. Ghazdali c, E. Moreau d,e a
Laboratoire de Mathématiques de Reims EA 4535, Université de Reims Champagne-Ardenne, France CReSTIC, Université de Reims Champagne-Ardenne, France LAMAI, FST, Université Cadi-Ayyad, Morocco d ENSAM, LSIS, UMR CNRS 7296, Université du Sud-Toulon-Var, France e Aix Marseille Université, France b c
a r t i c l e i n f o
abstract
Article history: Received 24 January 2014 Received in revised form 4 April 2014 Accepted 19 April 2014 Available online 30 April 2014
We introduce a new blind source separation approach, based on modified Kullback– Leibler divergence between copula densities, for both independent and dependent source component signals. In the classical case of independent source components, the proposed method generalizes the mutual information (between probability densities) procedure. Moreover, it has the great advantage to be naturally extensible to separate mixtures of dependent source components. Simulation results are presented showing the convergence and the efficiency of the proposed algorithms. & 2014 Elsevier B.V. All rights reserved.
Keywords: Blind source separation Modified Kullback–Leibler divergence between copulas Mutual information
1. Introduction Blind source separation (BSS) is an instrumental problem in signal processing which has been addressed in the last three decades. We consider an instantaneous linear mixture described by xðtÞ ≔AsðtÞ þnðtÞ A Rp ;
ð1Þ
where A A Rpp is an unknown non-singular mixing matrix, sðtÞ ≔ðs1 ðtÞ; …; sp ðtÞÞ > is the unknown vector of source signals to be estimated from xðtÞ ≔ðx1 ðtÞ; …; xp ðtÞÞ > , the vector of observed signals. The number of sources and the number of observations, for the present work, are assumed to be equal. The presence of additive noise nðtÞ within the mixing model complicates significantly the BSS problem. It is reduced by applying some form of n
Corresponding author. E-mail addresses:
[email protected] (A. Keziou),
[email protected] (H. Fenniri),
[email protected] (A. Ghazdali),
[email protected] (E. Moreau). http://dx.doi.org/10.1016/j.sigpro.2014.04.017 0165-1684/& 2014 Elsevier B.V. All rights reserved.
preprocessing such as denoising the observed signals through regularization approach, see e.g. [15]. The goal is to estimate the vector source signals sðtÞ using only the observed signals xðtÞ. The estimate yðtÞ of the source signals sðtÞ can be written as yðtÞ ¼ BxðtÞ;
ð2Þ pp
where B A R is the de-mixing matrix. The question is b which has to be how to obtain the de-mixing matrix B close to the ideal solution A 1 , using only the observed signals xðtÞ? It is well known, by Darmois theorem, that if the source components are mutually independent and at b most one component is Gaussian, a consistent estimate B of A 1 (up to scale and permutation indeterminacies of rows) is the one that makes the components of the vector yðtÞ independent; see e.g. [6]. The corresponding signals b b ðtÞ ≔BxðtÞ y are the estimate of the source signals sðtÞ. Under the above hypotheses, many procedures have been proposed in the literature. Some of these procedures use second or higher order statistics, see e.g. [13,2] and the references therein, others consist of optimizing (on the
320
A. Keziou et al. / Signal Processing 104 (2014) 319–324
de-mixing matrix space) an estimate of some measure of dependency structure of the components of the vector yðtÞ. As measures of dependence used in BSS, we find in the literature the criterion of mutual information (MI) [14,7], the criteria of α, β and Renyi's-divergences [5,19], and the criteria of ϕ-divergences [15]. The procedures based on minimizing estimates of MI are considered as the most efficient, since this criterion can be estimated efficiently, other procedures using divergences lead to robust method for appropriate choice of divergence criterion [15]. In this paper, we will focus on the criterion of MI (called also modified Kullback–Leibler divergence), viewed as measure of difference between copula densities, and we will use it to propose a new BSS approach that applies in both cases of independent or dependent source components. In the following, we will show that the mutual information of a random vector Y ≔ðY 1 ; …; Y p Þ > can be written as the modified Kullback–Leibler divergence (KLm-divergence) between the copula of independence and the copula of the vector. Then, we propose a separation procedure based on minimizing an appropriate estimate of KLm-divergence between the copula density of independence and the copula density of the vector. This approach applies in the standard case, and we will show that the proposed criterion can be naturally extended to separate mixture of dependent source components. The proposed approach can be adapted also to separate complex-valued signals. In all the sequel, we assume that at most one source is Gaussian, and we will treat separately the case of independent source components, and then the case of dependent source components. Chen et al. [3] proposed a BSS algorithm (for independent source components) based on minimizing a distance between the parameter of the copula of the estimated source and the value of the parameter corresponding to copula of independence. Ma and Sun [9] proposed a different criterion combining the MI between probability densities and Shannon entropy of semiparametric models of copulas. 2. Brief recalls on copulas Consider a random vector Y ≔ðY 1 ; …; Y p Þ > A Rp , p Z1, with joint distribution function (d.f.) FY ðÞ: y A Rp ↦FY ðyÞ ≔ FY ðy1 ; …; yp Þ ≔PðY 1 ry1 ; …; Y p ryp Þ, and continuous marginal d.f.'s F Y j ðÞ: yj A R↦F Y j ðyj Þ ≔PðY j r yj Þ, 8 j ¼ 1; …; p. The characterization theorem of Sklar [17] shows that there exists a unique p-variate function CY ðÞ: ½0; 1p ↦½0; 1, such that, FY ðyÞ ¼ CY ðF Y 1 ðy1 Þ; …; F Y p ðyp ÞÞ, 8 y ≔ðy1 ; …; yp Þ > A Rp . The function CY ðÞ is called a copula and it is in itself a joint d.f. on ½0; 1p with uniform marginals. We have for all u ≔ðu1 ; …; up Þ > A ½0; 1p , CY ðuÞ ¼ PðF Y 1 ðY 1 Þ r u1 ; …; F Y p ðY p Þ r up Þ. Conversely, for any marginal d.f.'s F 1 ðÞ; …; F p ðÞ, and any copula function CðÞ, the function CðF 1 ðy1 Þ; …; F p ðyp ÞÞ is a multivariate d.f. on Rp . On the other hand, since the marginal d.f.'s F Y j ðÞ, j ¼ 1; …; p, are assumed to be continuous, then the random variables F Y 1 ðY 1 Þ; …; F Y p ðY p Þ are uniformly distributed on the interval [0, 1]. So, if the components Y 1 ; …; Y p are statistically independent, then the corresponding copula writes C0 ðuÞ ≔∏pj¼ 1 uj ; 8 u A ½0; 1p . It is called the copula of independence. Define, when it exists, the copula density (of the random vector Y)
cY ðuÞ ≔ð∂p =∂u1 ⋯∂up ÞCY ðuÞ; 8 u A ½0; 1p . Hence, the copula density of independence c0 ðÞ is the function taking the value 1 on ½0; 1p and zero otherwise, namely, c0 ðuÞ ≔1½0;1p ðuÞ;
8 u A ½0; 1p :
ð3Þ
Let f Y ðÞ, if it exists, be the probability density on Rp of the random vector Y ≔ðY 1 ; …; Y p Þ > , and, respectively, f Y 1 ðÞ; …; f Y p ðÞ, the marginal probability densities of the random variables Y 1 ; …; Y p . Then, a straightforward computation shows that, for all y A Rp , we have p
f Y ðyÞ ¼ ∏ f Y j ðyj ÞcY ðuÞ;
ð4Þ
j¼1
where u ≔ðu1 ; …; up Þ > ≔ðF Y 1 ðy1 Þ; …; F Y p ðyp ÞÞ > . In the monographs by [11,8], the reader may find detailed ingredients of the modeling theory as well as surveys of the commonly used semiparametric copulas. 3. Mutual information and copulas The MI of a random vector Y ≔ðY 1 ; …; Y p Þ > A Rp is defined by Z ∏pj¼ 1 f Y j ðyj Þ f Y ðyÞ dy1 ⋯dyp : MI ðY Þ ≔ log ð5Þ f Y ðyÞ Rp It is called also the modified Kullback–Leibler divergence (KLm-divergence) between the product of the marginal densities and the joint density of the vector. Note also that MIðYÞ≕KLm ð∏pj¼ 1 f Y j ; f Y Þ is nonnegative and achieves its minimum value zero if and only if (iff) f Y ðÞ ¼ ∏pj¼ 1 f Y j ðÞ, i.e., iff the components of the random vector Y are statistically independent. An equivalent formula of (5) is ! ∏pj¼ 1 f Y j ðY j Þ MI ðY Þ ≔E log ; ð6Þ f Y ðYÞ where EðÞ is the mathematical expectation. Using the relation (4), and applying the change variable formula for multiple integrals, we can show that MIðYÞ can be written as Z 1 cY ðuÞ du≕KLm ðc0 ; cY Þ log MI ðY Þ ¼ cY ðuÞ ½0;1p ¼ Eðlog cY ðF Y 1 ðY 1 Þ; …; F Y p ðY p ÞÞÞ≕ HðcY Þ; R where HðcY Þ ≔ ½0;1p logðcY ðuÞÞcY ðuÞ du is the Shannon entropy of the copula density cY ðÞ. The relation above means that the MI of the random vector Y can be seen as the KLm-divergence between the copula density of independent c0 ðÞ, see (3), and the copula density cY ðÞ of the random vector Y. We summarize the above results in the following proposition. Proposition 1. Let Y A Rp be any random vector with continuous marginal distribution functions. Then, the MI of Y can be written as the KLm-divergence between the copula density c0 of independence and the copula density of the vector Y: Z 1 MI ðY Þ ¼ cY ðuÞ du≕KLm ðc0 ; cY Þ log cY ðuÞ ½0;1p ¼ Eðlog cY ðF Y 1 ðY 1 Þ; …; F Y p ðY p ÞÞÞ:
ð7Þ
A. Keziou et al. / Signal Processing 104 (2014) 319–324
Moreover, KLm ðc0 ; cY Þ is nonnegative and takes the value zero iff cY ðuÞ ¼ c0 ðuÞ≕1½0;1p ðuÞ, 8 u A ½0; 1p , i.e., iff the components of the vector Y are statistically independent.
4. A separation procedure for mixtures of independent sources through copulas In this section, we describe our approach based on minimizing a nonparametric estimate of the KLm-divergence KLm ðc0 ; cY Þ, assuming that the source components are independent. Denote by S ≔ðS1 ; …; Sp Þ > the random source vector, X ≔AS the observed random vector and Y ≔BX the estimated random source vector. The discrete (noise free) version of the mixture model (1) writes xðnÞ ≔ AsðnÞ; n ¼ 1; …; N. The source signals sðnÞ; n ¼ 1; …; N, will be considered as N i.i.d. copies of the random source vector S, and then xðnÞ; yðnÞ ≔BxðnÞ; n ¼ 1; …; N, are, respectively, N i.i.d. copies of the random vectors X and Y ≔BX. When the source signal data sðnÞ; n ¼ 1; …; N, are not i.i.d., such as cyclo-stationary sources, see e.g. [10], we can return to the above i.i.d. case, by developing the observed signals xðnÞ; n ¼ 1; …; N, following some harmonic or trigonometric basis, and then separate the error vector components ϵ ≔ðϵ1 ; …; ϵp Þ > from the data ϵðnÞ; n ¼ 1; …; N, which can be assumed i.i.d. copies of ϵ. Note that, by the linearity of the problem, the separating matrix will be the same for the original observed signals xðÞ or the errors signals ϵðÞ. The same methodology can be used for Section 5 below. In view of Proposition 1, the function B↦KLm ðc0 ; cY Þ is non-
density. A more appropriate choice of the kernel kðÞ, for estimating the copula density, can be done according to [12], which copes with the boundary effect. The bandwidth parameters H 1 ; …; H p in (9) and h1 ; …; hp in (10) will be chosen according to Silverman's rule of thumb, see e.g. [16], i.e., for all j ¼ 1; …; p, we take 1=ðp þ 4Þ 1=5 4 b j N 1=ðp þ 4Þ and hj ¼ 4 b j N 1=5 ; Hj ¼ s Σ p þ2 3 b j and s b j are, respectively, the empirical standard where Σ deviation of the data Fb Y ðy ð1ÞÞ; …; Fb Y ðy ðNÞÞ and j
1 N d KL ∑ log b c Y Fb Y 1 y1 ðnÞ ; …; Fb Y p yp ðnÞ ; m ðc0 ; cY Þ ≔ Nn¼1
ð8Þ
j
j
from the proper definitions of the estimates as follows: d b b b d dKL 1 N dB c Y F Y 1 y1 ðnÞ ; …; F Y p yp ðnÞ m ðc0 ; cY Þ ¼ ∑ ; dB Nn¼1 b c YðFb Y 1 ðy1 ðnÞÞ; …; Fb Y p ðyp ðnÞÞÞ ð11Þ where d=dB ≔ð∂=∂Bij Þij , i; j ¼ 1; …; p, and ∂b c Y ðFb Y 1 ðy1 ðnÞÞ; …; Fb Y p ðyp ðnÞÞÞ ∂Bij ¼
N 1 ∑ NH 1 ⋯H p m ¼ 1
0
negative and achieves its minimum value zero iff B ¼ A (up to scale and permutation indeterminacies). In other
b b ðnÞ ¼ BxðnÞ; mated source signals y n ¼ 1; …; N. Based on the equality (7), we propose to estimate the KLm-divergence KLm ðc0 ; cY Þ by a “plug-in” type procedure. We obtain then the estimate
j
b the estimate of the yj ð1Þ; …; yj ðNÞ. In order to compute B, de-mixing matrix, we can use a gradient descent algorithm taking as initial matrix B0 ¼ I p , the p p identity d matrix. The gradient in B of KL m ðc0 ; cY Þ can be explicated
1
words, we have A 1 ¼ arg inf B KLm ðc0 ; cY Þ. Hence, to achieve separation, the idea is to minimize on B some d statistical estimate KL m ðc0 ; cY Þ, of KLm ðc0 ; cY Þ, constructed from the data yð1Þ; …; yðnÞ. The separation matrix is then d b ≔arg inf B KL estimated by B m ðc0 ; cY Þ, leading to the esti-
321
k
p
∏
j ¼ 1;j a i
k
! Fb Y j ðyj ðmÞÞ Fb Y j ðyj ðnÞÞ Hj
1 ! b b Fb Y i ðyi ðmÞÞ Fb Y i ðyi ðnÞÞ 1 ∂ F Y i ðyi ðmÞÞ F Y i ðyi ðnÞÞ A ; ∂Bij Hi Hi
with ∂Fb Y i ðyi ðmÞÞ 1 N y ðℓÞ yi ðmÞ ¼ ∑ k i xj ðℓÞ xj ðmÞ ∂Bij Nhi ℓ ¼ 1 hi and ∂Fb Y i ðyi ðnÞÞ 1 N y ðℓÞ yi ðnÞ ¼ ∑ k i xj ðℓÞ xj ðnÞ : ∂Bij Nhi ℓ ¼ 1 hi
5. A solution to the BSS for dependent source components
where ! p N Fb Y j ðyj ðmÞÞ uj 1 b c Y ð uÞ ≔ ∑ ∏ k ; NH 1 ⋯H p m ¼ 1 j ¼ 1 Hj
ð9Þ
for all u A ½0; 1p , is the kernel estimate of the copula density cY ðÞ, and Fb Y j ðxÞ, 8 j ¼ 1; …; p, is the smoothed estimate of the marginal distribution function F Y j ðxÞ of the random variable Yj, at any real value x A R, defined by yj ðℓÞ x 1 N Fb Y j ðxÞ ≔ ∑ K ; ð10Þ Nℓ¼1 hj where KðÞ is the primitive of a kernel kðÞ, a symmetric centered probability density. In our forthcoming simulation study, we will take as kernel kðÞ a standard Gaussian
We give in the following subsections a solution to the BSS problem for mixtures of dependent component sources. The method depends on the knowledge about the dependency structure of the source components. 5.1. When both the model and the parameter are known Assume that we dispose of some prior information about the density copula of the random source vector S. Note that this is possible for many practical problems, it can be done, from realizations of S, by a model selection procedure in semiparametric copula density models fcθ ðÞ; θ A Θ Rd g, typically indexed by a multivariate parameter θ, see e.g. [4]. The parameter θ can be estimated
322
A. Keziou et al. / Signal Processing 104 (2014) 319–324
using maximum semiparametric likelihood, see [18,1]. Denote by θ0 the obtained value of θ and cθ0 ðÞ the copula density modeling the dependency structure of the source components. Obviously, since the source components are assumed to be dependent, cθ0 ðÞ is different from the density copula of independence c0 ðÞ. Hence, we naturally replace, in the above criterion function B↦KLm ðc0 ; cY Þ, c0 by cθ0 . Moreover, we can show that the criterion function B↦KLm ðcθ0 ; cY Þ is nonnegative and achieves its minimum value zero iff B ¼ A 1 (up to scale and permutation indeterminacies), i.e., A 1 ¼ arg inf B KLm ðcθ0 ; cY Þ, provided that the copula density cθ0 ðÞ of S satisfies the following assumption (C.A): for any regular matrix M, if the copula density of MS is equal to cθ0 ðÞ, then M ¼ DP, where D is diagonal and P is a permutation. Note also that the criterion KLm ðcθ0 ; cY Þ can be written as !! cY ðF Y 1 ðY 1 Þ; …; F Y p ðY p ÞÞ KLm cθ0 ; cY ¼ E log : cθ0 ðF Y 1 ðY 1 Þ; …; F Y p ðY p ÞÞ So as before, we propose to estimate the de-mixing matrix d b ≔arg inf B KL by B m ðcθ0 ; cY Þ, where bY ðFb Y 1 ðy1 ðnÞÞ; …; Fb Y p ðyp ðnÞÞÞ 1 N c d KL ∑ log : m cθ0 ; cY ≔ Nn ¼ 1 cθ0 ðFb Y 1 ðy1 ðnÞÞ; …; Fb Y p ðyp ðnÞÞÞ The estimates of copula density and the marginal distribub can be tion functions are defined as before. The solution B computed by a descent gradient algorithm. The gradient in B can be explicitly computed in a similar way as in (11) b b ðnÞ ¼ BxðnÞ; Section 4. The estimated source signals are y
indeterminacies), provided that the copula density model fcθ ðÞ; θA Θ Rd g of S satisfies the following identifiability assumption (C.B): for any regular matrix M, if the copula density of MS A fcθ ðÞ; θ A Θ Rd g, then M ¼ DP, where D is diagonal and P is a permutation. 5.3. When both the model and the parameter are unknown In this case, we can use the following methodology. We consider a class of L models of copula densities of the source components, denoted M1 ≔fc1θ1 ðÞ; θ1 A Θ1 g; …; M L ≔ fcLθL ðÞ; θL A ΘL g. Here, the goal is to separate the source signal using the true unknown model of dependency of the sources. Assume that each model Mi, i¼1,…,L, satisfies the identifiability condition (C.B) in the above subsection. Then, the criterion function B↦inf ℓ ¼ 1;…;L inf θℓ A Θℓ KLm ðcℓθℓ ; cY Þ is nonnegative and achieves its minimum value zero iff B ¼ A 1 (up to scale and permutation indeterminacies). Hence, we propose to apply the method described in the above subsection for each model, and then choose the solution that minimizes the criterion over all considered models, i.e., ℓn b ≔arg inf inf KL d B m ðcθ ; cY Þ B
θ A Θ ℓn
where
ℓ d ℓn ≔arg min inf inf KL m ðcθ ; cY Þ: ℓ ¼ 1;…;L B
θ A Θℓ
6. Simulation results
n ¼ 1; …; N. We obtain then the following algorithm. Algorithm 1. A copula based BSS algorithm for dependent component sources. Data : the observed signals xðnÞ; n ¼ 1; …; N b ðnÞ; n ¼ 1; …; N Result : the estimated sources y Initialization : y0 ðnÞ ¼ B0 xðnÞ, B0 ¼ I p . Given ε 4 0 and μ 40 suitably choosen Do Update B and y: cm ðcθ ;cY Þ dKL 0 Bk þ 1 ¼ Bk μ dB B ¼ Bk
yk þ 1 ðnÞ ¼ Bk þ 1 xðnÞ; n ¼ 1; …; N Until J Bk þ 1 Bk J o ε b ðnÞ ¼ yk þ 1 ðnÞ; n ¼ 1; …; N y
5.2. When the model is known and the parameter is unknown Denote by fcθ ðÞ; θ A Θ Rd g the semiparametric copula density model of the source components. Since the parameter θ is unknown, we propose to adapt the above criterion and to estimate the demixing matrix by b ≔arg inf B inf θ A Θ KL d B m ðcθ ; cY Þ, which can be computed using the gradient (on both B and θ) of the criterion d function ðB; θÞ↦KL m ðcθ ; cY Þ. Note that the criterion function
B↦inf θ A Θ KLm ðcθ ; cY Þ is nonnegative and achieves its mini-
mum value zero iff B ¼ A 1 (up to scale and permutation
In this section, we give simulation results for the proposed method. We dealt with instantaneous mixtures (2 mixtures of 2 sources) of five kinds of sample sources: uniform i.i.d. with independent components (Fig. 1a); i.i.d. sources with independent components where the marginals are drawn from the 4-ASK (Amplitude Shift Keying) alphabets ( 3, 1,1,3 with equal probabilities 0.25) at which was added a centered Gaussian random variable with variance equal to 0.25 (Fig. 1b); i.i.d. (with uniform marginals) vector sources with dependent components generated from Fairlie–Gumbel–Morgenstern (FGM)copula with θ0 ¼ 0:8 (Fig. 2); i.i.d. (with uniform marginals in Fig. 3a and binary phase-shift keying (BPSK)-marginals in Fig. 3b) vector sources with dependent components generated from Ali–Mikhail–Haq (AMH)-copula with θ0 ¼ 0:6. The accuracy of source estimation is evaluated through the signal-noise-ratio (SNR), defined by SNRi ≔ 2 2 N b 10 log10 ∑N In the n ¼ 1 si ðnÞ =∑n ¼ 1 ðy i ðnÞ si ðnÞÞ ; i ¼ 1; 2. standard case of independent source components, the proposed method described in Section 4 is compared with the MI one proposed by [14], under the same conditions, see Fig. 1. The used mixing matrix is A ≔½1 0:5; 0:5 1. The number of samples is N ¼2000, and all simulations are repeated 100 times. The gradient descent parameter is taken μ ¼0.1 in all cases. We observe from Fig. 1 that the proposed method gives good results for the standard case of independent component sources. Moreover, we see, from Figs. 2 and 3, that our proposed method is able to separate, with good performance, mixtures of dependent source components.
A. Keziou et al. / Signal Processing 104 (2014) 319–324
55
50
50
45
45
40 35
35
SNR (dB)
SNR (dB)
40
30 25
30 25 20
20
15
15
0
10
20
30
40
50
KLm_C MI
10
KLm_C MI
10 5
323
60
5
70
0
10
20
Iterations
30
40
50
Iterations
Fig. 1. SNRs vs iterations with independent sources. (a) Uniform sources. (b) ASK sources.
50
0.7
45
0.6
40 0.5 criterion value
SNR (dB)
35 30 25
0.4
0.3
20 0.2 15 10 5
0.1
SNR_y1 SNR_y2 0
10
20
30
40
50
0
60
0
10
20
Iterations
30
40
50
60
Iterations
55
55
50
50
45
45
40
40
35
35
SNR (dB)
SNR (dB)
Fig. 2. Separation of dependent sources from FGM copula. (a) SNRs vs iterations. (b) Criterion vs iterations.
30 25
30 25
20
20
15
15 SNR_y1 SNR_y2
10 5
0
50
100 Iterations
150
SNR_y1 SNR_y2
10 200
5
0
50
100 Iterations
Fig. 3. SNRs vs iterations with dependent sources from AMH copula. (a) Uniform sources. (b) BPSK sources.
150
200
324
A. Keziou et al. / Signal Processing 104 (2014) 319–324
7. Conclusion We have proposed a new BSS approach by minimizing the empirical KLm-divergence between copula densities. The approach is able to separate instantaneous linear mixtures of both independent and dependent source components. In Section 6, the accuracy and the consistency of the obtained algorithms are illustrated by simulation, for 2 2 mixture-source. Other simulation examples give similar results for the case of 3 3 mixture-source. It should be mentioned that our proposed algorithms based on copula densities, rather than the classical ones based on probability densities, are more time consuming, since we estimate both copulas density of the vector and the marginal distribution function of each component. The present approach can be extended to deal with convolutive mixtures, and it will be interesting also to investigate a theoretical study of the identifiability assumptions as well as the convergence of the proposed algorithms. These developments are beyond the scope of the present paper, and will be addressed in future communications.
Acknowledgments Thanks to Egide PHC-Volubilis program for funding the project 27093ND - MA/12/270. The authors wish to thank the reviewers for their constructive comments and criticisms leading to improvement of this paper. References [1] S. Bouzebda, A. Keziou, New estimates and tests of independence in some copula models via ϕ-divergences, Kybernetika 46 (1) (2010) 178–201.
[2] M. Castella, S. Rhioui, E. Moreau, J.-C. Pesquet, Quadratic higher order criteria for iterative blind separation of a MIMO convolutive mixture of sources, IEEE Trans. Signal Process. 55 (1) (2007) 218–232. [3] R.B. Chen, M. Guo, W. Hardle, S. Huang, Independent component analysis via copula techniques, SFB 649 Discussion Paper 2008-004, 2008. [4] X. Chen, Y. Fan, Pseudo-likelihood ratio tests for semiparametric multivariate copula model selection, Can. J. Stat. 33 (3) (2005) 389–414. [5] A. Cichocki, S.-i. Amari, Families of alpha- beta- and gammadivergences: flexible and robust measures of similarities, Entropy 12 (6) (2010) 1532–1568. [6] P. Comon, Independent component analysis, a new concept? Signal Process. 36 (3) (1994) 287–314. [7] M. El Rhabi, G. Gelle, H. Fenniri, G. Delaunay, A penalized mutual information criterion for blind separation of convolutive mixtures, Signal Process. 84 (10) (2004) 1979–1984. [8] H. Joe, Multivariate Models and Dependence Concepts, Monographs on Statistics and Applied Probability, vol. 73, Chapman & Hall, London, 1997. [9] J. Ma, J. Sun, Copula component analysis, in: ICA 2007, 2007, pp. 73–80. [10] W. Naanaa, J.-M. Nuzillard, Blind source separation of positive and partially correlated data, Signal Process. 85 (9) (2005) 1711–1722. [11] R.B. Nelsen, An Introduction to Copulas, Springer Series in Statistics, second ed., Springer, New York, 2006. [12] M. Omelka, I. Gijbels, N. Veraverbeke, Improved kernel estimation of copulas: weak convergence and goodness-of-fit testing, Ann. Stat. 37 (5B) (2009) 3023–3058. [13] J.-C. Pesquet, E. Moreau, Cumulant-based independence measures for linear mixtures, IEEE Trans. Inf. Theory 47 (5) (2001) 1947–1956. [14] D.-T. Pham, Mutual information approach to blind separation of stationary sources, IEEE Trans. Inf. Theory 48 (7) (2002). [15] M.E. Rhabi, H. Fenniri, A. Keziou, E. Moreau, A robust algorithm for convolutive blind source separation in presence of noise, Signal Process. 93 (4) (2013) 818–827. [16] B.W. Silverman, Density Estimation for Statistics and Data Analysis, Monographs on Statistics and Applied Probability, Chapman & Hall, London, 1986. [17] A. Sklar, Fonctions de répartition à n dimensions et leurs marges, Publ. Inst. Stat. Univ. Paris 8 (1959) 229–231. [18] H. Tsukahara, Semiparametric estimation in copula models, Can. J. Stat. 33 (3) (2005) 357–375. [19] F. Vrins, D.-T. Pham, M. Verleysen, Is the general form of Renyi's entropy a contrast for source separation? in: Lecture Notes in Computer Science, vol. 4666/2007, Springer, Berlin, 2007.