(Luận án tiến sĩ) Một Số Hàm Khoảng Cách Trong Lý Thuyết Thông Tin Lượng Tử Và Các Vấn Đề Liên Quan

120 0 0
Tài liệu đã được kiểm tra trùng lặp
(Luận án tiến sĩ) Một Số Hàm Khoảng Cách Trong Lý Thuyết Thông Tin Lượng Tử Và Các Vấn Đề Liên Quan

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

Thông tin tài liệu

Trang 1

MINISTRY OF EDUCATION AND TRAINING QUY NHON UNIVERSITY

VUONG TRUNG DUNG

SOME DISTANCE FUNCTIONS IN QUANTUM INFORMATION THEORY AND RELATED PROBLEMS

DOCTORAL DISSERTATION IN MATHEMATICS

BINH DINH Ð 2024

Trang 2

MINISTRY OF EDUCATION AND TRAINING QUY NHON UNIVERSITY

VUONG TRUNG DUNG

SOME DISTANCE FUNCTIONS IN QUANTUM INFORMATION THEORY AND RELATED PROBLEMS

Speciality: Mathematical Analysis Speciality code: 9 46 01 02

Reviewer 1: Prof Dang Duc Trong Reviewer 2: Prof Pham Tien Son

Reviewer 3: Assoc Prof Pham Quy Muoi

1 Assoc Prof Dr Le Cong Trinh 2 Assoc Prof Dr Dinh Trung Hoa

BINH DINH Ð 2024

Trang 3

This thesis was completed at the Department of Mathematics and Statistics, Quy Nhon Uni-versity under the supervision of Assoc Prof Dr Le Cong Trinh and Assoc Prof Dr Dinh Trung Hoa I hereby declare that the results presented in it are new and original Most of them were published in peer-reviewed journals, others have not been published elsewhere For using results from joint papers I have gotten permission from my co-authors.

Binh Dinh, 2024

Vuong Trung Dung

Trang 4

This thesis was undertaken during my years as a PhD student at the Department of Math-ematics and Statistics, Quy Nhon University Upon the completion of this thesis, I am deeply indebted to numerous individuals On this occasion, I would like to extend my sincere appreci-ation to all of them.

First and foremost, I would like to express my sincerest gratitude to Assoc Prof Dr Dinh Trung Hoa, who guided me into the realm of matrix analysis and taught me right from the early days Not only that, but he also devoted a signiÞcant amount of valuable time to engage in dis-cussions, and provided problems for me to solve He motivated me to participate in workshops and establish connections with senior researchers in the Þeld He guided me to Þnd enjoyment in solving mathematical problems and consistently nurtured my enthusiasm for my work I canÕt envision having a more exceptional advisor and mentor than him.

The second person I would like to express my gratitude to is Assoc Prof Dr Le Cong Trinh, who has been teaching me since my undergraduate days and also introduced me to Prof Hoa From the early days of sitting in lecture halls at university, Prof Trinh has been instilling inspiration and a love for mathematics in me ItÕs fortunate that now I have the opportunity to be mentored by him once again He has always provided enthusiastic support not only in my work but also in life Without that dedicated support, it would have been difÞcult for me to complete this thesis.

I would like to extend a special thank you to the educators at both the Department of

Trang 5

Math-ematics and Statistic and the Department of Graduate Training at Quy Nhon University for providing the optimal environment for a postgraduate student who comes from a distant loca-tion like myself Binh Dinh is also my hometown and the place where I have spent all my time from high school to university The privilege and personal happiness of coming back to Quy Nhon University for advanced studies cannot be overstated.

I am grateful to the Board and Colleagues of VNU-HCM High School for the Gifted for providing me with much support to complete my PhD study Especially, I would like to extend my heartfelt gratitude to Dr Nguyen Thanh Hung, who has assisted to me in both material and spiritual aspects since the very Þrst days I set foot in Saigon He is not only a mentor and colleague but also a second father to me, who not only supported me Þnancially and emotionally during challenging times but also constantly encouraged me to pursue a doctoral degree Without this immense support and encouragement, I wouldnÕt be where I am today.

I also want to express my gratitude to Su for the wonderful time weÕve spent together, which has been a driving force for me to complete the PhD program and strive for even greater achieve-ments that I have yet to attain.

Lastly, and most signiÞcantly, I would like to express my gratitude to my family They have always been by my side throughout work, studies, and life I want to thank my parents for giving birth to me and nurturing me to adulthood This thesis is a gift I dedicate to them.

Binh Dinh, 2024 Vuong Trung Dung

Trang 6

1.1 Matrix theory fundamentals 13

1.2 Matrix function and matrix mean 19

2 Weighted Hellinger distance 28 2.1 Weighted Hellinger distance 29

2.2 In-betweenness property 32

3 The³-z-Bures Wasserstein divergence 38 3.1 The ³-z-Bures Wasserstein divergence and the least squares problem 42

3.2 Data processing inequality and in-betweenness property 60

3.3 Quantum Þdelity and its parameterized versions 64

3.4 The ³-z-Þdelity between unitary orbits 75

4 A new weighted spectral geometric mean 82 4.1 A new weighted spectral geometric mean and its basic properties 83

Trang 7

4.2 The Lie-Trotter formula and weak log-majorization 87

Trang 8

Glossary of notation

Cn : The set of alln-tuples of complex numbers 8x, y9 : The scalar product of vectorsx and y Mn : The set ofn × n complex matrices

B(H) : The set of all bounded linear operators acting on Hilbert spaceH Hn : The set of alln × n Hermitian matrices

n : The set of alln × n positive semi-deÞnite matrices Pn : The set of alln × n positive deÞnite matrices I, O : The identity and zero elements of Mn, respectively A7 : The conjugate transpose (or adjoint) of the matrixA |A| : The positive semi-deÞnite matrix(A7A)1/2

Tr(A) : The canonical trace of matrixA

»(A) : The vector of eigenvalues of matrixA in decreasing order s(A) : The vector of singular values of matrixA in decreasing order Sp(A) : The spectrum of matrixA

�胀A�胀 : The operator norm of matrixA

|||A||| : The unitarily invariant norm of matrixA x z y : x is majorized by y

x zw y : x is weakly majorized by y

A�胉B : The geometric mean of two matricesA and B

Trang 9

A�胉tB : The weighted geometric mean of two matricesA and B A�胉B : The spectral geometric mean of two matricesA and B

A�胉tB : The weighted spectral geometric mean of two matricesA and B Ft(A, B) : TheF -mean of two matrices A and B

A'B : The arithmetic mean of two matricesA and B A!B : The harmonic mean of two matricesA and B A : B : The parallel sum of two matricesA and B µp(A, B, t) : The matrixp-power mean of matrices A and B

Trang 10

Quantum information stands at the conßuence of quantum mechanics and information the-ory, wielding the mathematical elegance of both realms to delve into the profound nature of information processing at the quantum level In classical information theory, bits are the funda-mental units representing 0 and 1 Quantum information theory, however, introduces the concept of qubits, the quantum counterparts to classical bits Unlike classical bits, qubits can exist in a superposition of states, allowing them to be both 0 and 1 simultaneously This unique property empowers quantum computers to perform certain calculations exponentially faster than classical computers.

Entanglement is a crucial phenomenon in quantum theory where two or more particles be-come closely connected When particles are entangled, changing the state of one immediately affects the state of the other, no matter the distance between them This has important impli-cations for quantum information and computing, offering new possibilities for unique ways of handling information.

Quantum algorithms, such as ShorÕs algorithm for factoring large numbers and GroverÕs algorithm for quantum search, exemplify the power of quantum information in tackling complex computational tasks with unparalleled efÞciency.

In order to treat information processing in quantum systems, it is necessary to mathemati-cally formulate fundamental concepts such as quantum systems, states, and measurements, etc Useful tools for researching quantum information are functional analysis and matrix theory First, we consider the quantum system It is described by a Hilbert space H, which is called a

representation space This will be advantageous because it is not only the underlying basis of

Trang 11

quantum mechanics but is also as helpful in introducing the special notation used for quantum

mechanics The (pure) physical states of the system correspond to unit vectors of the Hilbert

space This correspondence is not 1-1 When f1 andf2 are unit vectors, then the correspond-ing states are identical if f1 = zf2 for a complex number z of modulus 1 Such z is often

called phase The pure physical state of the system determines a corresponding state vector upto a phase Traditional quantum mechanics distinguishes between pure states and mixed states.Mixed states are described by density matrices A density matrix or statistical operator is a

pos-itive matrix of trace 1 on the Hilbert space This means that the space has a basis consisting of eigenvectors of the statistical operator and the sum of eigenvalues is 1 In quantum information theory, distance functions are used to measure the distance between two mixed states Addition-ally, these distance functions can be employed to characterize the properties of a given quantum state For instance, they can quantify the quantum entanglement between two parts of a state, representing the shortest distance between the state and the set of all separable states These distance functions naturally extend to the set of positive semi-deÞnite matrices, which is also the main focus of this thesis.

Nowadays, the signiÞcance of matrix theory has been widely recognized across various Þelds, including engineering, probability and statistics, quantum information, numerical analy-sis, biological and social sciences In image processing (subdivision schemes), medical imaging (MRI), radar signal processing, statistical biology (DNA/genome), and machine learning, data from numerous experiments are stored as positive deÞnite matrices To work with each set of data, we need to select its representative element In other words, we need to compute the aver-age of the corresponding positive deÞnite matrices Therefore, considering global solutions of the least-squares problems for matrices is of paramount importance (refer to [2, 8, 18, 28, 67, 73] for examples).

Let0 < a f x f b Consider the following least squares problem:

d2(x, a) + d2(x, b) ³ min, x * [a, b],

Trang 12

whered := dE(x, y) = |y 2 x|, or, d := dR(x, y) := | log(y) 2 log(x)|.

The arithmetic mean(a + b)/2 and the geometric mean:

ab are unique solutions to the above problem with respect to dE and dR distance, respectively Moreover, based on the AM-GM inequality for two non-negative numbersa and b, we have a new distance as follows

d(a, b) = a + b

2 2:ab ForA, B * Pn, some matrix analogs of scalar distances are:

¥ Euclidean distance induced from Euclidean/Frobenius inner product 8A, B9 = Tr(A7B) The associated norm is�胀A�胀F = 8A, A91/2 = (Tr(A7A))1/2.

¥ The Riemann distance [12] is ·R(A, B) = || log(A21B)||2 =

¥ The Log-Determinant metric [75] in machine learning and quantum information:

dl(A, B) = log detA + B

Trang 13

between two data points Such functions are not necessarily symmetric; and the triangle inequal-ity does not need to be true Divergences [11] are such distance-like functions

DeÞnition A smooth function § : Pn× Pn ³ R+is called a quantum divergence if

(i) §(A, B) = 0 if and only if A = B.

(ii) The derivativeD§ with respect to the second variable vanishes on the diagonal, i.e.,

D§(A, X)|X=A = 0.

(iii) The second derivativeD2

§ is positive on the diagonal, i.e.,

D2§(A, X)|X=A(Y, Y ) g 0 for all Hermitian matrix Y.

Some divergences that have recently received a lot of attention are in [11, 14, 35, 56] Now let us revisit the scalar mean theory which serves as a starting point for our next problem

A two-variable function M (x, y) satisfying condition 6) can be reduced to a one-variable function f (x) := M (1, x) Namely, M (x, y) is recovered from f as M (x, y) = xf (x21y).

Trang 14

Notice that the functionf , corresponding to M is monotone increasing on R+ And this relation forms a one-to-one correspondence between means and monotone increasing functions on R+.

The following are some desired properties of any object that is called a ÒmeanÓM on H+n (A1) Positivity: A, B �背 0 ó M(A, B) �背 0.

(A2) Monotonicity:A �背 A2, B �背 B2 ó M(A, B) �背 M (A2, B2) (A3) Positive homogeneity:M (kA, kB) = kM (A, B) for k * R+.

(A4) Transformer inequality:X7M (A, B)X �胍 M (X7AX, X7BX) for X * B(H).

(A5) Congruence invariance:X7M (A, B)X = M (X7AX, X7BX) for invertible X * B(H) (A6) Concavity: M (tA + (1 2 t)B, tA2+ (1 2 t)B2) �背 tM (A, A2) + (1 2 t)M (B, B2) for

t * [0, 1].

(A7) Continuity from above: ifAn³ A and Bn ³ B, then M (An, Bn) ³ M(A, B) (A8) Betweenness: ifA �胍 B, then A �胍 M (A, B) �胍 B.

(A9) Fixed point property:M (A, A) = A.

To study matrix or operator means in general, we must Þrst consider three classical means in mathematics: arithmetic, geometric, and harmonic means These means are deÞned in the following manner, respectively,

A'B = 12(A + B),

A�胉B = A1/2�胀A21/2BA21/2�胀1/2A1/2, and

A!B = 2(A21+ B21)21.

In the above deÞnitions, if matrixA is not invertible, we replace A with A�胈 = A+�胈I and then let �胈 tend to 0 (similarly for matrixB) It can be seen that the arithmetic, harmonic and geometric

Trang 15

means share the properties (A1)-(A9) in common In 1980, Kubo and Ando [54] developed an axiomatic theory of operator mean on H+n At Þrst, they deÞned a connection of two matrices

as follows (the term ÒconnectionÓ comes from the study of electrical network connections) DeÞnition A connection on H+n is a binary operation à on H+n satisfying the following axioms for allA, A2, B, B2, C * H+

(M1) Monotonicity:A �胍 A2, B �胍 B2 =ó AÃB �胍 A2ÃB2 (M2) Transformer inequality:C(AÃB)C �胍 (CAC)Ã(CBC).

(M3) Joint continuity from above: if An, Bn * B(H)+satisfy An ³ A and Bn ³ B, then AnÃBn ³ AÃB.

A mean is a connection with normalization condition

(M4) IÃI = I.

To each connection à corresponds its transpose Ã2 deÞned byAÃ2B = BÃA A connection Ã

is symmetric by deÞnition if à = Ã2 The adjoint of Ã, denoted by Ã7, is deÞned byAÃ7B = (A21ÃB21)21, for invertible A, B When à is a non-zero connection, its dual, in symbol ç, is deÞned by ç = (Ã2)7 = (Ã7)2.

However, Kubo-Ando theory of means still has many limitations In applied and engineering Þelds, people need more classes of means that are non Kubo-Ando For some non Kubo-Ando means we refer the interested readers to [17, 23, 25, 35, 37].

One of the famous non-Kubo-Ando means is the spectral geometric mean [37], denoted as A�胉B, introduced in 1997 by Fiedler and Pt«ak It is called the spectral geometric mean because (A�胉B)2 is similar toAB and that the eigenvalues of their spectral mean are the positive square roots of the corresponding eigenvalues ofAB In 2015, Kim and Lee [52] deÞned the weighted spectral mean:

A�胉tB :=�胀A21�胉B�胀tA�胀A21�胉B�胀t, t * [0, 1] In this thesis we focus on two problems:

Trang 16

1 Distance function generated by operator means We introduce some new distance on the set of positive deÞnite matrices in the relation to operator means, and their applications In addition, we also study some geometric properties for means such as the in-betweenness property, and data processing inequality in quantum information.

2 A new weighted spectral geometric mean We introduce a new weighted spectral geo-metric mean, denoted byFt(A, B) and study basic properties for this quantity We also establish a weak log-majorization relation involvingFt(A, B) and the Lie-Trotter formula forFt(A, B).

The main tools in our research are the spectral theorem for Hermitian matrices and the theory of Kubo-Ando means Some fundamental techniques in the theory of operator monotone func-tions and operator convex funcfunc-tions are also utilized in the dissertation We also employ basic knowledge in matrix theory involving unitarily invariant norms, trace, etc.

The main results in this thesis are presented in the following articles:

1 Vuong T.D., Vo B.K (2020), ÒAn inequality for quantum ÞdelityÓ, Quy Nhon Univ J Sci.,

4 (3).

2 Dinh T.H., Le C.T., Vo B.K, Vuong T.D (2021), ÒWeighted Hellinger distance and in

betweenness propertyÓ, Math Ine Appls., 24, 157-165.

3 Dinh T.H., Le C.T., Vo B.K., Vuong T.D (2021), ÒThe ³-z-Bures Wasserstein

diver-genceÓ, Linear Algebra Appl., 624, 267-280.

4 Dinh T.H., Le C.T., Vuong T.D., ³-z-Þdelity and ³-z-weighted right mean, Submitted.

5 Dinh T.H., Tam T.Y., Vuong T.D, On new weighted spectral geometric mean, Submitted.

They were presented on the seminars at the Department of Mathematics and Statistics at Quy Nhon University and at the following international workshops and conferences as follows:

1 First SIBAU-NU Workshop on Matrix Analysis and Linear Algebra, 15-17 October, 2021.

Trang 17

2 20th Workshop on Optimization and ScientiÞc Computing, April 21-23, 2022 - Ba Vi,

5 International Workshop on Matrix Analysis and Its Applications, July 7-8, 2023, Quy Nhon, Viet Nam.

6 10th Viet Nam Mathematical Congress, August 8-12, 2023, Da Nang, Viet Nam.

This thesis has introduction, three chapters, conclusion, further investigation, a list of the authorÕs papers related to the thesis and preprints related to the topics of the thesis, and a list of references.

The introduction provides a background on the topics covered in this work and explains why they are meaningful and relevant It also brießy summarizes the content of the thesis and highlights the main results from the main three chapters.

In the Þrst chapter, the author collects some basic preliminaries which are used in this thesis In the second chapter, we introduce the weighted Hellinger distance for matrices which is an interpolating between the Euclidean distance and the Hellinger distance In 2019, Minh [43] introduced the Alpha Procrustes distance as follows: For ³ > 0, and for positive semi-deÞnite

In this chapter, by employing this approach, we deÞne a new distance called the Weighted Hellinger distance as follows:

dh,³(A, B) = 1 ³dh

A2³, B2³�胀

Trang 18

and then studied its properties In the Þrst section of this chapter, we show that the weighted Hellinger distance, as ³ tends to zero, is exactly the Log-Euclidean distance (Proposition 2.1.1), that is for two positive semi-deÞnite matricesA and B,

³³0d2h,³(A, B) = || log(A) 2 log(B)||2F.

Afterwards, in Proposition 2.1.2 we demonstrate the equivalence between the weighted Hellinger distance and the Alpha Procrustes distance, it means

db,³(A, B) f dh,³(A, B) f:2db,³(A, B).

We say that a matrix mean à satisÞes the in-betweenness property with respect to the metricd if for any pair of positive deÞnite operatorsA and B,

d(A, AÃB) f d(A, B).

In the second section, we prove that the matrix power meanµp(t, A, B) = (tAp+ (1 2 t)Bp)1/p

satisÞes the in-betweenness property in the weighted Hellinger and Alpha Procrustes distances (Theorem 2.2.1 and Theorem 2.2.2) At the end of this chapter, we prove that if à is a symmetric mean and satisÞes the in-betweenness property with respect to the Alpha Procrustes distance or the Weighted Hellinger distance, then it can only be the arithmetic mean (Theorem 2.2.3).

In chapter 3, we study a new quantum divergence so-called the ³-z-Bures Wasserstein di-vergence In 2015, Audenaert and Datta [7] introduced the R«enyi power mean of matrices via the matrix functionP³,z(A, B) =�胁B12³2z A³

Based on this quantity, in this chapter, the ³-z-Bures Wasserstein divergence for positive semi-deÞnite matrices A and B is deÞned by

§(A, B) = Tr((12 ³)A + ³B) 2 Tr (Q³,z(A, B)) , where Q³,z(A, B) = �胁A12³2z B³

Then we prove that this quantity is a quantum di-vergence (Theorem 3.1.1) We also solve the least square problem with respect to §(A, B) and

Trang 19

showed that the solution of this problem is exactly the unique positive deÞnite solution of the matrix equation

wiQ³,z(X, Ai) = X (Theorem 3.1.2) In [49], M Jeong and co-authors investigated this solution and denoted it byR³,z(Ë, A)-called ³-z-weighted right mean In this thesis, we continue our study of this quantity and obtain some new results An important result is an inequality for R³,z(Ë, A), which can be considered a version of the AM-GM inequality (Theorem 3.1.3) Hwang and Kim [48] proved that for any weightedm-mean Gmbetween arith-metic mean and geometric mean, the functionGË

Notice that the ³-z-weighted right mean does not satisfy the above condition However, we do have a similar result for RË

³,z := R³,z(Ë, á) (Theorem 3.1.4) The well-known Lie-Trotter formula [76] states that forX, Y * Mn,

This formula plays an essential role in the development of Lie Theory, and frequently appears in different research Þelds [44, 47, 48] In [48], J.Hwang and S.Kim introduced the multi-variate Lie-Trotter mean on the convex cone Pn of positive deÞnite matrices For a positive probability vector Ë = (w1, , wm) and differentiable curves ³1, , ³m on Pn with ³i(0) = I (i = 1, á á á , m), a weighted m-mean Gm (form g 2) is the multivariate Lie-Trotter mean if

In the end of this section, we prove thatR³,z(Ë, A) is a multivariate Lie-Trotter mean (Theorem 3.1.5) In the second section of this chapter, we show that this divergence satisÞes the data processing inequality (DPI) in quantum information (Theorem 3.2.1) The data processing in-equality is an information-theoretic concept that states that the information content of a signal

Trang 20

cannot be increased via a local physical operation This can be expressed concisely as Òpost-processing cannot increase informationÓ, that is, for any completely positive trace preserving mapE and any positive semi-deÞnite matrices A and B,

§(E(A), E(B))f §(A, B).

Furthermore, we show that the matrix power meanµ(t, A, B) = ((1 2 t)Ap+ tBp)1/p satisÞes the in-betweenness property with respect to the ³-z-Bures Wasserstein divergence (Theorem 3.2.2) Quantum Þdelity is an important quantity in quantum information theory and quantum chaos theory It is a distance measure between density matrices, which are considered as quan-tum states Although it is not a metric, it has many useful properties that can be used to deÞne a metric on the space of density matrices In the next section, we give some properties of quantum Þdelity and its extended version An important results is we establish some variational principles for the quantum ³-z-Þdelity

f³,z(Ã, Ã) := Tr�胀ó/2zÃ(12³)/zó/2z�胀z = Tr�胀Ã(12³)/2zó/zÃ(12³)/2z�胀z,

where à and à are two postitive deÞnite matrices (Theorem 3.3.4) That is, it is the extremal value of two matrix functions

Let U (H) be the set of n × n unitary matrices, and Dn the set of density matrices For Ã* Dn, its unitary orbit is deÞned as

UÃ= {U ÃU7 : U * U(H)}.

Trang 21

In the last section we are going to obtain the maximum and minimum distance between orbits of two state à and à inDnvia the quantum ³-z-Þdelity and prove that the set of these distance is a close interval in R+(Theorem 2.4.2 and Theorem 3.4.3)

In chapter 4, we introduce a new weighted spectral geometric mean

Ft(A, B) = (A21�胉tB)1/2A222t(A21�胉tB)1/2, t * [0, 1],

where A and B are positive deÞnite matrices We study basic properties and inequalities for Ft(A, B) An important property that we obtain in this chapter is that Ft(A, B) satisÞes the Lie-Trotter formula (Theorem 4.2.1).

At the end of this chapter, we compare the weak-log majorization between theF-mean and the Wasserstein mean, which is the solution to the least square problem with respect to the Bures distance or Wasserstein distance (Theorem 4.2.3).

Trang 22

Chapter 1

1.1Matrix theory fundamentals

Let N be the set of all natural numbers For each n * N, we denote by Mn the set of all n × n complex matrices, Hn is the set of alln × n Hermitian matrices, H+

n is the set ofn × n positive semi-deÞnite matrices, Pnis the cone of positive deÞnite matrices in Mn, andDnis the set of density matrices which are the positive deÞnite matrices with trace equal to one Denote byI and O the identity and zero elements of Mn, respectively This thesis deals with problems for matrices, which are operators in Þnite-dimensional Hilbert spacesH We will indicate if the case is inÞnite-dimensional.

Recall that for two vectors x = (xj) , y = (yj) * Cn the inner product 8x, y9 of x and y is deÞned as8x, y9 c �胅

xjøj Now let A be a matrix in Mn, the conjugate transpose or the adjointA7 ofA is the complex conjugate of the transpose AT We have,8Ax, y9 = 8x, A7y9 DeÞnition 1.1.1 A matrixA = (aij)ni,j=1 * Mnis said to be:

(i) diagonal ifaij = 0 when i 7= j.

(ii) invertible if there exists an matrixB of order n × n such that AB = In In this situation A has a unique inverse matrix A21 * Mnsuch thatA21A = AA21 = In.

Trang 23

(iii) normal ifAA7 = A7A (iv) unitary ifAA7 = A7A = In.

(v) Hermitian ifA = A7.

(vi) positive semi-deÞnite if8Ax, x9 g 0 for all x * Cn (vii) positive deÞnite if8Ax, x9 > 0 for all x * Cn\{0}.

DeÞnition 1.1.2 (L¬ownerÕs Order, [86]) LetA and B be two Hermitian matrices of same order n We say that A g B if and only if A 2 B is a positive semi-deÞnite matrix.

DeÞnition 1.1.3 A complex number » is said to be an eigenvalue of a matrixA corresponding to its non-zero eigenvectorx if

Ax = »x.

The multiset of the eigenvalues ofA is denoted by Sp(A) and called the spectrum of A.

There are several conditions that characterize positive matrices Some of them are listed in theorem below [10].

Proposition 1.1.1.

(i)A is positive semi-deÞnite if and only if it is Hermitian and all its eigenvalues are

nonneg-ative Moreover,A is positive deÞnite if and only if it is Hermitian and all its eigenvalues

are positive.

(ii)A is positive semi-deÞnite if and only if it is Hermitian and all its principal minors are

nonnegative Moreover,A is positive deÞnite if and only if it is Hermitian and all its

principal minors are positive.

(iii)A is positive semi-deÞnite if and only if A = B7B for some matrix B Moreover, A is

positive deÞnite if and only ifB is nonsingular.

Trang 24

(iv)A is positive semi-deÞnite if and only if A = T7T for some upper triangular matrix T

Further,T can be chosen to have nonnegative diagonal entries If A is positive deÞnite,

thenT is unique This is called the Cholesky decomposition of A Moreover, A is positive

deÞnite if and only ifT is nonsingular.

(v)A is positive semi-deÞnite if and only if A = B2 for some positive matrixB Such a B

is unique We writeB = A1/2 and call it the (positive) square root ofA Moreover, A is

positive deÞnite if and only ifB is positive deÞnite.

(vi)A is positive semi-deÞnite if and only if there exist x1, , xninH such that

aij = 8xi, xj9

A is positive deÞnite if and only if the vectors xj, 1 f j f n, are linearly independent.

Let A * Mn, we denote the eigenvalues of A by »j(A), for j = 1, 2, , n For a matrix A * Mn, the notation »(A) c (»1(A), »2(A), , »n(A)) means that »1(A) g »2(A) g g »n(A) The absolute value of matrix A * Mnis the square root of matrixA7A and denoted by

|A| = (A7A)1.

We call the eigenvalues of|A| by the singular value of A and denote as sj(A), for j = 1, 2, , n For a matrix A * Mn, the notation s(A) c (s1(A), s2(A), , sn(A)) means that s1(A) g s2(A) g g sn(A).

There are some basic properties of the spectrum of a matrix.

Proposition 1.1.2 LetA, B * Mn, then(i)Sp(AB) = Sp(BA).

(ii) IfA is a Hermitian matrix then Sp(A) ¢ R.

(iii) A is a positive semi-deÞnite (respectively positive deÞnite) if and only ifA is a Hermitian

matrix andSp(A) ¢ Rg0(respectively Sp(A) ¢ R+).

Trang 25

(iv) IfA, B g 0 then Sp(AB) ¢ R+.

The trace of a matrixA = (aij) * Mn, denoted byTr(A), is the sum of all diagonal entries, or, we often use the sum of all eigenvalues »i(A) of A, i.e.,

Related to the trace of the matrix, we recall the Araki-Lieb-Thirring trace inequality [18] used consistently throughout the thesis.

Theorem 1.1.1 LetA and B be two positive semi-deÞnite matrices, and let q > 0, we have

where Snis the set of all permutations à of the set S= {1, 2, , n}.

Proposition 1.1.3 LetA, B * Hnwith»(A) = (»1, »2, , »n) and »(B) = (µ1, µ2, , µn).

(i) IfA > 0 and B > 0, then A g B if and only if B21 g A21.(ii) IfA g B, then X7AX g X7BX for every X * Mn.(iii) IfA g B, then »j g µj for eachj = 1, 2, , n.

(iv) IfA g B g 0, then Tr(A) g Tr(B) g 0.

Trang 26

(v) IfA g B g 0, then det(A) g det(B) g 0.

A function�胀 á �胀 : Mn³ R is said to be a matrix norm if for all A, B * Mnand"³ * C we have:

(i) �胀A�胀 g 0.

(ii) �胀A�胀 = 0 if and only if A = 0 (iii) �胀³A|| = |³| á ||A||.

(iv) �胀A + B�胀 f �胀A�胀 + �胀B�胀.

In addition, a matrix norm is said to be sub-multiplicative matrix norm if

�胀AB�胀 f �胀A�胀 á �胀B�胀.

A matrix norm is said to be a unitarily invariant norm if for every A * Mn, we have �胀UAV �胀 = �胀A�胀 for all U, V * Ununitary matrices It is denoted as�胀| á |�胀.

These are some important norms over Mn The operator norm ofA, deÞned by

Trang 27

Whenp = 2, we have the Frobenius norm or sometimes called the Hilbert-Schmidt norm :

Letx = (x1, x2, , xn) and y = (y1, y2, , yn) be in Rn Let x³ = �胀x[1], x[2], , x[n]�胀 denote a rearrangement of the components ofx such that x[1] �背x[2] �背 �背 x[n] We say thatx

In other words,x zlog y if and only if log x z log y.

MatrixP * Mnis called a projection ifP2 = P One says that P is a Hermitian projection

if it is both Hermitian and a projection; P is an orthogonal projection if the range of P is

orthogonal to its null space The partial ordering is very simple for projections IfP and Q are projections, then the relationP f Q means that the range of P is included in the range of Q An equivalent algebraic formulation isP Q = P The largest projection in Mn is the identityI and the smallest one is 0 Therefore0 f P f I for any projection P * Mn Assume thatP andQ are projections on the same Hilbert space Among the projections which are smaller than P and Q there is a maximal projection, denoted by P ' Q, which is the orthogonal projection onto the intersection of the ranges ofP and Q.

Trang 28

Theorem 1.1.2 [45] Assume thatP and Q are orthogonal projections Then

P ' Q = limn³>(P QP )n= lim

n³>(QP Q)n.

1.2Matrix function and matrix mean

Now let us recall the spectral theorem which is one of the most important tools in functional analysis and matrix theory.

Theorem 1.2.1 (Spectral decomposition, [9]) Let »1 > »2 > »kbe eigenvalues of a Hermi-tian matrix A Then

For a real-valued functionf deÞned on some interval K ¢ R, and for a self-adjoint matrix A * Mn with spectrum in K, the matrix f (A) is deÞned by means of the functional calculus,

We are now at the stage where we will discuss matrix/operator functions L¬owner was the Þrst to study operator monotone functions in his seminal papers [63] in 1930 In the same time, Kraus investigated the notion operator convex function [55].

DeÞnition 1.2.1 ([63]) A continuous functionf deÞned on an interval K(K ¢ R) is said to be

operator monotone of ordern on K if for two Hermitian matrices A and B in Mnwith spectras

Trang 29

inK, one has

A f B implies f (A) f f(B).

Iff is operator monotone of any orders then f is called operator monotone.

Theorem 1.2.2 (L¬owner-HeinzÕs Inequality, [86]) The functionf (t) = tris operator monotoneon[0, >) for 0 f r f 1 More speciÞcally, for two positive semi-deÞnite matrices such thatA f B Then

Ar f Br, 0 f r f 1.

DeÞnition 1.2.2 ([55]) A continuous functionf deÞned on an interval K(K ¢ R) is said to be

operator convex of ordern on K if for any Hermitian matrices A and B in Mnwith spectra in K, and for all real numbers 0 f » f 1,

f (»A + (1 2 »)B) f »f(A) + (1 2 »)f(B).

Iff is operator convex of any order n then f is called operator convex If 2f is operator convex

then we callf is operator concave.

Theorem 1.2.3 ([10]) Functionf (t) = trin[0, >) is operator convex when r * [21, 0]*[1, 2].

More speciÞcally, for any positive semi-deÞnite matricesA, B and for any » * [0, 1],

(»A + (1 2 »)B)r f »Ar+ (1 2 »)Br.

Another important example is the function f (t) = log t, which is operator monotone on (0, >) and the function g(t) = t log t is operator convex The relations between operator mono-tone and operator convex via the theorem below.

Theorem 1.2.4 ([9]) Letf be a (continuous) real function on the interval [0, ³) Then the

following two conditions are equivalent:(i)f is operator convex and f (0) f 0.

Trang 30

(ii) The functiong(t) = f (t)

t is operator monotone on(0, ³).

DeÞnition 1.2.3 ([10]) Letf (A, B) be a real valued function of two matrix variables Then, f

is called jointly concave, if for all0 f ³ f 1,

f (³A1 + (1 2 ³)A2, ³B1+ (1 2 ³)B2) g ³f(A1, B1) + (1 2 ³)f(A2, B2)

for allA1, A2, B1, B2 If 2f is jointly concave, we say f is jointly convex.

We will review very quickly some basic concepts of the Fr«echet differential calculus, with special emphasis on matrix analysis LetX, Y be real Banach spaces, and let L(X, Y ) be the space of bounded linear operators fromX to Y Let U be an open subset of X A continuous mapf from U to Y is said to be differentiable at a point u of U if there exists T * L(X, Y )

It is clear that if such aT exists, it is unique If f is differentiable at u, the operator T above is called the derivative of f at u We will use for it the notation Df (u), of "f (u) This is sometimes called the Fr«echet derivative Iff is differentiable at every point of U , we say that it is differentiable onU One can see that, if f is differentiable at u, then for every v * X,

This is also called the directional derivative off at u in the direction v Iff1, f2are two differentiable maps, thenf1 + f2is differentiable and

D (f1+ f2) (u) = Df1(u) + Df2(u).

The composite of two differentiable mapsf and g is differentiable and we have the chain rule

D(g ç f)(u) = Dg(f(u)) á Df(u).

Trang 31

One important rule of differentiation for real functions is the product rule: (f g)2 = f2g + gf2 If f and g are two maps with values in a Banach space, their product is not deÞned - unless the range is an algebra as well Still, a general product rule can be established Letf, g be two differentiable maps fromX into Y1, Y2, respectively LetB be a continuous bilinear map from Y1× Y2 intoZ Let × be the map from X to Z deÞned as ×(x) = B(f (x), g(x)) Then for all u, v in X

D×(u)(v) = B(Df (u)(v), g(u)) + B(f (u), Dg(u)(v)).

This is the product rule for differentiation A special case of this arises whenY1 = Y2 = L(Y ), the algebra of bounded operators in a Banach space Y Now ×(x) = f (x)g(x) is the usual product of two operators The product rule then is

D×(u)(v) = [Df (u)(v)] á g(u) + f (u) á [Dg(u)(v)]

Higher order Fr«echet derivatives can be identiÞed with multilinear maps Letf be a differen-tiable map from X to Y At each point u, the derivative Df (u) is an element of the Banach spaceL(X, Y ) Thus we have a map Df from X into L(X, Y ), deÞned as Df : u ³ Df(u) If this map is differentiable at a pointu, we say that f is twice differentiable at u The derivative of the mapDf at the point u is called the second derivative of f at u It is denoted as D2f (u) This is an element of the spaceL(X, L(X, Y )) Let L2(X, Y ) be the space of bounded bilinear maps fromX × X into Y The elements of this space are maps f from X × X into Y that are linear in both variables, and for whom there exists a constantc such that

�胀f (x1, x2)�胀 f c �胀x1�胀 �胀x2�胀

for all x1, x2 * X The inÞmum of all such c is called �胀f�胀 This is a norm on the space L2(X, Y ), and the space is a Banach space with this norm If × is an element of L(X, L(X, Y )), let

×(x1, x2) = [× (x1)] (x2) for x1, x2 * X.

Trang 32

Then×÷ * L2(X, Y ) It is easy to see that the map × ³ ÷× is an isometric isomorphism Thus the second derivative of a twice differentiable mapf from X to Y can be thought of as a bilinear map fromX × X to Y It is easy to see that this map is symmetric in the two variables; i.e.,

D2f (u) (v1, v2) = D2f (u) (v2, v1)

for allu, v1, v2 Derivatives of higher order can be deÞned by repeating the above procedure Thep th derivative of a map f from X to Y can be identiÞed with a p-linear map from the space X × X × á á á × X ( p copies) into Y A convenient method of calculating the p th derivative of f is provided by the formula

For the convenience of readers, let us provide some examples for the derivatives of matrices Example 1.2.1 In these examplesX = Y = L(H).

(i) Letf (A) = A2 Then

Trang 33

(iii) Letf (A) = A22 for each invertibleA Then

In connections with electrical engineering, Anderson and DufÞn [3] deÞned the parallel sum

of two positive deÞnite matricesA and B by

Trang 34

investi-gated the structure of the Riemannian manifold H+n They showed that the curve

³(t) = A�胉tB = A1/2�胀A21/2BA21/2�胀tA1/2 (t * [0, 1])

is the unique geodesic joining A and B, and called t-geometric mean or weighted geometric mean The weighted harmonic and the weighted arithmetic means are deÞned by

A!tB =�胀tA21+ (1 2 t)B21�胀21,

A'tB = tA + (1 2 t)B.

The well-known inequality related to these quantities is the harmonic, geometric, and arithmetic means inequality [47, 60] , that is,

A!tB f A�胉tB f A'tB.

These three means are Kubo-Ando means LetÕs collect the main content of the Kubo-Ando means theory in the general case [54] For x > 0 and t g 0, the function Ç(x, t) = x(1 + t)x + t is bounded and continuous on the extended half-line [0, >] The L¬owner theory ([9, 45]) on operator-monotone functions states that the mapm �胀³ f, deÞned by

f (x) = �胅

Ç(x, t)dm(t) for x > 0,

establishes an afÞne isomorphism from the class of positive Radon measures on[0, >] onto the class of operator-monotone functions In the representation abvove,f (0) = inf

Trang 35

monotone functionf : R+³ R+, satisfying

We callf the representing function of Ã.

The next theorem follows from the integral representation of matrix monotone functions and from the previous theorem.

Theorem 1.2.6 The map,m �胀³ Ã, deÞned by

IfP and Q are two projections, then the explicit formulation for P ÃQ is simpler.

Theorem 1.2.7 Ifà is a mean, then for every pair of projectionsP and Q

P ÃQ = a(P 2 P ' Q) + b(Q 2 P ' Q) + P ' Q,

Trang 36

Let f be the representing function of à Since xf (x21) is the representing function of the transpose Ã2, then à is symmetric if and only iff (x) = xf (x21) The next theorem gives the representation for a symmetric connection.

Theorem 1.2.8 The map,n �胀³ Ã, deÞned by

wherec = n({0}), establishes an afÞne isomorphism from the class of positive Radon measures

on the unit interval[0, 1] onto the class of symmetric connections.

Trang 37

Chapter 2

Weighted Hellinger distance

In recent years, many researchers have paid attention to different distance functions on the set Pnof positive deÞnite matrices Along with the traditional Riemannian metricdR(A, B) =

(where »i(A21B) are eigenvalues of the matrix A21/2BA21/2), there are other important functions Two of them are the Bures-Wasserstein distance [13], which are adapted from the theory of optimal transport :

db(A, B) =�胀Tr(A + B) 2 2 Tr((A1/2BA1/2)1/2)�胀1/2,

and the Hellinger metric or Bhattacharya metric [11] in quantum information :

dh(A, B) =�胀Tr(A + B) 2 2 Tr(A1/2B1/2)�胀1/2.

Notice that the metric dh is the same as the Euclidean distance between A1/2 and B1/2, i.e., �胀A1/22 B1/2�胀F.

Recently, Minh [43] introduced the Alpha Procrustes distance as follows: For ³> 0 and for two positive semi-deÞnite matricesA and B,

db,³ = 1

³db(A2³, B2³).

Trang 38

He showed that the Alpha Procrustes distances are the Riemannian distances corresponding to a family of Riemannian metrics on the manifold of positive deÞnite matrices, which encom-pass both the Log-Euclidean and Wasserstein Riemannian metrics Since the Alpha Procrustes

distances are deÞned based on the Bures-Wasserstein distance, we also call them the weighted

Bures-Wasserstein distances In that ßow, in this chapter we can deÞne the weighted Hellinger

metricfor two positive semi-deÞnite matrices as follows: dh,³(A, B) = 1

³dh(A2³, B2³), then investigate its properties within this framework.

The results of this chapter are taken from [32].

2.1Weighted Hellinger distance

DeÞnition 2.1.1 For two positive semi-deÞnite matricesA and B and for ³ > 0, the weighted Hellinger distance betweenA and B is deÞned as

dh,³(A, B) = 1

³dh(A2³, B2³) = 1

³(Tr(A2³+ B2³) 2 2 Tr(A³B³))1 (2.1.1) It turns out that dh,³(A, B) is an interpolating metric between the Log-Euclidean and the Hellinger metrics We start by showing that the limit of the weighted Hellinger distance as ³ tends to 0 is the Log-Euclidean distance We also show that the weighted Bures-Wasserstein and weighted Hellinger distances are equivalent (Proposition 2.1.2).

Proposition 2.1.1 For two positive semi-deÞnite matricesA and B,

³³0d2h,³(A, B) = || log(A) 2 log(B)||2F.

Proof. We rewrite the expression ofdh,³(A, B) as

Trang 39

Tending ³ to zero, we obtain

d2h,³(A, B) = || log A||2F + || log B||2B2 2�胄log A, log B�胄

F = || log A 2 log B||2F.

Trang 40

This completes the proof.

It is interesting to note that the weighted Bures-Wasserstein and weighted Hellinger distances are equivalent.

Proposition 2.1.2 For two positive semi-deÞnite matricesA and B,

db,³(A, B) f dh,³(A, B) f:2db,³(A, B).

Proof. According the Araki-Lieb-Thirring inequality [43] , we have Tr(A1/2BA1/2)r g Tr(ArBr), |r| f 1.

ReplaceA with A2³,B with B2³andr with 1

2 we obtain the following In the above inequality replace à with A2³

Tr(A2³) and à with

Tr(B2³) we have

Ngày đăng: 06/05/2024, 07:57

Tài liệu cùng người dùng

Tài liệu liên quan