Next Article in Journal
A Generic Formula and Some Special Cases for the Kullback–Leibler Divergence between Central Multivariate Cauchy Distributions
Next Article in Special Issue
Multi-User Measurement-Device-Independent Quantum Key Distribution Based on GHZ Entangled State
Previous Article in Journal
Nonlinear Modeling Study of Aerodynamic Characteristics of an X38-like Vehicle at Strong Viscous Interaction Regions
Previous Article in Special Issue
A Fisher Information-Based Incompatibility Criterion for Quantum Channels
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Belavkin–Staszewski Relative Entropy, Conditional Entropy, and Mutual Information

The College of Computer Science, Shaanxi Normal University, **’an 710062, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(6), 837; https://doi.org/10.3390/e24060837
Submission received: 23 May 2022 / Revised: 14 June 2022 / Accepted: 16 June 2022 / Published: 17 June 2022
(This article belongs to the Special Issue Quantum Information and Computation)

Abstract

:
Belavkin–Staszewski relative entropy can naturally characterize the effects of the possible noncommutativity of quantum states. In this paper, two new conditional entropy terms and four new mutual information terms are first defined by replacing quantum relative entropy with Belavkin–Staszewski relative entropy. Next, their basic properties are investigated, especially in classical-quantum settings. In particular, we show the weak concavity of the Belavkin–Staszewski conditional entropy and obtain the chain rule for the Belavkin–Staszewski mutual information. Finally, the subadditivity of the Belavkin–Staszewski relative entropy is established, i.e., the Belavkin–Staszewski relative entropy of a joint system is less than the sum of that of its corresponding subsystems with the help of some multiplicative and additive factors. Meanwhile, we also provide a certain subadditivity of the geometric Rényi relative entropy.

1. Introduction

Rényi proposed an axiomatic approach to derive the Shannon entropy, and he found a family of entropies with parameter α   ( α [ 0 , 1 ) ( 1 , ) ) , called Rényi entropy. Meanwhile, the same axiomatic approach was extended to relative entropy and obtained Rényi relative entropy [1]. Relative entropy (or Kullback–Leibler divergence [2]) is a special case of Rényi relative entropy, which is an important ingredient for a mathematical framework of information theory. It has operational meaning in information theoretical tasks and can be used to describe the level of closeness between two random variables [3,4]. The axiomatic approach introduced by Rényi can be readily generalized to quantum settings [5,6]. Because of the non-commutativity of the quantum states, there are at least three different and special ways to generalize the classical Rényi relative entropy [6,7,8,9,10], such as Petz-Rényi relative entropy [11,12], sandwiched Rényi relative entropy [6,13] and geometric Rényi relative entropy [14,15]. These quantities are very meaningful in different information-theoretic tasks, including source coding, hypothesis testing, state merging, and channel coding.
The fact is that quantum relative entropy, by taking the limit as α 1 , is a special case of the Petz-Rényi and sandwiched Rényi relative entropies. However, the geometric Rényi relative entropy converges to the Belavkin–Staszewski (BS) relative entropy by taking the same limit. It is noteworthy that both the quantum and BS relative entropies are important variants of the classical relative entropy extension to quantum settings [16,17,18]. Quantum relative entropy, a direct generalization of the classical relative entropy, has been studied extensively in recent decades. BS relative entropy is also an enticing and crucial entropy used to process quantum information tasks, which can be used to describe the effects of possible noncommutativity of the quantum states (the quantum relative entropy can not work well for this). Additionally, BS relative entropy has recently attracted the attention of researchers. More precisely, Katariya and Wilde employed BS relative entropy to discuss quantum channel estimation and discrimination [19], Bluhm and Capel contributed a strengthened data processing inequality for BS relative entropy [20]. This property was first established by Hiai and Petz [21]. Bluhm et al. produced some weak quasi-factorization results for BS relative entropy [22]. Fang and Fawzi studied quantum channel capacities with respect to geometric Rényi relative entropy [23].
It is commonly known that von Neumann entropy, quantum conditional entropy, and quantum mutual information play vital roles in quantum information theory. Apart from the above entropic measures derived from the quantum relative entropy, however, other useful entropy-like quantities have also been well studied recently, such as max-information [24], collision entropy [25], and min- and max-entropies [26,27,28]. All of these information measures were generated from quantum Rényi relative entropies by taking different limits.
BS relative entropy can be seen as a fresh and forceful tool to resolve some specific challenges of quantum information-processing tasks. Concurrently, the main use of the geometric Rényi and BS relative entropies is to establish upper bounds on the rates of feedback-assisted quantum communication protocols [29]. To our present knowledge, there is no systematic analysis and research for conditional entropy and mutual information defined from BS relative entropy. Therefore, this paper explores some basic but necessary results for BS relative entropy. More precisely, we first provide a class of new definitions of conditional entropy (called BS conditional entropy, see Definition 2) and new mutual information (called BS mutual information, see Definition 3) via BS relative entropy. Additionally, we showed that von Neumann entropy can be defined by BS relative entropy. Second, we built an order relation between the BS conditional entropy of the bipartite and tripartite quantum systems. Subsequently, since classical-quantum states play an essential role in quantum channel coding and classical data compression with quantum side information, we discussed some valuable properties of BS conditional entropy and BS mutual information in classical-quantum settings. We established the weak concavity of BS conditional entropy and obtained chain rules for BS mutual information. Last but not least, the subadditivity of the geometric Rényi and BS relative entropies is established with the help of some multiplicative and additive factors (the factors are different linear combinations of quantum max-relative entropy [30]), i.e., the geomertric Rényi/BS relative entropy of a joint system is less than the sum of that of its corresponding subsystems.
This paper is organized as follows. In Section 2, we present the mathematical terminology and formal definitions necessary for the formulation of our results. Our results are shown in Section 3. The paper ends with a conclusion.

2. Basic Notations and Definitions

We denote a finite-dimensional Hilbert space by H . Normalized quantum states are in the set S = ( H ) : = { ρ P ( H ) : Tr ρ = 1 } , and subnormalized states are in the set S ( H ) : = { ρ P ( H ) : 0 < Tr ρ 1 } . We use P + ( H ) and P ( H ) to denote the set of positive definite operators and the set of positive semi-definite operators on H , respectively. An identity operator is denoted by I. The Hilbert spaces corresponding to different physical systems are distinguished by different capital Latin letters as a subscript. A compound system is modeled using the Hilbert space H A B = H A H B . For a bipartite classical-quantum system H X B , the corresponding state ρ X B is formalized as
ρ X B = x p ( x ) | x x | X ρ B x ,
where { | x } corresponds to an orthonormal basis on the classical system H X , ρ B x is any quantum state on the quantum system H B , p ( x ) is the probability distribution, and x p ( x ) = 1 [17,29]. We also refer to a tripartite classical-quantum state,
ρ X A B = x p ( x ) | x x | X ρ A B x ,
where ρ A B x is any quantum state on the quantum system H A B .
In quantum information theory, one can generalize Rényi relative entropy to the quantum case, these quantities depend on a parameter α ( 0 , 1 ) ( 1 , ) , and one can evaluate their values at α { 0 , 1 , } by taking different limits. For Petz-Rényi relative entropy [11,12] and sandwiched Rényi relative entropy [6,13], if one takes the limit as α 1 , we can obtain the well-known quantum relative entropy. For ρ S = ( H ) and σ S ( H ) , if supp ( ρ ) supp ( σ ) , the quantum relative entropy of ρ and σ is defined as
D ( ρ σ ) = Tr [ ρ ( log ρ log σ ) ] ;
otherwise, it is defined as + . Throughout this paper, we take the logarithmic function to base 2. The quantum relative entropy is nonnegative and satisfies the data processing inequality, which has good applications in quantum hypothesis testing and quantum resource theory [5,31,32]. We now define the geometric Rényi relative entropy [5,14,15,29].
Definition 1.
For all α ( 0 , 1 ) ( 1 , ) , ρ S = ( H ) and σ S ( H ) , the geometric Rényi relative entropy is defined as
D ^ α ( ρ σ ) = 1 α 1 log Tr σ ( σ 1 2 ρ σ 1 2 ) α .
The term of σ 1 2 ( σ 1 2 ρ σ 1 2 ) α σ 1 2 is called the weighted matrix geometric mean of two positive definite operators ρ and σ , where α is the weight parameter [5,23,33]. The geometric Rényi relative entropy can be shown to be the maximal relative entropy among all quantum Rényi relative entropies satisfying the data processing inequality [5,15], so it is also called maximal quantum Rényi relative entropy [5]. The geometric Rényi relative entropy increases monotonically with respect to the parameter α . Specially, for the limit as α 1 , the geometric Rényi relative entropy converges to the BS relative entropy [15,19,29], i.e.,
D ^ ( ρ σ ) = lim α 1 D ^ α ( ρ σ ) = Tr ρ log ( ρ 1 2 σ 1 ρ 1 2 ) ,
where supp ( ρ ) supp ( σ ) ; otherwise, D ^ ( ρ σ ) = + . Here, the inverse σ 1 is taken on the support of σ .
Similar to the quantum relative entropy [17], the BS relative entropy is non-negative and satisfies the data processing inequality [15,19,29]. For every quantum channel E , we have
D ^ ( E ( ρ ) E ( σ ) ) D ^ ( ρ σ ) .
For more properties of the BS and geometric Rényi relative entropies, one can refer to [19,29]. In particular, the quantum relative entropy is never larger than the BS relative entropy [19,21], i.e.,
D ( ρ σ ) D ^ ( ρ σ ) .
Obviously, if ρ and σ can be commuted, the BS relative entropy will reduce to the quantum relative entropy. In this paper, we also need to employ the quantum max-relative entropy [5,6,30,34], which comes from the sandwiched Rényi relative entropy by taking the limit as α , and
D max ( ρ σ ) = log inf { λ : ρ λ σ } .

3. Main Results

Once again, the Petz-/sandwiched and geometric Rényi relative entropies are inconsistent when taking the limit as α 1 , which leads to many differences between the BS relative entropy D ^ ( ρ σ ) (generated from the geometric Rényi relative entropy) and the quantum relative entropy D ( ρ σ ) (generated from the Petz-/sandwiched Rényi relative entropy). However, we find that both the quantum relative entropy and the BS relative entropy reduce to the von Neumann entropy for any quantum state ρ when one takes σ = I , i.e.,
D ^ ( ρ I ) =   Tr ρ log ( ρ 1 2 I 1 ρ 1 2 ) =   Tr ρ log ρ =   S ( ρ ) .
Thus, we have
D ^ ( ρ I ) = D ( ρ I ) = S ( ρ ) .

3.1. Belavkin–Staszewski Conditional Entropy

One of the significant properties of the relative entropy is that it can derive the conditional entropy and mutual information in information theory. The quantum relative entropy is the quantum analogue of Kullback–Leibler divergence. We know that there is no similar concept for the joint probability distribution of two variables with different time in quantum mechanics; in other words, there is no real conditional quantum state to process quantum information tasks. Thus, we can consider a formal definition of the quantum conditional entropy [17,31], i.e.,
S ( A | B ) ρ = S ( ρ A B ) S ( ρ B ) ,
where ρ B = Tr A ( ρ A B ) is the reduced state for the bipartite quantum state ρ A B . The quantum conditional entropy S ( A | B ) ρ can be denoted as the quantum relative entropy [5,29], i.e.,
S ( A | B ) ρ = D ( ρ A B I A ρ B ) .
In fact, from the basic properties of the quantum relative entropy, the above equation has another equivalent expression [5,29],
S ( A | B ) ρ = max σ B D ( ρ A B I A σ B ) ,
where the maximum is taken over all sub-normalized states on H B .
Combining Equation (9) with Equation (10), we have
S ( A | B ) ρ = D ^ ( ρ A B I A B ) + D ^ ( ρ B I B ) .
However, from the property of Equation (7) and definition of Equation (11), intuitively, we find that conditional entropy defined by the BS relative entropy is different from the quantum conditional entropy of Equation (11) generally. Therefore, we define a new conditional entropy based on the BS relative entropy in the following: the so-called BS conditional entropy.
Definition 2.
For any quantum state ρ A B S = ( H A B ) , the BS conditional entropy is defined as
S ^ ( A | B ) ρ = D ^ ( ρ A B I A ρ B ) .
Similar to Equation (12), we can also define the alternative BS conditional entropy, i.e.,
S ^ m ( A | B ) ρ = max σ B D ^ ( ρ A B I A σ B ) ,
where σ B S ( H B ) . In general, the optimal state is not necessarily the state ρ B . We further have
S ^ m ( A | B ) ρ S ^ ( A | B ) ρ S ( A | B ) ρ ,
from the relation of Equation (7). Additionally, if one considers the above relations in the bipartite classical-quantum systems, they remain equal, as follows.
Lemma 1.
For any bipartite classical-quantum states ρ X B , we have
S ^ m ( B | X ) ρ = S ^ ( B | X ) ρ = S ( B | X ) ρ .
Proof. 
Without a loss of generality, letting σ X = x q ( x ) | x x | and ρ X = x p ( x ) | x x | , using the definition of Equation (1), we have
Tr ρ X B log ρ X B 1 2 I B σ X 1 ρ X B 1 2 = Tr x p ( x ) | x x | ρ B x · log x p ( x ) q ( x ) 1 | x x | ρ B x = Tr x p ( x ) log p ( x ) q ( x ) 1 | x x | ρ B x + x p ( x ) Tr ρ B x log ρ B x = D ( p ( x ) q ( x ) ) x p ( x ) S ( ρ B x ) ,
where the last equality follows from the fact that
Tr x p ( x ) log p ( x ) q ( x ) 1 | x x | ρ B x = Tr x p ( x ) log p ( x ) q ( x ) 1 | x x | · Tr ρ B x
and Tr ρ B x = 1 .
Next, taking the minimization of σ X for both sides of Equation (18), we have
min σ B Tr ρ X B log ρ X B 1 2 I B σ X 1 ρ X B 1 2 = min σ B D ( p ( x ) q ( x ) ) x p ( x ) S ( ρ B x ) = x p ( x ) S ( ρ B x ) .
The optimization of σ X for the first equality only depends on the first term. For all x, one takes the minimization if and only if p ( x ) = q ( x ) , which also implies that D ( p ( x ) q ( x ) ) = 0 . Furthermore, combining the definitions of Equation (15) with Equation (5) to Equation (19), it holds that
S ^ m ( B | X ) ρ = x p ( x ) S ( ρ B x ) .
For S ^ ( B | X ) ρ , we have
Tr ρ X B log ρ X B 1 2 I B ρ X 1 ρ X B 1 2 = Tr x p ( x ) | x x | ρ B x · log x p ( x ) p ( x ) 1 | x x | ρ B x = x p ( x ) S ( ρ B x ) .
Similarly, we have
S ^ ( B | X ) ρ = x p ( x ) S ( ρ B x ) .
Finally, we can obtain the same result for S ( B | X ) ρ . We thus complete this proof.   □
Since the BS relative entropy satisfies the data processing inequality (6), for any tripartite quantum systems H A B C , it is easy to obtain that conditioning reduces entropy, i.e.,
S ^ ( A | B C ) ρ S ^ ( A | B ) ρ .
This property also holds for the quantum conditional entropy S ( A | B ) ρ . However, there is not always true for S ( A | B ) ρ S ( A C | B ) ρ (Problem 11.3(2), in [17]). For the BS conditional entropy, this paper provides another result in the following.
Lemma 2.
For any tripartite quantum states ρ A B C S = ( H A B C ) , we have
S ^ ( A B | C ) ρ log d A S ^ ( B | C ) ρ
and
S ^ m ( A B | C ) ρ log d A S ^ m ( B | C ) ρ ,
where d A is the dimension of subsystem H A .
Proof. 
Since the BS relative entropy satisfies the data-processing inequality (Corollary 4.53 in [29]) and the fact that partial trace is a quantum channel [5], we have
D ^ ( ρ A B C I A B ρ C ) D ^ ( Tr A ( ρ A B C ) Tr A ( I A B ρ C ) ) .
Furthermore, we have
D ^ ( Tr A ( ρ A B C ) Tr A ( I A B ρ C ) ) = D ^ ( ρ B C d A I B ρ C ) ,
where d A is the dimension of subsystem H A . Applying the additivity of the BS relative entropy (Proposition 4.54, [29]), we then have
D ^ ( ρ B C d A I B ρ C ) = D ^ ( ρ B C I B ρ C ) log d A .
Recalling the definition of the BS conditional entropy S ^ ( A | B ) ρ to the above equality, we have
S ^ ( A B | C ) ρ log d A S ^ ( B | C ) ρ .
Similarly, for the alternative definition of the BS conditional entropy S ^ m ( A | B ) ρ , taking the minimization of σ X , we have
min σ B D ^ ( ρ A B C I A B σ C ) min σ B D ^ ( ρ B C I B σ C ) log d A ,
which implies that
S ^ m ( A B | C ) ρ log d A S ^ m ( B | C ) ρ .
  □
The quantum conditional entropy of Equation (10) also satisfies the concavity, which plays an important role in the quantum information processing [5,17,31]. Additionally, for the BS conditional entropy of the tripartite classical-quantum state, we obtain the following result.
Theorem 1.
For any tripartite classical-quantum states ρ X A B , we have
H ( X ) + x p ( x ) S ^ ( A x | B ) ρ S ^ ( A | B ) ρ + log d X
and
H ( X ) + x p ( x ) S ^ m ( A x | B ) ρ S ^ m ( A | B ) ρ + log d X ,
where H ( X ) is the Shannon entropy, and S ^ ( A x | B ) ρ is the BS conditional entropy for the quantum state ρ A B x .
Proof. 
For any tripartite classical-quantum states ρ X A B , we have
Tr ρ X A B log ρ X A B 1 2 I X A ρ B 1 ρ X A B 1 2 = Tr x p ( x ) | x x | ρ A B x · log x p ( x ) | x x | ρ A B x 1 2 I A ρ B 1 ρ A B x 1 2 = Tr x p ( x ) | x x | ρ A B x · log x p ( x ) | x x | + Tr x p ( x ) | x x | ρ A B x log ρ A B x 1 2 I A ρ B 1 ρ A B x 1 2 .
For the first term of the last equality, we have
Tr x p ( x ) | x x | ρ A B x · log x p ( x ) | x x | = x p ( x ) log p ( x ) · Tr ρ A B x = H ( X ) ,
where the last equality follows from the fact that Tr ρ A B x = 1 . Similarly, for the second term, we further have
Tr x p ( x ) | x x | ρ A B x log ρ A B x 1 2 I A ρ B 1 ρ A B x 1 2 = x p ( x ) Tr ρ A B x log ρ A B x 1 2 I A ρ B 1 ρ A B x 1 2 = x p ( x ) D ^ ( ρ A B x I A ρ B ) = x p ( x ) S ^ ( A x | B ) ρ ,
where S ^ ( A x | B ) ρ is the BS conditional entropy for the quantum state ρ A B x .
Therefore, we can obtain
Tr ρ X A B log ρ X A B 1 2 I X A ρ B 1 ρ X A B 1 2 = H ( X ) x p ( x ) S ^ ( A x | B ) ρ .
Using the definition of the BS conditional entropy S ^ ( X A | B ) ρ to the above equality, we have
S ^ ( X A | B ) ρ = H ( X ) + x p ( x ) S ^ ( A x | B ) ρ .
Applying Lemma 2, we further have
S ^ ( X A | B ) ρ S ^ ( A | B ) ρ + log d X .
Substituting Equation (27) into the above inequality (28), we can obtain the first inequality of Theorem 1.
We will replace ρ B with σ B to analyze the case of the alternative BS relative entropy in the same way, i.e.,
Tr ρ X A B log ρ X A B 1 2 I X A σ B 1 ρ X A B 1 2 = H ( X ) + x p ( x ) D ^ ( ρ A B x I A σ B ) ,
taking the optimization of σ B for the above equality and combining the definition of Equation (15). We then obtain the desired result.   □
From the above results, we know that there are two additional terms H ( X ) and log d X on each side of the inequality (24), which are different from the concavity of the quantum conditional entropy S ( A | B ) ρ . To make a distinction, we call it the weak concavity of the BS conditional entropy given by Theorem 1.
Combining the above fact with Theorem 1, we can establish the relationship between S ^ ( A | X B ) ρ and S ^ ( X A | B ) ρ . Using the direct sum property of the BS relative entropy (Proposition 4.54 in [29]), we have
S ^ ( A | X B ) ρ = x p ( x ) D ^ ( ρ A B x I A ρ B x ) .
We cannot determine the order relations between ρ B x with ρ B , but one can always compare D ^ ( ρ A B x I A ρ B x ) with D ^ ( ρ A B x I A ρ B ) . For the case that the former is less than the latter, we have
S ^ ( A | X B ) ρ x p ( x ) D ^ ( ρ A B x I A ρ B ) .
Applying Equation (26), we can obtain
S ^ ( X A | B ) ρ log d X S ^ ( A | X B ) ρ .
In addition, as a special case of the inequality (20), we can easily obtain that S ^ ( A | X B ) ρ S ^ ( A | B ) ρ . Therefore, it holds that
S ^ ( X A | B ) ρ log d X S ^ ( A | X B ) ρ S ^ ( A | B ) ρ .
Otherwise, for the case of D ^ ( ρ A B x I A ρ B x ) D ^ ( ρ A B x I A ρ B ) , we then obtain
H ( X ) + S ^ ( A | X B ) ρ S ^ ( X A | B ) ρ .
For the quantum conditional entropy of Equation (10), we know that any bipartite pure states are entangled if and only if S ( A | B ) < 0 . Here, we are also interested in the BS conditional entropy. Without a loss of generality, let | ψ A B = i λ i | i A | i B be any bipartite pure state, where λ i represents non-negative real numbers satisfying i λ i 2 = 1 , known as Schmidt coefficients, and | i A and | i B are orthonormal states for A and B, respectively. The number of non-zero values λ i is called the Schmidt number for the pure state | ψ A B [17]. We have
S ^ ( A | B ) ψ = Tr ψ A B log ψ A B 1 2 I A ρ B 1 ψ A B 1 2 = Tr ψ A B log I A ρ B 1 2 ψ A B I A ρ B 1 2 = log r ,
where r is the Schmidt number of | ψ . We remark that the bipartite pure state is entangled if the Schmidt number is greater than 1.

3.2. Belavkin–Staszewski Mutual Information

The quantum mutual information is another important measure in quantum information theory, which can describe total correlations in the bipartite quantum subsystems, and there are important applications in quantum channel capacity, quantum cryptography, and quantum thermodynamics [17,35]. Based on the property of the quantum relative entropy, there are four equal definitions for the quantum mutual information, i.e.,
I ( A ; B ) ρ = D ( ρ A B ρ A ρ B ) = min σ B D ( ρ A B ρ A σ B ) = min σ A D ( ρ A B σ A ρ B ) = min σ A , σ B D ( ρ A B σ A σ B ) ,
where the minimums are taken over all density operators σ A and σ B on quantum systems H A and H B , respectively. However, for other general relative entropies, these equalities do not hold in general, such as max-information [24]. In this section, we will consider a new mutual information via the BS relative entropy, Similar to Equation (35), we define four different BS information terms as follows.
Definition 3.
For any quantum state ρ A B S = ( H A B ) , the BS mutual information terms are defined as
I ^ 1 ( A ; B ) ρ = D ^ ( ρ A B ρ A ρ B ) ,
I ^ 2 ( A ; B ) ρ = min σ B D ^ ( ρ A B ρ A σ B ) ,
I ^ 2 ( A ; B ) ρ = min σ A D ^ ( ρ A B σ A ρ B ) ,
I ^ 3 ( A ; B ) ρ = min σ A , σ B D ^ ( ρ A B σ A σ B ) .
Notice that, for I ^ 2 ( A ; B ) ρ and I ^ 2 ( A ; B ) ρ , they can be thought of as swap** the positions of the optimization operators σ A and σ B , so we will consider only one of them. Intuitively, the remaining three definitions of the BS mutual information I ^ i ( A ; B ) ρ decreases with i, i.e.,
I ^ 3 ( A ; B ) ρ I ^ 2 ( A ; B ) ρ I ^ 1 ( A ; B ) ρ .
Additionally, recalling the inequality (7), we can obtain that there is not less BS mutual information than there is quantum mutual information, i.e.,
I ( A ; B ) ρ I ^ i ( A ; B ) ρ .
From the monotonicity of the BS relative entropy, it follows that discarding quantum systems does not increase the BS mutual information, i.e.,
I ^ i ( A ; B ) ρ I ^ i ( A ; B C ) ρ .
Subsequently, for the quantum mutual information (35), it holds that
I ( A ; B ) ρ = S ( ρ A ) S ( A | B ) ρ = S ( ρ B ) S ( B | A ) ρ .
The above two relations are called chain rules for the quantum mutual information. Chain rule can be regarded as a ‘bridge’ between conditional entropy with mutual information in information theory. We are also interested in exploring chain rules for the BS mutual information for bipartite classical-quantum system H X B (for all quantum scenarios, further discussion is needed as a remaining issue). It is well-known that the classical-quantum state only possesses classical correlation, and there is no quantum correlation, which leads us to find some significative and interesting results. We first give the following result before discussing the chain rules for the BS mutual information.
Lemma 3.
For any bipartite classical-quantum states ρ X B , we have
I ^ i ( X ; B ) ρ = min σ B x p ( x ) D ^ ( ρ B x σ B ) ,
where i = 2 , 3 .
Proof. 
The proof is similar to the proof of Lemma 1, thus we omit some calculation steps. We first consider the case I ^ 3 ( X ; B ) ρ . Let σ X = x q ( x ) | x x | , for any quantum state σ B , we have
Tr ρ X B log ρ X B 1 2 σ X 1 σ B 1 ρ X B 1 2 = Tr x p ( x ) | x x | ρ B x · log x p ( x ) q ( x ) 1 | x x | ( ρ B x ) 1 2 σ B 1 ( ρ B x ) 1 2 = Tr x p ( x ) log p ( x ) q ( x ) 1 | x x | ρ B x + x p ( x ) Tr ρ B x log ( ρ B x ) 1 2 σ B 1 ( ρ B x ) 1 2 = D ( p ( x ) q ( x ) ) + x p ( x ) D ^ ( ρ B x σ B ) .
Taking the minimum optimization for both sides of Equation (45) about σ X and σ B , respectively, we then have,
I ^ 3 ( X ; B ) ρ = min σ X , σ B D ( p ( x ) q ( x ) ) + x p ( x ) D ^ ( ρ B x σ B ) = min σ X D ( p ( x ) q ( x ) ) + min σ B x p ( x ) D ^ ( ρ B x σ B ) = min σ B x p ( x ) D ^ ( ρ B x σ B ) .
The last equality follows from the fact that the relative entropy D ( p ( x ) q ( x ) ) is nonnegative; i.e., it holds that
min σ X D ( p ( x ) q ( x ) ) = 0 ,
for all x, if and only if p ( x ) = q ( x ) .
For I ^ 2 ( X ; B ) ρ , only optimization is required for σ B from its definition, so we directly have
I ^ 2 ( X ; B ) ρ = min σ B Tr ρ X B log ρ X B 1 2 ρ X 1 σ B 1 ρ X B 1 2 = min σ B x p ( x ) D ^ ( ρ B x σ B ) .
  □
Notice that the BS mutual information I ^ 1 ( X ; B ) ρ does not involve any optimizations, so we have
I ^ 1 ( X ; B ) ρ = x p ( x ) D ^ ( ρ B x ρ B ) .
This result shows that the BS mutual information I ^ 1 ( X ; B ) ρ is identical in form with the quantum mutual information I ( X ; B ) ρ , while the latter is the well-known Holevo information. In addition, if we consider the tripartite classical-quantum state, we can also obtain a sum form of the BS mutual information for i = 1 , i.e.,
I ^ 1 ( X A ; B ) ρ = x p ( x ) I ^ 1 ( A x ; ρ B ) ,
where I ^ 1 ( A x ; ρ B ) is the BS mutual information between quantum states ρ A B x and ρ A x ρ B . Other cases of the BS mutual information are similar, so we will not go into detail. Based on the above results, we obtain the chain rules for the BS mutual information for bipartite classical-quantum states as follows.
Theorem 2.
For any bipartite classical-quantum state ρ X B , we have
I ^ i ( X ; B ) ρ + S ^ m ( X | B ) ρ = H ( X ) ,
where i = 2 , 3 .
Proof. 
The proof of this theorem is similar to the proof of Theorem 1, so we omit some of the repetition. For any bipartite classical-quantum state ρ X B , we have
D ^ ( ρ X B I X σ B ) = x p ( x ) D ^ ( ρ B x σ B ) H ( X ) .
Employing the definition of Equation (15), we further have
S ^ m ( X | B ) ρ = H ( X ) min σ B x p ( x ) D ^ ( ρ B x σ B ) .
Applying Lemma 3 to Equation (50), we can then complete the proof.   □
Similarly, for I ^ 1 ( X ; B ) ρ , we give a chain rule with respect to the definition of the BS conditional entropy S ^ ( X | B ) ρ as follows.
Corollary 1.
For any bipartite classical-quantum state ρ X B S ( H X B ) , we have
I ^ 1 ( X ; B ) ρ + S ^ ( X | B ) ρ = H ( X ) .
Proof. 
From Definition 2, we can directly obtain that
S ^ ( X | B ) ρ = Tr ρ X B log ρ X B 1 2 I X ρ B 1 ρ X B 1 2 = H ( X ) x p ( x ) D ^ ( ρ B x ρ B ) .
Combining Equation (46) with Equation (52), we then obtain the desired result.   □
Recalling Holevo information and applying the result of Lemma 1, we further have
I ( X ; B ) ρ = S ( ρ B ) S ^ ( B | X ) ρ = S ( ρ B ) S ^ m ( B | X ) ρ = S ( ρ B ) S ( B | X ) ρ .
This result shows that the BS conditional entropy with classical side information can be used to describe the Holevo information as well. In addition, employing the inequality (41), we have
S ( ρ B ) S ^ m ( B | X ) ρ I ^ 3 ( X ; B ) ρ .
Comparing this to Theorem 2, we find that, when the side information is classical, the equal sign of the chain rule for I ^ 3 ( X ; B ) ρ does not hold in general.

3.3. Subadditivity for the BS Relative Entropy

It is necessary to study the relationship of entropic measures between a joint system and its corresponding subsystems, which plays a vital role in estimating channel capacity bounds and analyzing error exponents. The quantum relative entropy satisfies subadditivity and superadditivity, both of which are fundamental properties of the quantum relative entropy [31]. More precisely, for any bipartite quantum states ρ A B and a product state σ A σ B , the superadditivity of the quantum relative entropy is
D ( ρ A σ A ) + D ( ρ B σ B ) D ( ρ A B σ A σ B ) .
This was extended to a more general setting [36]. This paper does not determine whether the BS relative entropy holds the same property. However, the following result shows an opposite relationship for the BS entropy, i.e., the subadditivity. We first give an equivalent definition of the BS relative entropy for obtaining the desired result.
Lemma 4.
For any quantum state ρ S = ( H ) and σ S ( H ) , we have
D ^ ( ρ σ ) = Tr σ ( σ 1 2 ρ σ 1 2 ) log σ 1 2 ρ σ 1 2 .
Proof. 
Let Π σ be the projection onto the support of σ . One can obtain that ρ = ρ Π σ = Π σ ρ from supp ( ρ ) supp ( σ ) . From Equation (5), we thus have
D ^ ( ρ σ ) = Tr ρ log ( ρ 1 2 σ 1 ρ 1 2 ) = Tr ρ 1 2 Π σ ρ 1 2 log ( ρ 1 2 σ 1 2 σ 1 2 ρ 1 2 ) = Tr ρ 1 2 σ 1 2 σ 1 2 ρ 1 2 log ( ρ 1 2 σ 1 2 σ 1 2 ρ 1 2 ) ,
where the equalities holds from the fact that Π σ = σ 1 2 σ 1 2 . Employing Lemma 2.6 in [29], we have
σ 1 2 ρ 1 2 log ( ρ 1 2 σ 1 2 σ 1 2 ρ 1 2 ) = log ( σ 1 2 ρ σ 1 2 ) σ 1 2 ρ 1 2 ,
where ρ 1 2 σ 1 2 = ( σ 1 2 ρ 1 2 ) . We then have
D ^ ( ρ σ ) = Tr ρ 1 2 σ 1 2 log ( σ 1 2 ρ σ 1 2 ) σ 1 2 ρ 1 2 = Tr log ( σ 1 2 ρ σ 1 2 ) σ 1 2 ρ σ 1 2 .
The equality holds from the cyclic property of the trace. Since
log ( σ 1 2 ρ σ 1 2 ) σ 1 2 ρ σ 1 2 = σ 1 2 ρ σ 1 2 log ( σ 1 2 ρ σ 1 2 ) ,
we have
D ^ ( ρ σ ) = Tr log ( σ 1 2 ρ σ 1 2 ) σ 1 2 ρ σ 1 2 σ = Tr σ 1 2 ρ σ 1 2 log ( σ 1 2 ρ σ 1 2 ) σ .
Finally, we obtain the desired result by applying the cyclic property of the trace again.   □
Theorem 3.
For any quantum state ρ A B S = ( H A B ) , σ A S ( H A ) , or σ B S ( H B ) , we have
D ^ ( ρ A B σ A σ B ) λ log λ + D ^ ( ρ A σ A ) + D ^ ( ρ B σ B ) ,
where λ = 2 D max ( ρ A B ρ A ρ B ) .
Proof. 
Let D max ( ρ A B ρ A ρ B ) = log λ for supp ρ A B supp ( ρ A ρ B ) , which implies that ρ A B λ ρ A ρ B . For any quantum state σ A or σ B that satisfies supp ρ A supp σ B and supp ρ A supp σ B , respectively, applying Lemma 4, we have
D ^ ( ρ A B σ A σ B ) = Tr σ A σ B σ A 1 2 σ B 1 2 ρ A B σ A 1 2 σ B 1 2 · log σ A 1 2 σ B 1 2 ρ A B σ A 1 2 σ B 1 2 .
Employing the basic operator inequalities of Lemma 2.13 in [29], we have
σ A 1 2 σ B 1 2 ρ A B σ A 1 2 σ B 1 2 λ σ A 1 2 σ B 1 2 ρ A ρ B σ A 1 2 σ B 1 2 .
Substituting the above inequality into Equation (61), we then have
D ^ ( ρ A B σ A σ B ) Tr σ A σ B λ σ A 1 2 ρ A σ A 1 2 σ B 1 2 ρ B σ B 1 2 · log λ σ A 1 2 ρ A σ A 1 2 σ B 1 2 ρ B σ B 1 2 = λ · Tr σ A σ B σ A 1 2 ρ A σ A 1 2 σ B 1 2 ρ B σ B 1 2 · log λ + log σ A 1 2 ρ A σ A 1 2 + log σ B 1 2 ρ B σ B 1 2 λ log λ + λ D ^ ( ρ A σ A ) + λ D ^ ( ρ B σ B ) .
The equality follows from the linearity of the trace. The last inequality holds based on
Tr σ A σ B σ A 1 2 ρ A σ A 1 2 σ B 1 2 ρ B σ B 1 2 1
and
Tr σ A σ B σ A 1 2 ρ A σ A 1 2 σ B 1 2 ρ B σ B 1 2 log σ A 1 2 ρ A σ A 1 2 = Tr σ A σ A 1 2 ρ A σ A 1 2 log σ A 1 2 ρ A σ A 1 2 = D ^ ( ρ A σ A ) .
Similarly,
Tr σ A σ B σ A 1 2 ρ A σ A 1 2 σ B 1 2 ρ B σ B 1 2 log σ B 1 2 ρ B σ B 1 2 = D ^ ( ρ B σ B ) .
  □
As mentioned above, the geometric Rényi relative entropy converges to the BS relative entropy when the limit is α 1 . More generally, we also provide an upper bound for the geometric Rényi relative entropy.
Theorem 4.
For any quantum state ρ A B S = ( H A B ) , σ A S ( H A ) , or σ B S ( H B ) , we have
D ^ α ( ρ A B σ A σ B ) α α 1 D max ( ρ A B ρ A ρ B ) + D ^ α ( ρ A σ A ) + D ^ α ( ρ B σ B ) .
Proof. 
Let D max ( ρ A B ρ A ρ B ) = log λ for supp ρ A B supp ( ρ A ρ B ) . This implies that ρ A B λ ρ A ρ B . For σ A and σ B with supp ρ A supp σ B and supp ρ A supp σ B , respectively, employing the relation of Eq. (4.6.21) in [29], we have
Tr σ A σ B σ A 1 2 σ B 1 2 ρ A B σ A 1 2 σ B 1 2 α = Tr ρ A B ρ A B 1 2 σ A 1 σ B 1 ρ A B 1 2 α 1 .
It then holds that
Tr ρ A B ρ A B 1 2 σ A 1 σ B 1 ρ A B 1 2 α 1 Tr λ ρ A ρ B λ ρ A 1 2 ρ B 1 2 σ A 1 σ B 1 ρ A 1 2 ρ B 1 2 α 1 = λ α Tr ρ A ρ B ρ A 1 2 ρ B 1 2 σ A 1 σ B 1 ρ A 1 2 ρ B 1 2 α 1 = λ α Tr ρ A ρ A 1 2 σ A 1 ρ A 1 2 α 1 Tr ρ B ρ B 1 2 σ B 1 ρ B 1 2 α 1 .
Combining the above result with Definition 1, we have
D ^ α ( ρ A B σ A σ B ) 1 α 1 log λ α Tr ρ A ρ A 1 2 σ A 1 ρ A 1 2 α 1 Tr ρ B ρ B 1 2 σ B 1 ρ B 1 2 α 1 = α α 1 D max ( ρ A B ρ A ρ B ) + D ^ α ( ρ A σ A ) + D ^ α ( ρ B σ B ) .
  □
Similarly, one can define a new mutual information term via the geometric Rényi relative entropy as
I ^ α ( A ; B ) ρ = min σ A , σ B D ^ α ( ρ A B σ A σ B ) .
It is then easy to draw the following conclusion.
Corollary 2.
For any quantum state ρ A B S = ( H A B ) , σ A S ( H A ) , or σ B S ( H B ) , we have
I ^ α ( A ; B ) ρ α α 1 D max ( ρ A B ρ A ρ B ) + D ^ α ( ρ A σ A ) + D ^ α ( ρ B σ B ) .
Notably, if one considers the classical-quantum state, there is no result as shown in Lemma 3 for the mutual information defined by the geometric Rényi relative entropy. Specifically, for α ( 0 , 1 ) , we have
D ^ α ( ρ X B ρ X ρ B ) = 1 α 1 log x p ( x ) Tr ρ B ρ B 1 2 ρ B x ρ B 1 2 α 1 α 1 x p ( x ) log Tr ρ B σ B 1 2 ρ B x ρ B 1 2 α = x p ( x ) D ^ α ( ρ B x ρ B ) ,
where the inequality comes from the Jensen inequality of log t . For α ( 1 , ) , we obtain the opposite result, i.e.,
D ^ α ( ρ X B ρ X ρ B ) x p ( x ) D ^ α ( ρ B x ρ B ) .
Furthermore, if one considers the conditional entropy defined by the geometric Rényi relative entropy, for α ( 0 , 1 ) , we have
D ^ α ( ρ X B I X ρ B ) x p ( x ) D ^ α ( ρ B x ρ B ) H ( X ) .
For α ( 1 , ) , we then have
D ^ α ( ρ X B I X ρ B ) x p ( x ) D ^ α ( ρ B x ρ B ) H ( X ) .

4. Conclusions

This paper investigates the subadditivity of the geometric Rényi and BS relative entropies and explores the indispensable properties of the BS conditional entropy and mutual information, especially in classical-quantum settings. The subadditivity of the geometric Rényi and BS relative entropies can provide new valuable bounds to estimate channel capacity and analyze the error exponent. As mentioned above, the BS relative entropy represents a different quantum generalization of classical relative entropy. The main use of BS relative entropy is in establishing upper bounds for the rates of feedback-assisted quantum communication protocols. The primary goal of further research on BS relative entropy is to explore the intrinsic properties of its relevant conditional entropy and mutual information and to gain a better understanding of their operational relevance. We hope that the formal tools provided in this paper will be useful for this purpose.
One question worth answering is whether there is a chain rule for the mutual information in terms of the geometric Rényi relative entropy, i.e.,
D ^ α ( ρ A B ρ A ρ B ) S ^ α ( ρ A ) + D ^ α ( ρ A B I A ρ B ) ,
or the other forms, where S ^ α ( ρ A ) is the quantum Rényi entropy. Subsequently, the duality of conditional entropy is an important property for a tripartite pure state system, which can be effectively applied in random number extraction and channel coding [26]. Further research will focus on the duality of the BS conditional entropy.

Author Contributions

Conceptualization, Y.Z. and Z.X.; methodology, Y.Z. and Z.X.; formal analysis, Y.Z.; writing—original draft preparation, Y.Z.; writing—review and editing, Y.Z., B.Y. and Z.X.; project administration, B.Y. and Z.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Nos. 62171266, U2001205) and the Funded Projects for the Academic Leaders and Academic Backbones, Shaanxi Normal University (No. 16QNGG013).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

Z.X. is supported by the National Natural Science Foundation of China (No. 62171266) and by the Funded Projects for the Academic Leaders and Academic Backbones, Shaanxi Normal University (No. 16QNGG013). B.Y. is supported by the National Natural Science Foundation of China (No. U2001205).

Conflicts of Interest

The authors declare that there is no conflict of interest. The funders had no role in the design of the study, in the collection, analyses, or interpretation of data, in the writing of the manuscript, or in the decision to publish the results.

References

  1. Rényi, A. On measures of information and entropy. Proc. Symp. Math. Stat. Probab. 1961, 4, 547–561. [Google Scholar]
  2. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  3. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley Sons, Inc.: Hoboken, NJ, USA, 2006. [Google Scholar]
  4. Csiszár, I.; Körner, J. Information Theory: Coding Theorems for Discrete Memoryless Systems; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  5. Tomamichel, M. Quantum information processing with finite resources. ar**s on matrices. Infin. Dimens. Anal. Quantum Probab. Relat. Top. 1998, 1, 83–89. [Google Scholar] [CrossRef]
  6. Matsumoto, K. A new quantum version of f-divergence. ar**v 2013, ar**v:1311.4722. [Google Scholar]
  7. Umegaki, H. Conditional expectations in an operator algebra IV (entropy and information). Kodai Math. J. 1962, 14, 59–85. [Google Scholar] [CrossRef]
  8. Nielsen, M.A.; Chuang, I.L. Quantum Computation and Quantum Information; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
  9. Belavkin, V.P.; Staszewski, P. C*-algebraic generalization of relative entropy and entropy. Ann. Inst. Henri Poincare 1982, 37, 51–58. [Google Scholar]
  10. Katariya, V.; Wilde, M.M. Geometric distinguishability measures limit quantum channel estimation and discrimination. ar**v 2020, ar**v:2004.10708. [Google Scholar] [CrossRef]
  11. Bluhm, A.; Capel, A. A strengthened data processing inequality for the Belavkin-Staszewski relative entropy. Rev. Math. Phys. 2020, 32, 2050005. [Google Scholar] [CrossRef] [Green Version]
  12. Hiai, F.; Petz, D. The proper formula for relative entropy and its asymptotics in quantum probability. Commun. Math. Phys. 1991, 143, 99–114. [Google Scholar] [CrossRef]
  13. Bluhm, A.; Capel, A.; Perez-Hernandez, A. Weak quasi-factorization for the Belavkin-Staszewski relative entropy. In Proceedings of the 2021 IEEE International Symposium on Information Theory (ISIT), Melbourne, Australia, 12–20 July 2021; pp. 118–123. [Google Scholar]
  14. Fang, K.; Fawzi, H. Geometric Rényi divergence and its applications in quantum channel capacities. Commun. Math. Phys. 2021, 384, 1615–1677. [Google Scholar] [CrossRef]
  15. Ciganović, N.; Beaudry, N.J.; Renner, R. Smooth max-information as one-shot generalization for mutual information. IEEE Trans. Inform. Theory 2014, 60, 1573–1581. [Google Scholar] [CrossRef] [Green Version]
  16. Watanabe, Y. Randomness Extraction via a Quantum Generalization of the Conditional Collision Entropy. IEEE Trans. Inform. Theory 2020, 66, 1171–1177. [Google Scholar] [CrossRef]
  17. Tomamichel, M.; Colbeck, R.; Renner, R. Duality between smooth min- and max-entropies. IEEE Trans. Inf. Theory 2010, 56, 4674–4681. [Google Scholar] [CrossRef] [Green Version]
  18. König, R.; Renner, R.; Schaffner, C. The operational meaning of min- and max-entropy. IEEE Trans. Inf. Theory 2009, 55, 4337–4347. [Google Scholar] [CrossRef] [Green Version]
  19. Vitanov, A.; Dupuis, F.; Tomamichel, M.; Renner, R. Chain rules for smooth min- and max-entropies. IEEE Trans. Inf. Theory 2013, 59, 2603–2612. [Google Scholar] [CrossRef] [Green Version]
  20. Khatri, S.; Wilde, M.M. Principles of quantum communication theory: A modern approach. ar**v 2020, ar**v:2011.04672v1. [Google Scholar]
  21. Datta, N. Min- and max-relative entropies and a new entanglement monotone. IEEE Trans. Inf. Theory 2009, 55, 2816–2826. [Google Scholar] [CrossRef] [Green Version]
  22. Wilde, M.M. Quantum Information Theory, 2nd ed.; Cambridge University Press: Cambridge, UK, 2017. [Google Scholar]
  23. Chitambar, E.; Gour, G. Quantum resource theories. Rev. Mod. Phys. 2019, 91, 025001. [Google Scholar] [CrossRef] [Green Version]
  24. Lawson, J.D.; Lim, Y. The geometric mean, matrices, metrics, and more. Am. Math. Mon. 2001, 108, 797–812. [Google Scholar] [CrossRef]
  25. Renner, R. Security of Quantum Key Distribution. Ph.D. Thesis, ETH Zurich, Zurich, Switzerland, 2005. [Google Scholar]
  26. Sagawa, T. Entropy, divergence, and majorization in classical and quantum thermodynamics. ar**v 2020, ar**v:2007.09974v1. [Google Scholar]
  27. Capel, Á.; Lucia, A.; Pérez-García, D. Superadditivity of quantum relative entropy for general states. IEEE Trans. Inf. Theory 2018, 64, 4758–4765. [Google Scholar] [CrossRef] [Green Version]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhai, Y.; Yang, B.; **, Z. Belavkin–Staszewski Relative Entropy, Conditional Entropy, and Mutual Information. Entropy 2022, 24, 837. https://doi.org/10.3390/e24060837

AMA Style

Zhai Y, Yang B, ** Z. Belavkin–Staszewski Relative Entropy, Conditional Entropy, and Mutual Information. Entropy. 2022; 24(6):837. https://doi.org/10.3390/e24060837

Chicago/Turabian Style

Zhai, Yuan, Bo Yang, and Zhengjun **. 2022. "Belavkin–Staszewski Relative Entropy, Conditional Entropy, and Mutual Information" Entropy 24, no. 6: 837. https://doi.org/10.3390/e24060837

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop