# Quantum relative entropy

The quantum relative entropy, in analogy with the classical relative entropy or Kullback-Leibler divergence, is a measure of closeness between two states. It was defined by Denis Petz as :$D\left(\rho||\sigma\right)=Tr\rho\left(\log\rho-\log\sigma\right)$ where $\rho$ and $\sigma$ are two density matrices. It has a number of applications including its use in proving strong sub-additivity. One can also define a relative entropy distance to some set of states, where one minimises $D\left(\rho||\sigma\right)$ over all $\sigma$ from some convex set. If the convex set is chosen to be the set of seperable states, then the relative entropy distance is an entanglement measure, called the relative entropy of entanglement. If the set of states are product states, then the relative entropy distance is the quantum mutual information == See also == * Entropy of entanglement {{stub}} Category:Quantum Information Theory Category:Handbook of Quantum Information Category:Entropy