Secret Key Agreement¶
One of the only methods of encrypting a message from Alice to Bomb such that no third party (Eve) can possibly decrypt it is a one-time pad.
This technique requires that Alice and Bob have a secret sequence of bits, \(S\), which Alice then encrypts by computing the exclusive-or of it with the plaintext, \(P\), to produce the cyphertext, \(C\): \(C = S \oplus P\).
Bob can then decrypt by xor
ing again: \(P = S \oplus C\).
In order to pull this off, Alice and Bob need to construct \(S\) out of some sort joint randomness, \(p(x, y, z)\), and public communication, \(V\), which is assumed to have perfect fidelity. The maximum rate at which \(S\) can be constructed in the secret key agreement rate.
Background¶
Given \(N\) IID copies of a joint distribution governed by \(p(x, y, z)\), let \(X^N\) denote the random variables observed by Alice, \(Y^N\) denote the random variables observed by Bob, and \(Z^N\) denote the random variables observed by Even. A secret key agreement scheme consists of functions \(f\) and \(g\), as well as a protocol for public communication (producing \(V\)), and is considered \(R\)-achievable if:
The maximum rate \(R\) such that there exists a \(R\)-achievable scheme is known as the secret key agreement rate. Intuitively, this means there exists some procedure such that, for every \(N\) observations, Alice and Bob can publicly converse and then construct \(S\) bits which agree almost surely, and are almost surely independent of everything Eve has access to. \(S\) is then known as a secret key.
There are three general classes of secret key agreement rates, depending on which parties are permitted to communicate. We discuss them below.
No Communication¶
In the case that neither Alice nor Bob are permitted communication, the no-communication secret key agreement rate is given by:
where \(X \meet Y\) is the Gács-Körner Common Information variable.
Secrecy Capacity¶
Consider the situation that no party is allowed to communication, but rather than passively observing \(p(X, Y, Z)\) Alice has full access to driving the channel \(p(Y, Z | X)\). In this case we arrive at a maximum secret-key agreement rate known as the secrecy capacity, which is given by:
One-Way Communication¶
If only Alice is allowed to publicly broadcast information, the secret key agreement rate is given by:
Two-Way Communication¶
When both Alice and Bob are permitted communication, the secret key agreement rate, \(\operatorname{S}[X \leftrightarrow Y || Z]\), is much more difficult to compute, and in fact only upper and lower bounds on this rate are known.
Lower Bounds¶
The first few lower bounds on two-way secret key agreement rate are simply symmetrized forms of the more restricted secret key agreement rates.
Lower Intrinsic Mutual Information¶
The first lower bound on the secret key agreement rate is known in dit
as the lower_intrinsic_mutual_information()
, and is given by:
Secrecy Capacity¶
Next is the secrecy capacity:
This gives the secret key agreement rate when communication is not allowed.
Necessary Intrinsic Mutual Information¶
A tighter bound is given by the necessary_intrinsic_mutual_information()
[GGunluK17], which is the maximum of the two one-way secret key agreement rates:
Interactive Intrinsic Mutual Information¶
Upper Bounds¶
Upper Intrinsic Mutual Information¶
The secret key agreement rate is trivially upper bounded by:
Intrinsic Mutual Information *****************&******
The intrinsic_mutual_information()
[MW97] is defined as:
It is straightforward to see that \(p(\overline{z} | z)\) being a constant achieves \(\I{X : Y}\), and \(p(\overline{z} | z)\) being the identity achieves \(\I{X : Y | Z}\).
Reduced Intrinsic Mutual Information¶
This bound can be improved, producing the reduced_intrinsic_mutual_information()
[RSW03]:
This bound improves upon the Intrinsic Mutual Information when a small amount of information, \(U\), can result in a larger decrease in the amount of information shared between \(X\) and \(Y\) given \(Z\) and \(U\).
Minimal Intrinsic Mutual Information¶
The Reduced Intrinsic Mutual Information can be further reduced into the minimal_intrinsic_total_correlation()
[GA17]:
Two-Part Intrinsic Mutual Information¶
All Together Now¶
Taken together, we see the following structure:
Generalizations¶
Most of the above bounds have straightforward multivariate generalizations. These are not necessarily bounds on the multiparty secret key agreement rate. For example, one could compute the minimal_intrinsic_dual_total_correlation()
:
Examples¶
Let us consider a few examples:
In [1]: In [1]: from dit.multivariate.secret_key_agreement import *
First, we consider the distribution intrinsic_1
:
In [2]: In [3]: print(intrinsic_1)
---------------------------------------------------------------------------
NameError Traceback (most recent call last)
<ipython-input-2-b266ad93d4d5> in <module>
----> 1 print(intrinsic_1)
NameError: name 'intrinsic_1' is not defined
In [3]: Class: Distribution
In [4]: Alphabet: ('0', '1', '2', '3') for all rvs
...: Base: linear
...: Outcome Class: str
...: Outcome Length: 3
...: RV Names: None
...:
File "<ipython-input-4-f997bcd614eb>", line 1
Alphabet: ('0', '1', '2', '3') for all rvs
^
SyntaxError: invalid syntax
With upper bounds:
In [5]: In [4]: upper_intrinsic_mutual_information(intrinsic_1, [[0], [1]], [2])
---------------------------------------------------------------------------
NameError Traceback (most recent call last)
<ipython-input-5-adf4d13a797d> in <module>
----> 1 upper_intrinsic_mutual_information(intrinsic_1, [[0], [1]], [2])
NameError: name 'intrinsic_1' is not defined
In [6]: Out[4]: 0.5
We see that the trivial upper bound is 0.5, because without conditioning on \(Z\), \(X\) and \(Y\) can agree when the observe either a \(2\) or a \(3\), which results in \(\I{X : Y} = 0.5\). Given \(Z\), however, that information is no longer private. But, given \(Z\), a conditional dependence is induced between \(X\) and \(Y\): \(Z\) knows that if she is a \(0\) that \(X\) and \(Y\) agree, and if she is a \(1\) they disagree. This results \(\I{X : Y | Z} = 0.5\). In either case, however, \(X\) and \(Y\) can not agree upon a secret key: in the first case the eavesdropper knows their correlation, while in the second they are actually independent.
The intrinsic_mutual_information()
, however can detect this:
In [7]: In [5]: intrinsic_mutual_information(intrinsic_1, [[0], [1]], [2])
---------------------------------------------------------------------------
NameError Traceback (most recent call last)
<ipython-input-7-cf75a74ddcab> in <module>
----> 1 intrinsic_mutual_information(intrinsic_1, [[0], [1]], [2])
NameError: name 'intrinsic_1' is not defined
In [8]: Out[5]: 0.0
Next, let’s consider the distribution intrinsic_2
:
In [9]: In [7]: print(intrinsic_2)
---------------------------------------------------------------------------
NameError Traceback (most recent call last)
<ipython-input-9-094f23e395aa> in <module>
----> 1 print(intrinsic_2)
NameError: name 'intrinsic_2' is not defined
In [10]: Class: Distribution
In [11]: Alphabet: (('0', '1', '2', '3'), ('0', '1', '2', '3'), ('0', '1'))
In [12]: Base: linear
---------------------------------------------------------------------------
NameError Traceback (most recent call last)
<ipython-input-12-822bc640d06e> in <module>
----> 1 Base: linear
NameError: name 'linear' is not defined
In [13]: Outcome Class: str
....: Outcome Length: 3
....: RV Names: None
....: