1 Introduction
The problem of compression has been extensively studied in the field of information theory and, more recently, in computational complexity and cryptography [23, 27, 40, 42]. Informally, given a distribution , compression aims to efficiently encode samples from
as short strings while at the same time being able to efficiently recover these samples. While the typical information-theoretic study of compression considers the case of compressing multiple independent samples from the same source X, its study in computer science, and in particular in this work, considers the “single-shot” case. Compression in this setting is closely related to randomness condensers [18, 34, 38, 39] and resource-bounded Kolmogorov complexity [32, 33] – two well-studied problems in computational complexity. Randomness condensers, which relax randomness extractors, are functions that efficiently map an input distribution into an output distribution with a higher entropy rate. A randomness condenser can be viewed as an efficient compression algorithm, without a corresponding efficient decompression algorithm. The resource-bounded Kolmogorov complexity of a string is the smallest description length of an efficient program that outputs this string. This program description can be viewed as a compressed string, such that decoding is efficiently possible, while finding the compressed string may be inefficient.
An important property of efficient compression algorithms, which combines the efficiency features of randomness condensers and resource-bounded Kolmogorov complexity, is their ability to efficiently produce “random-looking” outputs while allowing the original input to be efficiently recovered. Despite the large body of work on compression and its computational variants, this fundamental property has, to our knowledge, never been the subject of a dedicated study. In this work, we fill this gap by initiating such a study. Before formalizing the problem, we give a simple motivating example.
Consider the goal of encrypting a sample x from a distribution X (say, a random 5-letter English word from the Merriam-Webster Dictionary) using a low-entropy secret key k. Applying a standard symmetric-key encryption scheme with a key derived from k gives rise to the following brute-force attack: Try to decrypt with different keys until obtaining in the support of X. In the typical case that wrong keys always lead to
outside the support of X, this attack successfully recovers x. Variants of this attack arise in different scenarios, including password-authenticated key exchange [4], honey encryption [30], subliminal communication and steganography [26], and more. A natural solution is to use perfect compression: if x can be compressed to a uniformly random string
before being encrypted, it cannot be distinguished from another random string
obtained by trying the wrong key. Note, however, that compression may be an overkill for this application. Instead, it suffices to efficiently encode x into a (possibly longer) pseudorandom string from which x can be efficiently decoded. This more general solution motivates the question we consider in this work.



![$$\begin{aligned} \Pr \big [y\leftarrow X_\lambda :\mathsf {D}_{X}(\mathsf {E}_{X}(y))=y&\big ]\text { is overwhelming and} \end{aligned}$$](../images/508076_1_En_23_Chapter/508076_1_En_23_Chapter_TeX_Equ1.png)












Our motivation for studying pseudorandom encodings stems from the fact that this very natural problem appears in a wide variety of – sometimes seemingly unrelated – problems in cryptography. We already mentioned steganography, honey encryption, and password-authenticated key exchange; we will cover more such connections in this work. Yet, this notion of encoding has to our knowledge never been studied systematically. In this work we study several natural flavors of this notion, obtain positive and negative results about realizing them, and map their connections with other problems in cryptography.
The main focus of this work is on the hypothesis that all efficiently samplable distributions admit a pseudorandom encoding. Henceforth, we refer to this hypothesis the pseudorandom encoding hypothesis .
For describing our results, it will be convenient to use the following general notion of efficiently samplable distributions. A distribution family ensemble (for some language
) is efficiently samplable if there exists a probabilistic polynomial time (PPT) algorithm
such that
is distributed according to
for every
. In case the distribution does not depend on additional inputs,
can be considered equal to
.
Overview of Contributions. Following is a brief summary of our main contributions. We will give an expanded overview of the contributions and the underlying techniques in the rest of this section.
We provide a unified study of different flavors of pseudorandom encodings (PRE) and identify computational, randomized PRE in the CRS model as a useful and achievable notion.
We establish a two-way relation between PRE and the previously studied notion of invertible sampling This reveals unexpected connections between seemingly unrelated problems in cryptography (e.g., between adaptively secure computation for general functionalities and “honey encryption”).
We bootstrap “adaptive PRE” from “static PRE” As a consequence, one can base succinct adaptively secure computation on standard
as opposed to subexponential
[15].
We use PRE to obtain a compiler from standard secure multiparty computation (MPC) protocols to covert MPC protocols.
1.1 Flavors of Pseudorandom Encoding
The notion of pseudorandom encoding has several natural flavors, depending on whether the encoding algorithm is allowed to use randomness or not, and whether the pseudorandomness property satisfies a computational or information-theoretic notion of indistinguishability. We denote the corresponding hypotheses that every efficiently samplable distribution can be pseudorandomly encoded according to the above variants as ,
,
and
.2
Further, we explore relaxations which rely on a trusted setup assumption: we consider the pseudorandom encoding hypothesis in the common reference string model, in which a common string sampled in a trusted way from some distribution is made available to the parties. This is the most common setup assumption in cryptography and it is standard to consider the feasibility of cryptographic primitives in this model to overcome limitations in the plain model. That is, we ask whether for every efficiently samplable distribution , there exists an efficiently samplable CRS distribution and efficient encoding and decoding algorithms
as above, such that correctness and pseudorandomness hold, where the encoding and decoding algorithm as well as the distinguisher receive the CRS as input, and the distributions in Eqs. (1) and (2) are additionally over the choice of the CRS.
Considering distributions which may depend on an input further entails two different flavors. On the one hand, we consider the notion where inputs m are chosen adversarially but statically (that is, independent of the CRS) and, on the other hand, we consider the stronger notion where inputs m are chosen adversarially and adaptively depending on the CRS. We henceforth denote these variants by the prefix “
” and “
”, respectively.
Static-to-Adaptive Transformation. The adaptive notion, where inputs may be chosen depending on the CRS, is clearly stronger than the static notion. However, surprisingly, the very nature of pseudorandom encodings allows one to apply an indirection argument similar to the one used in [11, 12, 25], which yields a static-to-adaptive transformation.
(informal). If all efficiently samplable distributions can be pseudorandomly encoded in the CRS model with a static choice of inputs, then all efficiently samplable distributions can be pseudorandomly encoded in the CRS model with an adaptive choice of inputs.
Static-to-adaptive transformations in cryptography are generally non-trivial, and often come at a big cost in security when they rely on a “complexity leveraging” technique. This connection and its application we will discuss below are a good demonstration of the usefulness of the notion of pseudorandom encodings.
Relaxing Compression. The notion of statistical deterministic pseudorandom encodings recovers the notion of optimal compression. Hence, this conflicts with the existence of one-way functions.3 In our systematic study of pseudorandom encodings, we gradually relax perfect compression in several dimensions, while maintaining one crucial property – the indistinguishability of the encoded distribution from true randomness.
Example. To illustrate the importance of this property, we elaborate on the example we outline at the beginning of the introduction, focusing more specifically on password-authenticated key exchange (PAKE). A PAKE protocol allows two parties holding a (low entropy) common password to jointly and confidentially generate a (high entropy) secret key, such that the protocol is resilient against offline dictionary attacks, and no adversary can establish a shared key with a party if he does not know the matching password. A widely used PAKE protocol due to Bellovin and Merritt [4] has a very simple structure: the parties use their low-entropy password to encrypt the flows of a key-exchange protocol using a block cipher. When the block cipher is modeled as a random cipher, it has the property that decrypting a ciphertext (of an arbitrary plaintext) under an incorrect secret key yields a fresh random plaintext. Thus, Bellovin and Merritt point out that the security of their PAKE protocol requires that “the message to be encrypted by the password must be indistinguishable from a random number.” This is easy to achieve for Diffie-Hellman key exchange over the multiplicative group of integers modulo a prime p. However, for elliptic curve groups this is no longer the case, and one needs to resort to alternative techniques including nontrivial point compression algorithms that compress the representation of a random group element into a nearly uniform bitstring [6].
Clearly, our relaxation of compression does not preserve the useful property of obtaining outputs that are shorter than the inputs. However, the remaining pseudorandomness property is good enough for many applications.
In the following, we elaborate on our weakest notion of pseudorandom encodings, that is, pseudorandom encodings allowing the encoding algorithm to be randomized and providing a computational pseudorandomness guarantee. We defer the discussion on the stronger statistical or deterministic variants to Sect. 1.3, where we derive negative results for most of these stronger notions, which leaves computational randomized pseudorandom encodings as the “best possible” notion that can be realized for general distributions.
Randomized, Computational Pseudorandom Encodings. Computational randomized pseudorandom encodings allow the encoding algorithm to be randomized and require only computational pseudorandomness.






(informal). A distribution admits a pseudorandom encoding if and only if it admits invertible sampling.
Intuitively, the efficient encoding algorithm corresponds to the inverse sampling algorithm, and decoding an encoded string corresponds to sampling with the de-randomized alternative sampler . This equivalence immediately extends to all variants of pseudorandom encodings and corresponding variants of invertible sampling we introduce in this work. Invertible sampling is itself connected to other useful cryptographic notions, such as oblivious sampling, trusted common reference string generations, and adaptively secure computation (which we will elaborate upon below).
Building on this connection, the impossibility result of [29] translates to our setting. On a high level, extractable one-way functions (EOWFs) conflict with invertible sampling because they allow to extract a “secret” (in this case a pre-image) from an image, independently of how it was computed. This conflicts with invertible sampling because invertible sampling is about sampling without knowing the secrets.
(informal, [29]). Assuming the existence of extractable one-way functions (EOWF) and a non-interactive zero-knowledge proof system, does not hold.
This suggests that towards a realizable notion of pseudorandom encodings, a further relaxation is due. Thus, we ask whether the above impossibility result extends to the CRS model. In the CRS model, the above intuition why ISH conflicts with EOWFs fails, because the CRS can contain an obfuscated program that samples an image using some secret, but does not output this secret.
Dachman-Soled, Katz, and Rao [16] (building on the universal deniable encryption construction of Sahai and Waters [35]) construct a so-called “explainability compiler” that implies based on indistinguishability obfuscation4 (
). By our equivalence theorem above, this implies pseudorandom encodings for all efficiently samplable distributions in the CRS model, with static choice of inputs, from
. Invoking the static-to-adaptive transformation detailed above, this also applies to the adaptive variant.
(informal). Assuming the existence of (polynomially secure) indistinguishability obfuscation and one-way functions, holds.
Note that [29] claim that their impossibility result extends to the CRS model, whereas the above theorem seems to suggest the opposite. We show that technically the result of [29] does extend to the CRS model at the cost of assuming unbounded auxiliary-input extractable one-way functions, a strong flavor of EOWFs that seems very unlikely to exist but cannot be unconditionally ruled out.
(informal). Assuming the existence of extractable one-way functions with unbounded common auxiliary input and a non-interactive zero-knowledge proof system, does not hold.
In fact, this apparent contradiction has been the source of some confusion in previous works: the work of [29] makes an informal claim that their impossibility result for ISH extends to the CRS model. However, due to the connection between ISH and adaptively secure MPC (which we will discuss in more details later on), this claim was challenged in [16]: the authors achieve a construction of adaptively secure MPC for all functionalities assuming , which seemingly contradicts the claim of [29]. The authors of [16] therefore stated that the “impossibility result of Ishai et al. [...] does not hold in the CRS model.” Our extension clarifies that the distinction is in fact more subtle: the result of [29] does extend to the CRS model, but at the cost of assuming EOWF with unbounded auxiliary inputs. This does not contradict the constructions based on
, because
and EOWF with unbounded auxiliary inputs are known to be contradictory [5].

An overview of the relations between the pseudorandom encoding hypothesis and other fields of cryptography and computational complexity theory. For simplicity, our static-to-adaptive transformation only appears in the computational, randomized setting in this overview, but also applies to the other settings. (Since the deterministic variants of the pseudorandom encoding hypothesis are impossible for some pathologic samplers, the arrows between deterministic and randomized variants of the pseudorandom encoding hypothesis are to be read as if the deterministic variant is true for some sampler, then the corresponding randomized variant is true for that sampler.)
Further Relaxation. We further study an additional relaxation of pseudorandom encodings, where we allow the encoding algorithm to run in super-polynomial time. We show that this relaxed variant can be achieved from cryptographic primitives similar to extremely lossy functions [45], which can be based on the exponential hardness of the decisional Diffie-Hellman problem – a strong assumption, but (still) more standard than indistinguishability obfuscation. However, the applicability of the resulting notion turns out to be rather restricted.
1.2 Implications and Applications of Our Results
In this section, we elaborate on the implications of the techniques we develop and the results we obtain for a variety of other cryptographic primitives.
New Results for Adaptively Secure Computation. As mentioned above, a sampler admits invertible sampling if and only if it can be pseudorandomly encoded. A two-way connection between invertible sampling and adaptively secure MPC for general randomized functionalities was established in [29]. An MPC protocol allows two or more parties to jointly evaluate a (possibly randomized) functionality on their inputs without revealing anything to each other except what follows from their inputs and outputs. This should hold even in the presence of an adversary who can corrupt any number of parties in an adaptive (sequential) fashion. When we write “adaptive MPC”, we mean adaptive MPC for all randomized functionalities. This should be contrasted with weaker notions of adaptive MPC for strict subsets of corrupted parties [3, 9, 20] or for adaptively well-formed functionalities5 [10] which can both be done from mild assumptions. The connection from [29] shows that adaptive MPC for all randomized functions is possible if and only if every PPT sampler admits invertible sampling, i.e., the invertible sampling hypothesis is true.
We show that this result generalizes to the global CRS model. More precisely, we prove the adaptive variant of the pseudorandom encoding hypothesis in the CRS model is equivalent to adaptive MPC in the global CRS model.6
As detailed above, the static pseudorandom encoding hypothesis in the CRS model follows from
(and one-way functions). Applying our static-to-adaptive transformation, the same holds for the adaptive variant. Thus, we obtain the first instantiation of an adaptive explainability compiler [16] without complexity leveraging and, hence, based only on polynomial hardness assumptions. The recent work of Cohen, shelat, and Wichs [15] uses such an adaptive explainability compiler to obtain succinct adaptive MPC, where “succinct” means that the communication complexity is sublinear in the complexity of the evaluated function. Due to our instantiation of
from polynomial
, we improve the results of [15] by relaxing the requirement for subexponentially secure
to polynomially secure
in a black-box way.
(informal). Assuming the existence of polynomially secure indistinguishability obfuscation and the adaptive hardness of the learning with errors problem, then malicious, two-round, UC-secure adaptive MPC and sublinear communication complexity is possible (in the local CRS model, for all deterministic functionalities).
Steganography and Covert Multi-party Computation. We explore the connection of the pseudorandom encoding hypothesis to various flavors of steganography. The goal of steganography, informally, is to embed secret messages in distributions of natural-looking messages, in order to hide them from external observers. While the standard setting for steganography relies on shared secret keys to encode the messages, we show that pseudorandom encodings naturally give rise to a strong form of keyless steganography. Namely, one can rely on pseudorandom encodings to encode any message into an innocent-looking distribution, without truly hiding the message (since anyone can decode the stream), but providing plausible deniability, in the sense that, even with the decoded message, it is impossible to tell apart whether this message was indeed encoded by the sender, or whether it is simply the result of decoding the innocent distribution.
(informal). Assuming pseudorandom encodings, then there exists a keyless steganographic protocol which provides plausible deniability.
Plausible deniability is an important security notion; in particular, an important cryptographic primitive in this area is the notion of (sender-)deniable encryption [8], which is known to exist assuming indistinguishability obfuscation [35]. Deniable encryption enables to “explain” ciphertexts produced for some message to any arbitrary other message by providing corresponding random coins for a faked encryption process. We view it as an interesting open problem to build deniable encryption under the pseudorandom encoding hypothesis together with more standard cryptographic primitives; we make a first step in this direction and show the following: the statistical variant of pseudorandom encodings, together with the existence of public-key encryption, implies deniable encryption. Interestingly, we also show that the computational randomized pseudorandom encoding hypothesis suffices to imply non-committing encryption, a weaker form of deniable encryption allowing to explain only simulated ciphertexts to arbitrary messages [9].
Covert Secure Computation. Covert MPC [13, 41] is an intriguing flavor of MPC that aims at achieving the following strong security guarantee: if the output of the protocol is not “favorable,” the transcript of the interaction should not leak any information to the parties parties, including whether any given party was actually taking part in the protocol. This strong form of MPC aims at providing security guarantees when the very act of starting a computation with other parties should remain hidden. As an example [41], suppose that a CIA agent who infiltrated a terrorist group wants to make a handshake with another individual to find out whether she is also a CIA agent. Here, we show that pseudorandom encodings give rise to a general compiler transforming a standard MPC protocol into a covert one, in a round-preserving way. The idea is to encode each round of the protocol such that encoded messages look random. Together with the equivalence between adaptively secure MPC and pseudorandom encodings, this gives a connection between two seemingly unrelated notions of secure computation.
(informal). Assuming adaptively secure MPC for all functionalities, there exists a round-preserving compiler that transforms a large class of “natural” MPC protocols into covert MPC protocols (in the static, semi-honest setting).
Other Results. Due to our infeasibility results of , distribution transforming encoders (DTEs) for all efficiently samplable distributions are infeasible. Even the computational relaxation of DTEs is infeasible assuming extractable one-way functions. Since all currently known constructions of honey encryption rely on DTEs, we conditionally refute the existence of honey encryption based on the DTE-then-encrypt framework from [30]. On the positive side, due to our feasibility result of
, computational honey encryption is feasible in the CRS model.
(informal). Assuming and a suitable symmetric-key encryption scheme (modeled as a random cipher), computational honey encryption for all efficiently samplable distributions exists in the CRS model.
1.3 Negative Results for Stronger Notions of Pseudorandom Encodings
Below we describe how we gradually relax optimal compression via different notions of pseudorandom encodings and derive infeasibility results for all variants of pseudorandom encodings which restrict the encoding algorithm to be deterministic or require an information-theoretic pseudorandomness guarantee. This leaves computational randomized pseudorandom encodings as the best possible achievable notion.
Deterministic, Statistical Pseudorandom Encodings. The notion of pseudorandom encodings with a deterministic encoding algorithm and information-theoretic indistinguishability is perhaps the simplest notion one can consider. As we will prove in this paper, this notion recovers the notion of optimal compression: since the encoding algorithm for some source is deterministic, it can be seen with an entropy argument that the output size of
must be at most
, the min-entropy of
; otherwise, the distribution
can necessarily be distinguished from random with some statistically non-negligible advantage. Therefore,
is an optimal and efficient compression algorithm for
, with decompression algorithm
; this is true even for the relaxation in the CRS model. The existence of efficient compression algorithms for various categories of samplers was thoroughly studied [40]. In particular, the existence of compression algorithms for all efficiently samplable sources implies the inexistence of one-way functions (this is an observation attributed to Levin in [23]) since compressing the output of a pseudorandom generator to its entropy would distinguish it from a random string, and the existence of one-way functions implies the existence of pseudorandom generators [24]).
(informal). Assuming the existence of one-way functions, neither nor
hold.
This is a strong impossibility result, as one-way functions dwell among the weakest assumptions in cryptography, [28]. One can circumvent this impossibility by studying whether compression can be achieved for more restricted classes of distributions, as was done e.g. in [40]. Our work can be seen as pursuing an orthogonal direction. We seek to determine whether a relaxed notion of compression can be achieved for all efficiently samplable distributions. The relaxations we consider comprise the possibility to use randomness in the encoding algorithm, and weakening the requirement on the encoded distribution to being only computationally indistinguishable from random. Clearly, these relaxations remove one of the most important features of compression algorithms, which is that their outputs are smaller than their inputs (i.e., they compress). Nevertheless, the indistinguishability of the encoded distribution from the uniform distribution is another crucial feature of optimal compression algorithms, which has independent applications.
Deterministic, Computational Pseudorandom Encodings. We now turn towards a relaxation where the encoded distribution is only required to be computationally indistinguishable from random, but the encoding algorithm is still required to be deterministic. This flavor is strongly connected to an important problem in cryptography: the problem of separating HILL entropy [24] from Yao entropy [44]. HILL and Yao entropy are different approaches of formalizing computational entropy, i.e., the amount of entropy a distribution appears to have from the viewpoint of a computationally bounded entity. Informally, a distribution has high HILL entropy if it is computationally close to a distribution with high min-entropy; a distribution has high Yao entropy if it cannot be compressed efficiently. Finding a distribution which, under standard cryptographic assumptions, has high Yao entropy, but low HILL entropy constitutes a long standing open problem in cryptography. Currently, only an oracle separation [42] and a separation for conditional distributions [27] are known. To establish the connection between and this problem, we proceed as follows: informally, a deterministic pseudorandom encoding must necessarily compress its input to the HILL entropy of the distribution. That is, the output size of the encoding cannot be much larger than the HILL entropy of the distribution. This, in turn, implies that a distribution which admits such a pseudorandom encoding cannot have high Yao entropy.
In this work, we formalize the above argument, and show that the conditional separation of HILL and Yao entropy from [27] suffices to refute . This separation holds under the assumption that non-interactive zero-knowledge proofs with some appropriate structural properties exist (which in turn can be based on standard assumptions such as the quadratic residuosity assumption). Thus, we obtain the following infeasibility result:
(informal). If the quadratic residuosity assumption holds, then does not hold.
Hence, we may conclude that towards a feasible variant of pseudorandom encodings for all efficiently samplable distributions, requiring the encoding algorithm to be deterministic poses a strong restriction.
Randomized, Statistical Pseudorandom Encodings. We now consider the relaxation of perfect compression by allowing the encoding algorithm to be randomized while still requiring information-theoretic indistinguishability from randomness. This flavor of pseudorandom encoding was used in the context of honey encryption [30]. Honey encryption is a cryptographic primitive which has been introduced to mitigate attacks on encryption schemes resulting from the use of low-entropy passwords. Honey encryption has the property that decrypting a ciphertext with an incorrect key always yields a valid-looking plaintext which seems to come from the expected distribution, thereby mitigating brute-force attacks. This is the same property that was useful in the previous PAKE example.
The study of honey encryption was initiated in [30], where it was shown that honey encryption can naturally be constructed by composing a block cipher (modeled as a random cipher) with a distribution transforming encoder (DTE), a notion which is equivalent to our notion of pseudorandom encoding with randomized encoding and statistical pseudorandomness. The focus of [30] was on obtaining such DTEs for simple and useful distributions. In contrast, we seek to understand the feasibility of this notion for arbitrary distributions. Intuitively, it is not straightforward to encode any efficient distribution into the uniform distribution; consider for example the distribution over RSA moduli, i.e., products of two random n-bit primes. Since no efficient algorithm is known to test membership in the support of this distribution, natural approaches seem to break down. In fact, we show in this work that this difficulty is inherent: building on techniques from [5, 29], we demonstrate the impossibility of (randomized, statistical) pseudorandom encodings for all efficiently samplable distributions, under a relatively standard cryptographic assumption.
(informal). Assuming the sub-exponential hardness of the learning with errors (LWE) problem, does not hold.
This result directly implies that under the same assumption, there exist efficiently samplable distributions (with input) for which no distribution transforming encoder exists. We view it as an interesting open problem whether this result can be extended to rule out the existence of honey encryption for arbitrary distributions under the same assumption.
1.4 Open Questions and Subsequent Work
The most intriguing question left open by our work is whether the weakest variant of the pseudorandom encoding hypothesis , which is implied by
, also implies
. Very recently, this question was settled in the affirmative by Wee and Wichs [43] under the LWE assumption. More concretely, by modifying a heuristic
construction of Brakerski et al. [7], they show that
is implied by LWE if one is additionally given an oblivious LWE-sampler in the CRS model. Such a sampler, given a matrix
, generates outputs that are indistinguishable from LWE samples
without knowing the secrets
or the noise
. The existence of an oblivious LWE sampler is nontrivial even under the LWE assumption, because
can be such that
is not pseudorandom; however, such a sampler still follows from the invertible sampling hypothesis [29], which we show to be equivalent to the pseudorandom encoding hypothesis. By proposing an explicit heuristic construction of (a relaxed flavor of) an oblivious LWE sampler, the end result of [43] is a construction of
from a new “falsifiable” assumption.
Whether implies
under weaker or different assumptions than LWE remains open. A potentially easier goal is using
to construct public-key encryption from one-way functions. This is related to the possibility of constructing oblivious transfer from any public-key encryption in which public keys and ciphertexts are obliviously samplable [19, 22], which is implied by public-key encryption and
. Here
is used to bypass the black-box separation between public-key encryption and oblivious transfer [22].
Finally, there is a lot of room for relaxing the intractability assumptions we use to rule out the statistical () and deterministic (
) flavors of pseudorandom encodings.
Organization. In Sect. 2, we provide a technical overview of a selection of our results. In Sect. 3, we provide condensed definitions of pseudorandom encodings and invertible sampling and a formal proof of their equivalence and in Sect. 4 we describe the static-to-adaptive transformation. We refer the reader to the full version [1] for more details and for the other results we described.
2 Overview of Techniques
In this section, we elaborate on some of our technical results in more detail. In the following, we identify a PPT sampler with the distribution (family) ensemble it induces.

















The concepts of inverse samplability and pseudorandom encodings are tightly connected. Suppose a PPT algorithm is inverse samplable. Then, there exists an alternative and an inverse sampler
satisfying closeness and invertibility. Invertibility guarantees that the inverse sampler
on input of a sample y from
, outputs a computationally well-distributed random tape r. Hence, with overwhelming probability over the choice of
and
, the alternative sampler on input of r, recovers y. In other words, the inverse sampler
can be seen as encoding a given sample y, whereas the de-randomized alternative sampler
given this encoding as random tape, is able to recover y. Looking through the lens of pseudorandom encoding, this almost proves correctness except that y is sampled according to
instead of
. This difference can be bridged due to closeness. We now turn towards showing pseudorandomness of the encoded distribution. Due to closeness, the distributions
and
are computationally indistinguishable. Invertibility guarantees that, given a sample y from
, an encoding of y is indistinguishable to uniformly chosen randomness conditioned on the fact that decoding yields y. Removing y from this distribution, almost corresponds to pseudorandomness, except that y is sampled according to
instead of
. Again, we are able to bridge this gap due to closeness. Note that we crucially use the fact that the initial randomness used by
resides outside of the view of an adversary. Summing up, if a PPT sampler
is inverse samplable, then it can be pseudorandomly encoded.






















Likewise to the variations and relaxations described for pseudorandom encodings, we vary and relax the notion of invertible sampling. The inverse sampler can be required to be deterministic or allowed to be randomized. Further, closeness and invertibility can be required to hold information theoretically or computationally. We denote these variants as and
. To circumvent impossibilities in the plain model, we also define the relaxations in the common reference string model in static and adaptive flavors, denoted the prefix “
” and “
”, respectively. The above equivalence extends to all introduced variations of the pseudorandom encoding and invertible sampling hypotheses.
The Static-to-Adaptive Transformation. The static variant of pseudorandom encodings in the CRS model only guarantees correctness and pseudorandomness as long as the input m for the sampler is chosen independently of the CRS. The adaptive variant, on the other hand, provides correctness and pseudorandomness even for adaptive choices of inputs. Adaptive notions always imply their static analogues. Interestingly, for pseudorandom encodings, the opposite direction is true as well. The core idea is to use an indirection argument (similar to [11, 12, 25]) to delay CRS generation until during the actual encoding process. Thus, the advantage stemming from adaptively choosing the input is eliminated.
Suppose that the static variant of the pseudorandom encoding hypothesis in the CRS model is true and let be some PPT sampler. Since
can be pseudorandomly encoded in the CRS model with static choice of inputs, there exist algorithms
such that static correctness and pseudorandomness hold. Further, the algorithm
can also be pseudorandomly encoded as above. Let
be the corresponding algorithms such that static correctness and pseudorandomness hold. Note that since the sampler
does not expect an input, static and adaptive guarantees are equivalent.
Then, the sampler can be pseudorandomly encoded in the CRS model with adaptive choice of inputs as follows. Initially, we sample a common reference string
via
and make it available to the parties. Given
and a sample y from
, adaptive encoding works in two phases. First, a fresh CRS
is sampled via
and pseudorandomly encoded via
. Second, the given sample y is pseudorandomly encoded via
. The encoding of y then consists of
. To decode, the CRS
is restored via
. Then, using
, the original sample y is recovered via
.
Since is chosen freshly during the encoding process, the input m which may depend on
, cannot depend on
. Further, the distribution
does not expect an input. Hence, static guarantees suffice.
To realize that adaptive pseudorandomness holds, consider the encoding of for some adaptively chosen message m. Since the view of
when choosing the message m is independent of
, static pseudorandomness can be applied to replace the distribution
with uniform randomness. Further, since the sampler
does not expect any input, static pseudorandomness suffices to replace the distribution
with uniform randomness. This proves adaptive pseudorandomness.
The adaptive variant of correctness follows similarly from the static variant of correctness. Consider the distribution of decoding an encoded sample of , where m is adaptively chosen. Since the sampler
does not expect an input, static correctness can be applied to replace decoding
with the
sampled during encoding. Again, since
does not lie in the view of the adversary when choosing the message m, static correctness guarantees that decoding succeeds with overwhelming probability. This proves adaptive correctness.
On Deterministic Pseudorandom Encoding and Compression. The notion of pseudorandom encoding is inspired by the notion of compression. A tuple of deterministic functions is said to compress a source
to length
with decoding error
, if (i)
and (ii)
, see [40, 42]. Pseudorandom encoding partially recovers the notion of compression if we require the encoding algorithm to be deterministic. If a source
can be pseudorandomly encoded with a deterministic encoding algorithm having output length
, then
is compressible to length
. Note, however, that the converse direction is not true. Compression and decompression algorithms for a compressible source do not necessarily encode that source pseudorandomly. The output of a compression algorithm is not required to look pseudorandom and, in some cases, admits a specific structure which makes it easily distinguishable from uniform randomness, e.g. instances using Levin search, [40].
Clearly, the requirement for correctness, poses a lower bound on the encoding length , [36]. Conversely, requiring the encoding algorithm
to be deterministic means that the only source of entropy in the distribution
originates from the source
itself. Hence, for the distributions
and the uniform distribution over
to be indistinguishable, the encoding length
must be “sufficiently small”. We observe that correctness together with the fact that
is deterministic implies that the event
occurs with overwhelming probability. Applying pseudorandomness yields that
holds with overwhelming probability, wherefore we can conclude that
operates almost injectively on the set
. Hence, the (smooth) min-entropy of
is at least
.
Considering information theoretical pseudorandomness, the distributions and
are statistically close. Hence, by the reasoning above, the encoding length
is upper bounded by the (smooth) min-entropy of the source
. In conclusion, if a distribution can be pseudorandomly encoded such that the encoding algorithm is deterministic satisfying statistical pseudorandomness, then this distribution is compressible to its (smooth) min-entropy. Using a technical “Splitting Lemma”, this extends to the relaxed variant of the pseudorandom encoding hypothesis in the CRS model.












Unfortunately, we do not know whether this extends to the relaxed variants of the pseudorandom encoding hypothesis admitting access to a CRS. On a high level, the problem is that the HILL entropy, in contrast to the min-entropy, does not remain untouched when additionally conditioning on some common reference string distribution, even though the initial distribution is independent of the CRS. Hence, the splitting technique can not be applied here.
3 Pseudorandom Encodings and Invertible Sampling
In this section, we formally define pseudorandom encodings and invertible sampling. We will work with the hypothesis that every efficiently samplable distribution can be pseudorandomly encoded and invertible sampled and we refer to these hypotheses as the pseudorandom encoding hypothesis and the invertible sampling hypothesis, respectively. This section is a condensed and much less detailed version of [1].







- Correctness.
For all inputs
,
is negligible.
- Pseudorandomness.
- For all PPT adversaries
and all inputs
,
whereand
are defined below.






- Closeness.
- For all PPT adversaries
and all inputs
,
whereand
are defined below.
- Invertibility.
- For all PPT adversaries
and all inputs
,
whereand
are defined below.

is true if and only if
is true.
If holds, then
holds.
Assume holds. Let
be a PPT algorithm.
implies that there exists an alternative sampler
(with randomness space
) and a corresponding inverse sampler
satisfying closeness and invertibility.
For , we define the algorithms
(potentially randomized) and
(deterministic).

Hybrids used in the proof of correctness.
Game is identical to
and game
is identical to
. Hence,
.
For all PPT adversaries , for all
, there exists a PPT adversary
, such that
.
Construct an adversary on closeness. On input of (m, y),
computes
, calls
on input of
and outputs the resulting output. If y is sampled using
(for
),
perfectly simulates game
for
. If y is sampled using
(for
),
perfectly simulates game
for
. Therefore,
and
.
Thus, we have that for some PPT adversaries
.
Consider the adversary distinguishing between game
and game
that on input of (m, r, y), outputs 0 if
and outputs 1 otherwise. By definition,
always outputs 0 in
. Hence,
.



Hybrids used in the proof of pseudorandomness.
For all PPT adversaries , for all
, there exists a PPT adversary
, such that
.
Construct a PPT adversary on the closeness property as follows. On input of (m, y),
calls
on input of
and outputs the resulting output.
If ,
simulates game
for
, and if
,
simulates game
for
. Hence,
and
.
For all PPT adversaries , for all
, there exists a PPT adversary
, such that
.
We construct a PPT adversary on the invertibility property. On input of (m, r, y),
calls
on input of (m, r) and outputs its output.
If for
,
simulates game
for
. If
,
simulates game
for
. Therefore,
and
.
Hence, for some PPT adversaries
and
.
If holds, then
holds.
We prove the statement for the computational randomized case. The remaining cases are similar.
Assume holds. Let
be a PPT algorithm.
implies that for
there exist efficient algorithms
(potentially randomized) with output length
and
(deterministic) satisfying correctness and pseudorandomness.
For , we define the alternative sampler as
(randomized) and the corresponding inverse sampler
(potentially randomized).

Hybrids used in the proof of closeness.
Closeness. Let be an adversary on closeness. We consider a sequence of games starting from
and concluding in
, see Fig. 4.
The difference between game and game
is only conceptional, hence,
.
and
proceed exactly identical if
. More formally, let F be the event that
. We have that
. Hence, the Difference Lemma (due to Shoup, [37]) bounds
. Correctness guarantees that for all
,
is negligible.
For all PPT adversaries , for all
, there exists a PPT adversary
, such that
.
Construct an adversary on pseudorandomness as follows. On input of
,
calls
on input
and outputs the resulting output. If
for
,
perfectly simulates game
for
. Otherwise, if u is uniformly random over
,
perfectly simulates game
for
. Hence,
and
.
Hence, for some PPT adversary
.

Hybrids used in the proof of invertibility.
For all PPT adversaries , for all
, there exists a PPT adversary
, such that
.
Let be an adversary distinguishing
and
. Construct an adversary
on the closeness property. On input of (m, y),
computes
and calls
on input
. If
,
simulates game
for
. Else, if
,
simulates game
for
. Hence,
.
The difference between and
is purely conceptional. Hence,
.
and
behave identical if
. Let F denote the failure event
. We have that
. The Difference Lemma (due to Shoup, [37]) bounds
. Due to correctness, for all
,
is negligible.
For all PPT adversaries , for all
, there exists a PPT adversary
, such that
.
Construct a PPT adversary on the pseudorandomness property. On input of (m, u),
calls
on input
and outputs the resulting output. If
for
,
perfectly simulates game
for
. Otherwise, if u is uniformly random over
,
perfectly simulates game
for
. Hence,
and
.
The difference between and
is again only conceptional and
. Hence,
for some PPT adversary
.
4 Static-to-Adaptive Transformation
We obtain a natural relaxation of the pseudorandom encoding hypothesis by introducing public parameters. That is, a distribution defined via can be pseudorandomly encoded in this relaxed sense, if there exists a probabilistic setup algorithm
and encode and decode algorithms as before such that for all
, the event
occurs with overwhelming probability, where the probability is also over the choice of
, and the distribution
is indistinguishable from the distribution
. See the full version [1] for more details.
There are two variants of this definition. The input m can be required to be chosen independently of or allowed to be chosen depending on
. Clearly, the adaptive variant implies the non-adaptive (or static) variant. Interestingly, the opposite direction is true as well by an “indirection” argument similar to the one from the work on universal samplers [25]. A similar technique was used in the context of non-committing encryption [11] and adaptively secure MPC [12].
Let and
. If
is true, then
is true.
We prove the statement for the computational randomized case. A very similar proof applies to the remaining cases.
Let be a PPT sampler with input space
. Since
is true, for the PPT sampler
, there exist
with output length
such that correctness and pseudorandomness hold (statically). Again, since
is true, for the PPT sampler
, there exist
with output length
such that correctness and pseudorandomness hold (statically).8 Note that
does not expect an input.
In Fig. 6, we define algorithms satisfying adaptive correctness and pseudorandomness.

Adaptive pseudorandom encodings.



Hybrid games for the proof of adaptive correctness.
Adaptive correctness. We define a series of hybrid games to prove correctness, see Fig. 7. Game corresponds to encoding and subsequently decoding a sample y (for adaptively chosen input m) and game
is simply a reordering of the commands of
. The game hop from
to
only conceptional and
.
For all PPT adversaries , there exists a PPT adversary
, such that
.






![$$\begin{aligned} \Pr \left[ \begin{array}{ccl} crs ''&{}\leftarrow &{}\mathsf{Setup}(1^{\lambda })\\ crs '&{}\leftarrow &{}\mathsf{Setup}_{S}'(1^{\lambda })\\ r_1&{}\leftarrow &{}\mathsf {E}_{}''( crs '', crs ')\\ crs '_D&{}:=&{}\mathsf {D}_{}''( crs '', r_1) \end{array} : crs '_D\ne crs ' \right] \end{aligned}$$](../images/508076_1_En_23_Chapter/508076_1_En_23_Chapter_TeX_Equ8.png)
![$$\begin{aligned} |*|{\Pr [out_{2}=1]-\Pr [out_{1}=1]}\le \Pr [E]\text {.} \end{aligned}$$](../images/508076_1_En_23_Chapter/508076_1_En_23_Chapter_TeX_Equ9.png)

The game hop from to
only conceptional and
.
For all PPT adversaries , there exists a PPT adversary
, such that
.


![$$\begin{aligned} \Pr \left[ \begin{array}{ccl} m&{}\leftarrow &{}\overline{\mathcal {A}}(1^{\lambda })\\ crs '&{}\leftarrow &{}\mathsf{Setup}_S'(1^{\lambda })\\ y&{}\leftarrow &{}S(m)\\ r&{}\leftarrow &{}\mathsf {E}_{S}'( crs ', m, y)\\ y_D&{}:=&{}\mathsf {D}_{S}'( crs ', m, r) \end{array} :y=y_D\right] \end{aligned}$$](../images/508076_1_En_23_Chapter/508076_1_En_23_Chapter_TeX_Equ10.png)

![$$\Pr [out_{3}=1]$$](../images/508076_1_En_23_Chapter/508076_1_En_23_Chapter_TeX_IEq523.png)


Hybrid games for the proof of adaptive pseudorandomness.
Game corresponds to the adaptive pseudorandomness game. That is,
first samples
, the adversary
chooses the message m adaptively depending on
, and
then samples y using
, encodes that sample and gives the encoding to
.
For all PPT adversaries , there exists a PPT adversary
, such that
.
Construct an adversary on static pseudorandomness relative to
as follows. On input of
,
samples
calls
on input of
, and outputs the message m produced by
. In return,
receives
and either
or a uniform random string
from
.
computes
, calls
on input of
and returns
’s output.
If plays
, then it perfectly simulates
. On the other hand, if
plays
, then it perfectly simulates
.
The game hop from to
is only conceptional and
.
For all PPT adversaries , there exists a PPT adversary
, such that
.
Construct an adversary on static pseudorandomness relative to
as follows. On input of
,
returns
since the input space
of the sampler
is empty. In return,
receives
sampled via
and u which is either produced via
or via uniform sampling from
.
calls
on input of
and receives a message m from
. Finally,
samples
, calls
on input of
and outputs his output.
If plays
, then it perfectly simulates
. On the other hand, if
plays
, then it perfectly simulates
.
We thank Daniel Wichs for suggesting the unconditional static-to-adaptive transformation, as well as anonymous reviewers for extremely useful feedback.