Persuasive Privacy

University of Melbourne, 23rd March 2026

Joshua J Bon

Adelaide University

My research

G SBI SBI SMC SMC SBI--SMC Privacy Privacy SMC--Privacy Privacy--SBI

My research

Statistics

  • Bayesian modelling1 2
  • Efficient computation3 4 5
  • Approximate models3 4

G SBI SBI SMC SMC SBI--SMC

My research

Privacy

  • Federated & transfer learning1 2
  • Private (Bayesian) computation3
  • Approximate models4

G Privacy Privacy SMC SMC Privacy--SMC SBI SBI Privacy--SBI

My research

G SBI SBI SMC SMC SBI--SMC Privacy Privacy SMC--Privacy Privacy--SBI

Today’s talk

G SBI SBI SMC SMC SBI--SMC Privacy Privacy SMC--Privacy Privacy--SBI

Today’s talk1

  1. Differential privacy
  2. Privacy from the ground up
  3. Persuasive privacy
  4. Differential privacy redux
  5. Privacy without noise

Bon, Bailie, Rousseau & Robert (2026) arXiv 2601.22945

Under review at ICML 2026

Differential privacy

DP framework

\text{Dataset}~x \overset{\text{release}}{\longrightarrow} \text{statistic}~T

  • T \sim M(x,\cdot), with some mechanism M

Example

  • T = \frac{1}{n}\sum_{i=1}^{n} x_i + Z with Z \sim \mathcal{N}(0,\sigma^2)

M(x,\cdot) = \mathcal{N}(\overline{x},\sigma^2)

DP Intuition

To limit information leaked about data x by statistic T,

control the sensitivity of the mechanism M(x,\cdot) w.r.t. x.

DP Definition

Borrow from Lipschitz continuity (Bailie and Gong 2024).

d_\text{P}[M(x,\cdot),M(x^\prime,\cdot)] \leq \epsilon~d_\text{X}[x,x^\prime]

for all x,x^\prime \in \mathcal{D}.

DP Definition

A mechanism M is \epsilon-DP if

\sup_{S\in\mathcal{S}}\left\vert\ln\frac{M(x,S)}{M(x^\prime,S)}\right\vert \leq \epsilon

for all x,x^\prime \in \mathcal{D}, differing by at most one element.

Pure Differential Privacy

M(x,S) \leq \exp\{\epsilon\} M(x^\prime, S) for all x \sim x^\prime and S \in \mathcal{S}.

Some DP Properties

Focusing on (pure) \epsilon-DP.

  1. Composition: If M_1 and M_2 are \epsilon-DP then

M_1 \otimes M_2 ~\text{is}~ 2\epsilon\text{-DP}

  1. Post-processing: If K is independent of the data then

M_1 K ~\text{is}~ \epsilon\text{-DP}

Some DP Properties

Focusing on \epsilon-DP. Examples:

  1. Composition:

If releasing the mean and variance of a dataset is \epsilon-DP respectively, then the joint release is 2\epsilon-DP.

  1. Post-processing:

If M constructs a histogram that is \epsilon-DP, then any quantile from the histogram will also be \epsilon-DP.

Variants of DP

There are many variants and relaxations of DP.

\inf_{x\sim x^\prime}\mathbb{P}_{x}\left[ m(x,T) \leq \exp\{\epsilon\} m(x^\prime, T) \right] \geq 1 - \delta

m(x,S) \leq \exp\{\epsilon\} m(x^\prime, S) + \delta for all S \in \mathcal{S}.

Variants of DP

There are many variants and relaxations of DP.

  • Rényi DP
  • Gaussian DP
  • Concentrated DP
  • Many more!

All related to Lipschitz interpretation

Pure DP Semantics

Limitations of DP

  • DP Semantics are post-hoc explanations
  • Deterministic mechanisms are not covered by DP
    • What about the mean of a large dataset?

Our approach is a semantics-first understanding of data privacy, where

  • Assumptions can be tested against real-world
  • Privacy definitions are easy to interpret, communicate and tailor, because
  • Our framework is constructed from an agent-based game

Privacy from the ground up

Overview of statistical data privacy

Sender

Government agency release small area statistics on a disease

Receiver

Insurance company raises premiums using inferred disease prevalence in a small town

Effect on data privacy

Statistic release \longrightarrow adversarial decisions \longrightarrow privacy effect

Anatomy of statistical data privacy

The players

Sender

  • Custodian of data
  • Will use mechanism to release information about data

Receiver

  • Makes decision based on revealed information
  • Decision affects Sender’s privacy

Privacy function

Sender’s privacy measured with a privacy function

\rho:\mathcal{D} \times \mathsf{X} \rightarrow \mathbb{R}

  • Receiver’s decision \quad d_i \in \mathcal{D}

  • Data value \quad x \in \mathsf{X}

\rho orders preferences of decisions for a given dataset

  • If d_1 is preferred to d_2 then \rho(d_1,x) > \rho(d_2,x)

Example 1: Privacy function

For some fixed \kappa>0,

\rho(d,x) = \begin{cases} 0, & \text{if } x \in d, \vert d \vert < \kappa\\ 1, & \text{otherwise}, \end{cases}

  • x \in \mathsf{X} = \mathbb{R}

  • d \in \mathcal{D} = \{[a,b]:a,b\in \mathbb{R}, a\leq b\}

Example 2: Privacy function


\rho(d,x) = -\log d(x)

  • x \in \mathsf{X} = \mathbb{R}

  • d \in \mathcal{D} = \{\text{probability density functions on } \mathbb{R}\}

The statistic

A statistic T \in \mathsf{T} is output from a mechanism M given x

T \sim M(x,\cdot)

  • M:\mathcal{T} \times \mathsf{X} \rightarrow [0,1] for (\mathsf{T},\mathcal{T}) measurable space

The statistic

M can be deterministic or randomised

  • Summary statistics \quad\tau:\mathsf{X}\rightarrow\mathsf{T}
    • M(x,\cdot) = \delta_{\tau(x)}(\cdot)
  • Noisy summary statistics
    • M(x,\cdot) = \mathcal{N}(~\cdot \mid \tau(x), \sigma^2)
  • Composition of multiple mechanisms

The statistic

M can be deterministic or randomised

  • Result of optimisation
    • MLE: \quad M(x,\cdot) = \delta_{\tau(x)}(\cdot)
    • \tau(x) = \arg\max_{\theta \in \Theta} \log L(x \mid \theta)
  • Monte Carlo output
    • MCMC: \quad M(x,\mathrm{d}t_{1:N}) = \prod_{i=1}^N K(\mathrm{d}t_{i} \mid t_{i-1}, x)
  • Noisy optimisation
    • SGD: \quad M(x,\mathrm{d}t_N) = \int_{\mathsf{T}_{-n}}\prod_{i=1}^N K(\mathrm{d}t_{i} \mid t_{i-1}, x)

Transparency

Privacy class: A set of mechanisms that satisfy a given privacy definition.

If \mathfrak{C} is the privacy class generated by “\text{D}” then

M \in \mathfrak{C} \Longleftrightarrow M~\text{satisfies}~\text{D}

Assumption 1: Transparency

Sender shares the mechanism M and privacy class \mathfrak{C}, for which M \in \mathfrak{C}, with Receiver. Further, the definitions of M and \mathfrak{C} do not depend on the data.

Receiver

Assumption 2: Bayesian adversary

Receiver makes Bayesian decisions

  1. Holds a prior distribution over the data Q
  2. Constructs a posterior distribution Q_T from (M,T)
  3. Has loss function \ell: \mathcal{D} \times \mathsf{X} \rightarrow \mathbb{R}

\mathfrak{C} may affect Receiver’s data posterior. Here assumed not to.

Anatomy of statistical data privacy

Receiver’s optimal decision

From Assumption 1:

Receiver’s optimal decision

d^{P} \in \arg\inf_{d \in \mathcal{D}} \mathbb{E}_{z \sim Q_T}[ \ell(d,z)]

To model Receiver’s decision:

We need assumptions on Q and \ell

Receiver model

Assumption 3: Adversarial loss function

Receiver has loss function \ell(d,x) = \rho(d,x)

  • Interpretation 1: Receiver targets privacy

If d^P_{\ell^\prime} is Receiver’s optimal decision under (P,\ell^\prime) then

\mathbb{E}_{x \sim P}[ \rho(d^{P}_\rho,x)] \leq \mathbb{E}_{x \sim P}[ \rho(d^P_{\ell^\prime},x)]

  • Interpretation 2: data-averaged worst-case for Sender

Receiver model

Assumption 4: Adversarial prior class

Let the data-prior Q \in \mathcal{Q}_x

Implies privacy outcome in terms of

Receiver’s decision

d^{Q_T} \in \arg\inf_{d \in \mathcal{D}} \mathbb{E}_{z \sim Q_T}[ \rho(d,z)]

Privacy outcome

\inf_{Q \in \mathcal{Q}_x} \rho(d^{Q_T},x)

Anatomy of statistical data privacy

Anatomy of statistical data privacy

Measuring privacy

Privacy outcome

\inf_{Q \in \mathcal{Q}_x} \rho(d^{Q_T},x)

  • Let S: \mathcal{P} \times \mathsf{X} \rightarrow \mathbb{R} such that S(Q_T,x) = \rho(d^{Q_T},x)
  • S is a (negatively-orientated) proper scoring rule1
    • A “privacy score”

Measuring privacy

Privacy outcome

\inf_{Q \in \mathcal{Q}_x} S(Q_T,x)

  • Use known proper scoring rules to measure privacy outcome
  • Create new proper scoring rules from \rho
  • For example when d is a PDF/PMF:

\rho(d,x) = -\log d(x) then S(Q_T,x)=-\log q_T(x).

Without loss of generality, we focus on proper scoring rules.

Measuring privacy

Privacy outcome

\inf_{Q \in \mathcal{Q}_x} S(Q_T,x)

  • This is an absolute measure of privacy
  • We will look at a relative measure of privacy

Measuring privacy

Relative privacy outcome

\inf_{Q \in \mathcal{Q}_x} \left[ S(Q_T,x) - S(Q,x) \right]

  • Considers Receiver’s information gain
  • Sender’s relative privacy change (worst-case)

Measuring privacy

Relative privacy outcome

\inf_{Q \in \mathcal{Q}_x} \left[ S(Q_T,x) - S(Q,x) \right]

But…

  • Q_T may be random since T\sim M(x, \cdot)

Persuasive privacy

A privacy definition

  • Let \mathcal{Q}_x \subset \mathscr{P}(\mathsf{X},\mathcal{X})

  • M: \mathcal{T}\times \mathsf{X} \rightarrow \mathbb{R}_+ be a mechanism,

  • S be a privacy score, constants \kappa \geq 0, 0\leq\delta \ll 1.

Definition: Persuasive Privacy

We say M is (\mathcal{Q}_x, S, \kappa, \delta)-PP if

\inf_{x\in\mathsf{X}}\inf_{Q\in\mathcal{Q}_x}\mathbb{P}_x\left[S(Q, x) - S(Q_{T}, x) \leq \kappa \right] \geq 1 - \delta,

where \mathbb{P}_x is w.r.t. T \sim M(x,\cdot).

A privacy definition

Definition: Persuasive Privacy

We say M is (\mathcal{Q}_x, S, \kappa, \delta)-PP if

\inf_{x\in\mathsf{X}}\inf_{Q\in\mathcal{Q}_x}\mathbb{P}_x\left[S(Q, x) - S(Q_{T}, x) \leq \kappa \right] \geq 1 - \delta,

where \mathbb{P}_x is w.r.t. T \sim M(x,\cdot).

  • Change in privacy: S(Q, x) - S(Q_{T}, x)
  • Maximum allowable privacy change: \kappa
  • Maximum probability of leakage: \delta\quad

Why “persuasive”?

Sender chooses a mechanism to persuade Receiver to make decisions having limited impact on privacy.

Similar to Bayesian persuasion (Kamenica and Gentzkow 2011) However,

  • asymmetry between Sender and Receiver
  • utility functions of Sender and Receiver are related
  • Sender assesses decisions (mechanism) robustly: worst-case not expected value

Properties of Persuasive Privacy

  1. Composition

  2. Receiver post-processing

Definition: Receiver Post-Processing

A guarantee \mathrm{D} satisfies the receiver post-processing property if M \in \mathfrak{C}(\mathrm{D}) implies that M\otimes K \in \mathfrak{C}(\mathrm{D}) for all Markov kernels K independent of the data x.

  1. Do not satisfy (Sender) post-processing

Returning to DP

Recall the Persuasive Privacy elements (\mathcal{Q}_x, S, \kappa, \delta), and consider

  1. Family of alternative hypothesis priors

\mathcal{H}_x = \{ Q \in \mathcal P_2:x^\prime \sim x, Q(\{ x , x^\prime \})= 1 \}

  1. Log-probability scoring rule

S(P,x) = -\log p(x)

Returning to DP

Alternative hypothesis prior class and log-probability score recover PrDP as instance of Persuasive Privacy

\inf_{x\sim x^\prime}\mathbb{P}_{x}\left[ m(x,T) \leq \exp\{\epsilon\} m(x^\prime, T) \right] \geq 1 - \delta

  • Pure DP is a special case (\delta = 0)
  • PrDP implies Approximate DP

Returning to DP

  • Pure, Probabilistic, and approximate DP are notions of privacy relative to change in Receiver’s knowledge
  • Can be derived by considering persuasive privacy game
  • ADP satisfies (Sender) post-processing property
    • PrDP does not (unless pure DP)
    • Typically a criticism of PrDP
  • PrDP does have (Receiver) post-processing property
    • A simple fix: control Receiver’s decision by including pre-processed statistic

Privacy for deterministic mechanisms

  • DP can’t be defined for deterministic mechanisms
  • Persuasive privacy framework flexible enough
  • Demonstrate privacy for mean \bar x

M(x,\cdot) = \delta_{\bar x}(\cdot)

Privacy for deterministic mechanisms

Use class of Gaussian distributions

\mathcal{G}_{x}^{r} = \left\{ \mathcal{N}(\mu,\Sigma): \frac{(\bar{x}-\bar{\mu})^2}{\overline{\Sigma}} \leq r_1 , c_\Phi \leq r_2 \left(1- \frac{\sigma_i^2}{\Vert\sigma \Vert_2^2} \right) \right\}

for r_1 > 0 and r_2 > 1.

  • Data posterior after observing \bar x is multivariate normal but degenerate
  • Support only on subspace \{z\in \mathbb{R}^n: \bar z = \bar x\}

Privacy for deterministic mechanisms

Use (marginal) Dawid–Sebastiani Score

D_i(Q,x) = \log \sigma^2_i(Q) + \frac{[x_i - \mu_i(Q)]^2}{\sigma^{2}_i(Q)}

for each element of data, and consider the worst-case element for privacy.

Privacy for deterministic mechanisms

Proposition

The average mechanism M(x,\cdot) = \delta_{\bar{x}} satisfies (\mathcal{I},\mathcal{G}_{x}^{r},r_1+ \log r_2,0)-PP.

  • Possible to have a rigorous guarantee for a deterministic mechanism
  • Opens path for privacy guarantees for convergent iterative algorithms
    • e.g. MCMC without privacy perturbations (posterior summary statistics)
    • A “decomposition rule”

Summary

  • Framework for purpose-driven privacy definitions (rigorously justified)
    • All assumptions stated, anatomy used as backbone
  • Assessment of existing privacy guarantees with game theory
  • New interpretations of post-processing property
    • Easy fix for PrDP post-processing
  • Established that privacy guarantees possible for deterministic algorithms

Ongoing work

Short term

  • Decomposition rule (for deterministic mechanisms)

    • Use noise in MCMC as additional privacy
  • Nonparametric prior families

  • Independent Commissioner Against Corruption (SA)

Medium term

  • Privacy-utility trade-off

  • Computation (SMC, SBI) versus privacy approximation trade-off

Thank you!

References

Bailie, James, and Ruobin Gong. 2024. “General Inferential Limits Under Differential and Pufferfish Privacy.” International Journal of Approximate Reasoning 172: 109242.
Bon, Joshua J, Bernard Baffour, Melanie Spallek, and Michele Haynes. 2020. “Analysing Sensitive Data from Dynamically-Generated Overlapping Contingency Tables.” Journal of Official Statistics 36 (2): 275–96.
Bon, Joshua J, Timothy Ballard, and Bernard Baffour. 2019. “Polling Bias and Undecided Voter Allocations: US Presidential Elections, 2004–2016.” Journal of the Royal Statistical Society: Series A (Statistics in Society) 182 (2): 467–93.
Bon, Joshua J, and Anthony Lee. 2025. “Knots and Variance Ordering of Sequential Monte Carlo Algorithms.”
Bon, Joshua J, Anthony Lee, and Christopher Drovandi. 2021. “Accelerating Sequential Monte Carlo with Surrogate Likelihoods.” Statistics and Computing 31 (5): 1–26.
Bon, Joshua J, J Rousseau, and Christian P Robert. 2026. “Persuasive Privacy.”
Bon, Joshua J, David J Warne, David J Nott, and Christopher Drovandi. 2025. “Bayesian Score Calibration for Approximate Models.” Journal of Machine Learning Research 26 (301): 1–40. http://jmlr.org/papers/v26/24-1179.html.
Bretherton, Adam, Joshua J Bon, David J Warne, Kerrie Mengersen, and Christopher Drovandi. 2026. “A Principled Approach to Bayesian Transfer Learning.” Bayesian Analysis.
Dwork, Cynthia, Krishnaram Kenthapadi, Frank McSherry, Ilya Mironov, and Moni Naor. 2006. “Our Data, Ourselves: Privacy via Distributed Noise Generation.” In Advances in Cryptology - EUROCRYPT 2006, edited by Serge Vaudenay, 486–503. Berlin, Heidelberg: Springer Berlin Heidelberg.
Dwork, Cynthia, Frank McSherry, Kobbi Nissim, and Adam Smith. 2006. “Calibrating Noise to Sensitivity in Private Data Analysis.” In Theory of Cryptography Conference, 265–84. Springer.
Grünwald, Peter D, and A Philip Dawid. 2004. “Game Theory, Maximum Entropy, Minimum Discrepancy and Robust Bayesian Decision Theory.” Annals of Statistics 32 (4): 1367–1433.
Guingona, Vincent, Alexei Kolesnikov, Julianne Nierwinski, and Avery Schweitzer. 2023. “Comparing Approximate and Probabilistic Differential Privacy Parameters.” Information Processing Letters 182: 106380.
Hassan, Conor, Joshua J Bon, Elizaveta Semenova, Antonietta Mira, and Kerrie Mengersen. 2024. “Federated Learning for Non-Factorizable Models Using Deep Generative Prior Approximations.”
Kamenica, Emir, and Matthew Gentzkow. 2011. “Bayesian Persuasion.” American Economic Review 101 (6): 2590–2615.
Kasiviswanathan, Shiva P, and Adam Smith. 2014. “On the ’Semantics’ of Differential Privacy: A Bayesian Formulation.” Journal of Privacy and Confidentiality 6 (1).
O’Flaherty, Martin, Sara Kalucza, and Joshua J Bon. 2023. “Does Anyone Suffer from Teenage Motherhood? Mental Health Effects of Teen Motherhood in the UK Are Small and Homogeneous.” Demography 60 (3): 707–29.
Wasserman, Larry, and Shuheng Zhou. 2010. “A Statistical Framework for Differential Privacy.” Journal of the American Statistical Association 105 (489): 375–89.
Zhang, Lefeng, Tianqing Zhu, Ping Xiong, Wanlei Zhou, and Philip S Yu. 2021. “More Than Privacy: Adopting Differential Privacy in Game-Theoretic Mechanism Design.” ACM Computing Surveys (CSUR) 54 (7): 1–37.