Introduction Markov random field Let $G=(V,E)$ be a graph with $p$ nodes and $C{t}(G)$ be the set of cliques of size at most $t$ in $G$. A Markov random field with alphabet size $k$ and $t$-order interactions is a dis...
2024. 1. 18.6 min read
Introduction
Markov random field
Let G=(V,E) be a graph with p nodes and Ct(G) be the set of cliques of size at most t in G. A Markov random field with alphabet size k and t-order interactions is a distribution D over [k]p such that
p(x)∝expc∈Ct(G)∑ψc(xc).(1)
where xc is the subset of x that belongs to clique c. The equation (1) is also known as Hammersley-Clifford theorem.
t=2 : MRF is pairwise.
k=2 : binary MRF
k=t=2 : Ising model (binary and pairwise)
It is easy to think that for pairwise MRF, ψc(xc) is given as wixixj for xc={xi,xj} since the maximal clique size is 2.
Learning problem
Structure learning : Recover the set of non-zero edges in G (learning adjacency).
Parameter learning : Learn the structure and ψI for all cliques I of size at most t.
Preliminaries
Markov random field
Ising model
The p-variable Ising model is a distribution D(A,θ) on {−1,1}p that satisfies
p(x)∝exp1≤i≤j≤p∑Ai,jxixj+i∈[p]∑θixi,
where A is symmetric weight matrix with Aii=0 for i=1,…,p and θ∈Rp is a mean-field vector. The corresponding graph is MRF G=(V,E).
The width of D(A,θ) is defined as
λ(A,θ)=i∈[p]maxj∈[p]∑∣Ai,j∣+∣θi∣
The minimum edge weight is defined as
η(A,θ)=i,j∈[p],Ai,j=0min∣Ai,j∣
Pairwise graphical model
For a set of weight matrices W, the p-variable pairwise graphical model is a distribution D(W,θ) on [k]p that satisfies
Given samples X1,…,Xn∼D, the parameter learning algorithm outputs a matrix A^ such that
i,j∈[p]max∣Ai,j−A^i,j∣≤α.
Privacy
Concentrated Differential Privacy(zCDP)
A randomized algorithm A:Xn→S satisfies ρ-zCDP if for every neighboring datasets X,X′∈Xn,
∀α∈(1,∞)Dα(M(X)∥M(X′))≤ρα
where Dα is the α-Rényi divergence.
The Rényi divergence of order α (or called α-divergence) is defined as below.
Dα(P∥Q)=α−11log(i=1∑nqiα−1piα)
where P,Q are discrete probability distribution with pi=Pr(X=xi),X∼P.
(Bun & Steinke, 2016)
If A satisfies ϵ-DP, then A is 2ϵ2-zCDP.
If A satisfies 2ϵ2-zCDP, then A satisfies (2ϵ2+ϵ2log(δ1),δ)-DP for every δ>0.
(Dwork et al., 2010)
For a sequence of DP algorithms A1,⋯,AT and A=AT∘⋯∘A1, the following hold.
If A1,⋯,AT are (ϵ0,δ1),⋯,(ϵ0,δT)-DP respectively, then for every δ0>0, A is (ϵ,δ)-DP for ϵ=ϵ06Tlog(δ01) and δ=δ0+∑tδt.
If A1,⋯,AT are ρ1,⋯,ρT−zCDP respectively, then A is ρ-zCDP for ρ=∑tρt.
Parameter Learning of Pairwise MRF
Private sparse logistic regression
Consider training data D={d1,…dn}∼iidP where di=(xi,yi),∥xi∥∞≤1,yi∈{−1,1}. To minimize population logistic loss E[log(1+e−Y⟨w,X⟩], consider the following empirical risk minimization.
w=wargminL(w:D)=n1j∑log(1+e−yj⟨w,xj⟩)
Then, the following algorithm satisfies ρ−zCDP. At spatial statistics this model is so-called as auto-logistic model.
Algorithm
Algorithm Private Frank-Wolfe algorithmInput: D,L,convex set C={w∈Rp:∥w∥1≤λ}For t=1 to T−1 do:∀s∈S,αs←⟨s,∇L(w;D)⟩+Laplace(0,nρL1∥C∥1T)w~t←s∈Sargminαswt+1=(1−μt)wt+μtw~twhere μt=t+22end forOutput: wpriv=wT
Theorem
For wpriv from algorithm above, with probability at least 1−β the following holds.
The implementation algorithm is based on the private FW algorithm, and it satisfies ρ-zCDP.
Structure Learning of Graphical Models
Theorem (Wu et al., 2019)
There exists an algorithm that learns the structure of a pairwise graphical model w.p. at least 32. It requires n=O(η4λ2k4e14λlog(pk)) samples.
From the theorem above, for (ϵ,δ)-DP structure learning algorithm the sample size is given as follows.
n=O(ϵη4λ2k4log(pk)log(δ1))
Lower Bounds
Theorem
Any ϵ-DP structure learning algorithm of an Ising model with minimum edge weight η requires
n=Ω(ηϵp+ϵp)
samples. Also, for ρ-zCDP algorithm n=Ω(ρp) samples are required.
For p-variable pairwise MRF, ϵ-DP algorithm requires n=Ω(ηϵp+ϵk2p) samples and ρ-zCDP algorithm requires n=Ω(ρk2p) samples.
References
C. Dwork, G. N. Rothblum, and S. Vadhan, “Boosting and Differential Privacy,” in 2010 IEEE 51st Annual Symposium on Foundations of Computer Science, Las Vegas, NV, USA: IEEE, Oct. 2010, pp. 51–60. doi: 10.1109/FOCS.2010.12.
H. Zhang, G. Kamath, J. Kulkarni, and S. Wu, “Privately learning Markov random fields,” in International conference on machine learning, PMLR, 2020, pp. 11129–11140. Accessed: Jan. 17, 2024. [Online]. Available: http://proceedings.mlr.press/v119/zhang20l.html
S. Wu, S. Sanghavi, and A. G. Dimakis, “Sparse Logistic Regression Learns All Discrete Pairwise Graphical Models.” arXiv, Jun. 18, 2019. doi: 10.48550/arXiv.1810.11905.