当前位置:高等教育资讯网  >  中国高校课件下载中心  >  大学文库  >  浏览文档

《计算机科学》相关教学资源(参考文献)Local Distributed Sampling from Locally-Defined Distribution

资源类别:文库,文档格式:PDF,文档页数:38,文件大小:9.4MB,团购合买
点击下载完整版文档(PDF)

Local Distributed Sampling from Locally-Defined Distribution Yitong Yin Nanjing University

Local Distributed Sampling !om Locally-Defined Distribution Yitong Yin Nanjing University

Counting and Sampling RANDOM GENERATION OF COMBINATORIAL STRUCTURES FROM A UNIFORM DISTRIBUTION LGORITHMS Mark R.JERRUM Department f Computer Science,Universiry of Edinburgh,Edinburgh EH93 Unied Kingdom Leslie G.VALIANT* Aiken Computation Laboratory,Harard Universiry,Cambridge,MA 02138,U.S.A V灯jayV.VAZIRANI* Computer Scence Department,Comell Universiry,Ithac,NY 14853,U.S.A. [Jerrum-Valiant-Vazirani86]: (For self-reducible problems) approx.counting (approx.,exact)sampling is tractable is tractable

Counting and Sampling approx. counting is tractable (approx., exact) sampling is tractable (For self-reducible problems) [Jerrum-Valiant-Vazirani ’86]:

Computational Phase Transition Sampling almost-uniform independent set in graphs with maximum degree△: ●[Weitz2006]:lf△≤5,poly-time. [Sly 2010]:If A=6,no poly-time algorithm unless NP=RP. A phase transition occurs when△:5→6. Local Computation?

Computational Phase Transition • [Weitz 2006]: If ∆≤5, poly-time. • [Sly 2010]: If ∆≥6, no poly-time algorithm unless NP=RP. Sampling almost-uniform independent set in graphs with maximum degree ∆: A phase transition occurs when ∆: 5→6. Local Computation?

Local Computation "What can be computed locally?"[Naor,Stockmeyer'93] the LOCAC model [Linial '87]: Communications are synchronized. In each round:each node can exchange unbounded messages with all neighbors,perform unbounded local computation,and read/write to unbounded local memory. Complexity:of rounds to terminate in the worst case. In t rounds:each node can collect information up to distance t. PLOCAL:t=polylog(n)

Local Computation • Communications are synchronized. • In each round: each node can exchange unbounded messages with all neighbors, perform unbounded local computation, and read/write to unbounded local memory. • Complexity: # of rounds to terminate in the worst case. • In t rounds: each node can collect information up to distance t. the LOCAL model [Linial ’87]: “What can be computed locally?” [Naor, Stockmeyer ’93] PLOCAL: t = polylog(n)

A Motivation: Distributed Machine Learning ●Data are stored in a distributed system. Distributed algorithms for: sampling from a joint distribution(specified by a probabilistic graphical model); ● inferring according to a probabilistic graphical model

A Motivation: Distributed Machine Learning • Data are stored in a distributed system. • Distributed algorithms for: • sampling from a joint distribution (specified by a probabilistic graphical model); • inferring according to a probabilistic graphical model

Example:Sample Independent Set u:uniform distribution of independent sets in G. Ye0,1 indicates an independent set Each ve∈Vreturns a Y∈{0,l}, such that Y=(YvEr -u Or:drv(Y,u)<1/poly(n) network G(E)

Example: Sample Independent Set • Each v∈V returns a Yv∈ {0,1}, such that Y = (Yv)v∈V ∼ µ • Or: dTV(Y, µ) < 1/poly(n) µ: uniform distribution of independent sets in G. network G(V,E) Y ∈ {0,1}V indicates an independent set

Inference (Local Counting) w:uniform distribution of independent sets in G. g:marginal distribution at v conditioning on o e0,1)5. y∈{0,1}:g(y)=Pr[Y=y|Ys=o] ●Each y∈S receives ov as input. ● Each v E D returns a marginal distributionsuch that: drv(g,hg)≤poim Z=40-ΠyX.=01<i:X,=0 i=1 network G(E) Z:of independent sets

Inference (Local Counting) network G(V,E) µ: uniform distribution of independent sets in G. • Each v ∈ S receives σv as input. • Each v ∈ V returns a marginal distribution such that: µ ˆ￾ v dTV(ˆµ￾ v , µ￾ v )  1 poly(n) : marginal distribution at v conditioning on σ ∈{0,1}S µ . ￾ v 0 1 1 0 8y 2 {0, 1} : µ￾ v (y) = Pr Y ⇠µ [Yv = y | YS = ￾] 1 Z = µ(;) = Y n i=1 Pr Y ⇠µ [Yvi = 0 | 8j<i : Yvj = 0] Z: # of independent sets

Decay of Correlation g:marginal distribution at v conditioning on o0,1)5. strong spatial mixing(SSM): V boundary condition Be∈{0,l}r-sphere(W: drv(g,ug,B)≤poly(n)·exp(-2(r) SSM(if△≤5 when u is uniform distribution of ind.sets) approx.inference is solvable in O(log n)rounds in the OCA model

Decay of Correlation strong spatial mixing (SSM): SSM approx. inference is solvable in O(log n) rounds in the LOCAL model G v r B σ : marginal distribution at v conditioning on σ ∈{0,1}S µ . ￾ v ∀ boundary condition B∈{0,1}r-sphere(v) : dTV(µ￾ v , µ￾,B v )  poly(n) · exp(￾⌦(r)) (iff ∆≤5 when µ is uniform distribution of ind. sets)

Gibbs Distribution (with pairwise interactions) Each vertex corresponds to a variable with finite domain g. network G,E): ● Each edge e=(u,v)EE has a matrix (binary constraint): Ae:[q]×[q→[0,l] bN ●Each vertex ve∈has a vector (unary constraint): bm:[g小→[0,1] Gibbs distribution u:Voelg] u(o)xIAe(o,o)Ib(ou) e=(u,v)∈E u∈V

Gibbs Distribution network G(V,E): • Each vertex corresponds to a variable with finite domain [q]. • Each edge e=(u,v)∈E has a matrix (binary constraint): • Each vertex v∈V has a vector (unary constraint): µ(￾) / Y e=(u,v)2E Ae(￾u, ￾v) Y v2V bv(￾v) Ae u bv v (with pairwise interactions) Ae: [q] × [q] → [0,1] bv: [q] → [0,1] • Gibbs distribution µ : ∀σ∈[q]V

Gibbs Distribution (with pairwise interactions) Gibbs distribution u:o∈[g]' network G,E): L(a)Ae(ou;ou)bo() e=(u,v)∈E v∈V independent set: bv a-B司6-图 coloring: 10 1 1 Ae= bu= .H Ae:[q]×[q]→[0,1] 1 bm:[q]→[0,1]

Gibbs Distribution • Gibbs distribution µ : ∀σ∈[q]V µ(￾) / Y e=(u,v)2E Ae(￾u, ￾v) Y v2V bv(￾v) • independent set: bv =  1 1 ￾ Ae =  1 1 1 0￾ • coloring: network G(V,E): Ae u bv v Ae: [q] × [q] → [0,1] bv: [q] → [0,1] (with pairwise interactions) Ae = 2 6 6 6 4 0 0 ... 0 3 7 7 7 5 1 1 bv = 2 6 4 1 . . . 1 3 7 5

点击下载完整版文档(PDF)VIP每日下载上限内不扣除下载券和下载次数;
按次数下载不扣除下载券;
24小时内重复下载只扣除一次;
顺序:VIP每日次数-->可用次数-->下载券;
共38页,可试读13页,点击继续阅读 ↓↓
相关文档

关于我们|帮助中心|下载说明|相关软件|意见反馈|联系我们

Copyright © 2008-现在 cucdc.com 高等教育资讯网 版权所有