Soft configuration model

From testwiki
Revision as of 20:25, 15 January 2024 by imported>The Banner (β†’Model formulation)
(diff) ← Older revision | Latest revision (diff) | Newer revision β†’ (diff)
Jump to navigation Jump to search

Template:Short description Template:Network science

In applied mathematics, the soft configuration model (SCM) is a random graph model subject to the principle of maximum entropy under constraints on the expectation of the degree sequence of sampled graphs.[1] Whereas the configuration model (CM) uniformly samples random graphs of a specific degree sequence, the SCM only retains the specified degree sequence on average over all network realizations; in this sense the SCM has very relaxed constraints relative to those of the CM ("soft" rather than "sharp" constraints[2]). The SCM for graphs of size n has a nonzero probability of sampling any graph of size n, whereas the CM is restricted to only graphs having precisely the prescribed connectivity structure.

Model formulation

The SCM is a statistical ensemble of random graphs G having n vertices (n=|V(G)|) labeled {vj}j=1n=V(G), producing a probability distribution on 𝒒n (the set of graphs of size n). Imposed on the ensemble are n constraints, namely that the ensemble average of the degree kj of vertex vj is equal to a designated value k^j, for all vjV(G). The model is fully parameterized by its size n and expected degree sequence {k^j}j=1n. These constraints are both local (one constraint associated with each vertex) and soft (constraints on the ensemble average of certain observable quantities), and thus yields a canonical ensemble with an extensive number of constraints.[2] The conditions kj=k^j are imposed on the ensemble by the method of Lagrange multipliers (see Maximum-entropy random graph model).

Derivation of the probability distribution

The probability β„™SCM(G) of the SCM producing a graph G is determined by maximizing the Gibbs entropy S[G] subject to constraints kj=k^j, j=1,,n and normalization G𝒒nβ„™SCM(G)=1. This amounts to optimizing the multi-constraint Lagrange function below:

β„’(α,{ψj}j=1n)=G𝒒nβ„™SCM(G)logβ„™SCM(G)+α(1G𝒒nβ„™SCM(G))+j=1nψj(k^jG𝒒nβ„™SCM(G)kj(G)),

where α and {ψj}j=1n are the n+1 multipliers to be fixed by the n+1 constraints (normalization and the expected degree sequence). Setting to zero the derivative of the above with respect to β„™SCM(G) for an arbitrary G𝒒n yields

0=β„’(α,{ψj}j=1n)β„™SCM(G)=logβ„™SCM(G)1αj=1nψjkj(G)  β„™SCM(G)=1Zexp[j=1nψjkj(G)],

the constant Z:=eα+1=G𝒒nexp[j=1nψjkj(G)]=1i<jn(1+e(ψi+ψj))[3] being the partition function normalizing the distribution; the above exponential expression applies to all G𝒒n, and thus is the probability distribution. Hence we have an exponential family parameterized by {ψj}j=1n, which are related to the expected degree sequence {k^j}j=1n by the following equivalent expressions:

kq=G𝒒nkq(G)β„™SCM(G)=logZψq=jq1eψq+ψj+1=k^q, q=1,,n.

References

Template:Reflist