Introduction to ERGM/p* model Kayo Fujimoto, Ph.D. Based on presentation slides by Nosh Contractor and Mengxiao Zhu.

Slides:



Advertisements
Similar presentations
CS188: Computational Models of Human Behavior
Advertisements

Ricard V. Solè and Sergi Valverde Prepared by Amaç Herdağdelen
An introduction to exponential random graph models (ERGM)
Where we are Node level metrics Group level metrics Visualization
Exact Inference in Bayes Nets
Likelihood Ratio, Wald, and Lagrange Multiplier (Score) Tests
Dynamic Bayesian Networks (DBNs)
Markov-Chain Monte Carlo
Structural Inference of Hierarchies in Networks BY Yu Shuzhi 27, Mar 2014.
Statistical inference for epidemics on networks PD O’Neill, T Kypraios (Mathematical Sciences, University of Nottingham) Sep 2011 ICMS, Edinburgh.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Exponential random graph (p*) models for social networks Workshop Harvard University February 2002 Philippa Pattison Garry Robins Department of Psychology.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Computational statistics 2009 Random walk. Computational statistics 2009 Random walk with absorbing barrier.
Joint social selection and social influence models for networks: The interplay of ties and attributes. Garry Robins Michael Johnston University of Melbourne,
Machine Learning CUNY Graduate Center Lecture 7b: Sampling.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Today Introduction to MCMC Particle filters and MCMC
Exponential Random Graph Models (ERGM) Michael Beckman PAD777 April 9, 2010.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Sunbelt 2009statnet Development Team ERGM introduction 1 Exponential Random Graph Models Statnet Development Team Mark Handcock (UW) Martina.
Chapter 9 Hypothesis Testing II. Chapter Outline  Introduction  Hypothesis Testing with Sample Means (Large Samples)  Hypothesis Testing with Sample.
6. Experimental Analysis Visible Boltzmann machine with higher-order potentials: Conditional random field (CRF): Exponential random graph model (ERGM):
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 2 – Slide 1 of 25 Chapter 11 Section 2 Inference about Two Means: Independent.
1 Institute of Engineering Mechanics Leopold-Franzens University Innsbruck, Austria, EU H.J. Pradlwarter and G.I. Schuëller Confidence.
Social Network Analysis and Complex Systems Science
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
Markov Random Fields Probabilistic Models for Images
Neighbourhood-based models for social networks: model specification issues Pip Pattison, University of Melbourne [with Garry Robins, University of Melbourne.
Exponential Random Graph Models Under Measurement Error Zoe Rehnberg with Dr. Nan Lin Washington University in St. Louis ARTU 2014.
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
A two minute introduction to: Exponential random graph (p*)models for social networks SNAC Workshop, Illinois, November 2005 Garry Robins, University of.
Probabilistic Models for Discovering E-Communities Ding Zhou, Eren Manavoglu, Jia Li, C. Lee Giles, Hongyuan Zha The Pennsylvania State University WWW.
Lecture 3: Statistics Review I Date: 9/3/02  Distributions  Likelihood  Hypothesis tests.
CS Statistical Machine learning Lecture 24
Lecture 2: Statistical learning primer for biologists
Discrepancy between Data and Fit. Introduction What is Deviance? Deviance for Binary Responses and Proportions Deviance as measure of the goodness of.
Introduction to Statistical Models for longitudinal network data Stochastic actor-based models Kayo Fujimoto, Ph.D.
Week 41 How to find estimators? There are two main methods for finding estimators: 1) Method of moments. 2) The method of Maximum likelihood. Sometimes.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
CS246 Latent Dirichlet Analysis. LSI  LSI uses SVD to find the best rank-K approximation  The result is difficult to interpret especially with negative.
Machine Learning 5. Parametric Methods.
Diversity Loss in General Estimation of Distribution Algorithms J. L. Shapiro PPSN (Parallel Problem Solving From Nature) ’06 BISCuit 2 nd EDA Seminar.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Latent Feature Models for Network Data over Time Jimmy Foulds Advisor: Padhraic Smyth (Thanks also to Arthur Asuncion and Chris Dubois)
Bootstrapping James G. Anderson, Ph.D. Purdue University.
Hierarchical Models. Conceptual: What are we talking about? – What makes a statistical model hierarchical? – How does that fit into population analysis?
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Markov Chain Monte Carlo in R
ERGM conditional form Much easier to calculate delta (change statistics)
Exponential random graph models for multilevel networks
A p* primer: logit models for social networks
Recovering Temporally Rewiring Networks: A Model-based Approach
CAP 5636 – Advanced Artificial Intelligence
Latent Dirichlet Analysis
Lecture 5 Unsupervised Learning in fully Observed Directed and Undirected Graphical Models.
Ch13 Empirical Methods.
CS 188: Artificial Intelligence
Expectation-Maximization & Belief Propagation
The Most General Markov Substitution Model on an Unrooted Tree
Markov Networks.
Maximum Likelihood We have studied the OLS estimator. It only applies under certain assumptions In particular,  ~ N(0, 2 ) But what if the sampling distribution.
Longitudinal Social Network Data
Presentation transcript:

Introduction to ERGM/p* model Kayo Fujimoto, Ph.D. Based on presentation slides by Nosh Contractor and Mengxiao Zhu

Four parts of ERGM Observed network data Network statistics (or counts) of each configuration ERG Modeling Conditional probability and Change statistics Estimation and Simulation Estimate Parameters by Simulation Method: MCMC ML estimation Goodness of fit test (convergence t-test) Compare observed and simulated graphs Recent development in ERGM New model specification

Exponential Random Graph Model (ERGM) ERGMs take the form of a probability distribution of graphs: ERGMs take the form of a probability distribution of graphs: Y is a set of tie indicator variables Y Y is a set of tie indicator variables Y y is a realization, the observed network y is a realization, the observed network g( y ) is a vector of network statistics g( y ) is a vector of network statistics θ is a parameter vector corresponding to g( y ) θ is a parameter vector corresponding to g( y ) k( θ ) is a normalizing factor calculated by summing up k( θ ) is a normalizing factor calculated by summing up exp{ θ ’g(y)} over all possible network configurations

Observed network Graph statistics (or counts) of each configuration

Network Statistics Examples for Undirected Networks Example: Edge: 6 2-Star: =11 3-Star: =5 4-Star: 1 Triangle: 2 a b d c e

Num of Edges Undirected Network Configurations A Simple Example of ERGM Homogeneous Assumption Number of configurations: Directed Network: Undirected Network:

A Simple ERG model Predict network using edge count θ can take different values: θ = 0, θ = -0.69, θ = 0.69 L(y) can the following values: L(y) = 0, L(y) = 1, L(y) = 2, L(y) = 3

Example 1: θ = 0, L=0 Model: Model: ERGM Formula ERGM Formula Probability of getting networks with 0 edge θ = 0

Example 1: θ = 0, L=1 Model: Model: ERGM Formula ERGM Formula θ = 0 Probability of getting networks with 1 edge

Example 1: θ = 0, L=2 Model: Model: ERGM Formula ERGM Formula θ = 0 Probability of getting networks with 2 edge

Probability of getting networks with 3 edge Example 1: θ = 0, L=3 Model: Model: ERGM Formula ERGM Formula θ = 0

Example 1: θ = 0 Model: Model: ERGM Formula ERGM Formula θ = 0

Example 2: θ = Model: Model: ERGM Formula ERGM Formula θ = -0.69

Example 3: θ = 0.69 Model: Model: ERGM Formula ERGM Formula θ = 0.69

Why Change Statistics? Huge Sample Space Huge Sample Space Num of configurations :

ERG modeling Conditional Probability and Change Statistics

Conditional Probability vs. Total Probability Total probability of the whole network Total probability of the whole network It is impossible to calculate when the size of the network gets large It is impossible to calculate when the size of the network gets large Introduce the Conditional Probability of edges Introduce the Conditional Probability of edges Reduce sample space Reduce sample space

Avoid the Calculation on Sample Space Conditional Probability of an Edge to exist Conditional Probability of an Edge to exist Conditional Probability of an Edge to be absent is Conditional Probability of an Edge to be absent is Logit p* model: model log odds ratio of Yij exists Logit p* model: model log odds ratio of Yij exists

Change Statistics (logit p* model) From the end of last slide, we have: From the end of last slide, we have: Define Change Statistics as: Define Change Statistics as: Model log odds of a tie being present to absent: Model log odds of a tie being present to absent:

Estimation and Simulation (Monte Carlo Markov Chain Maximum Likelihood Method)

Review: Maximum Likelihood Estimation (MLE) Likelihood functions Likelihood functions Estimate parameter θ given the observed network. Estimate parameter θ given the observed network. Maximum Likelihood Estimation Maximum Likelihood Estimation Find θ values such that the observed statistics are equal to the expected statistics Find θ values such that the observed statistics are equal to the expected statistics Approximate MLE by simulation Approximate MLE by simulation

Procedures for simulating ERG distribution Markov Chain Monte Carlo Maximum Likelihood Estimation (MCMCMLE) Markov Chain Monte Carlo Maximum Likelihood Estimation (MCMCMLE) 1. Simulate a distribution of random graphs from a starting set of parameter values 2. Refine the parameter values by comparing the distribution of graphs against the observed graph 3. Repeat this process until the parameter estimate stabilize

Convergence T-statistics Test adequacy of parameter values estimated Test adequacy of parameter values estimated 1. T-statistics for each configuration T <|.1|  good fit NOTE: If the parameter estimates do not converge, the model is degenerate

A Simple Example of MCMCMLE Model: Model: Observed Network y: Observed Network y: Goal: Find θ value such that the observed number of edges are equal to the expected number of edges Goal: Find θ value such that the observed number of edges are equal to the expected number of edges

If θ can be chosen from the following 3 cases, θ=-0.69 is preferred because it gives the highest probability for the observed network θP(Y=y) Given the observed Network y: Given the observed Network y:

Markov dependence (Frank and Strauss, 1986) Potential ties are dependent only if they share a common actor Two possible network ties are conditionally independent unless they share a common actor Once homogeneity assumption is imposed, we obtain the following configurations…

Markov random graph models (non-directed networks) Density or edge(  ) Two-star(  2 ) Three-star(  3 ) Triangle(  )

Problems of degeneracy for Markov random models Certain parameter values place almost all of the probability mass on either the empty or the full graph Simulation studies showed that Markov random graph models are degenerate for many empirical networks with high level of clustering A few very high degree nodes Some regions of high triangulation

Two possibilities for the degeneracy problem (Snijders, et al 2006) Makov dependence assumption may be too restrictive The representation of transitivity by the total number of triangles might be too simplistic  New specification of higher order network dependency

New development in ERGM Partial conditional dependence assumption and new model specification

Partial conditional dependence (Social circuit dependence) Two possible network ties being conditionally dependent if their observation would lead to a 4-cycle i k j l = possible edges = observed edges

Partial conditional dependence (Example) Daughter A Father B Father A Daughter B

Difference between the two types of dependence assumptions i k j l Markov dependence assumptions Partial conditional dependence assumptions = potential tie = ties which affect the formation of the potential tie = ties with no effect on the potential tie i k j l

New Specifications of ERGM Represent structural parameters similar to the Markov parameters Effects are incorporated within the one configuration parameter Three new statistics for non-directed network Alternating k-stars Alternating k-triangles Alternating independent two-paths

Alternating k-star configuration (degree dist’n): Alternating k-triangle (tendency to form triads): Alternating k-two-path (tendency to form cycles) Examples of new specifications

Interpretation of the parameter Positive alternating k-star parameter Networks with some higher degree nodes are highly probable.  Core-periphery structure Positive alternating k-triangle parameter Triangulation in the network as well as tendencies for triangles themselves group together in larger higher order “clump” Positive alternating k-path parameter Tendency for 4-cycles in the network

Summary for model construction Random variables Each network tie (Yij) among nodes of a network A random tie variable Yij=1 if a tie form i to j exist, Yij=0 otherwise y ij the observed value of the variable Yij Dependence assumptions Define contingencies among network variables Determine the type of parameters in the model Ties also depends on node-level attributes (homophily) Homogeneity assumption Simplify parameters by imposing homogeneity constraints. Estimation procedures Find the best parameter values based on the observed network Use simulation (MCMLE)

Software for ERGM SIENA (Snijders, and colleagues) PNet (Robbins, and colleagues) Statnet (Butts, and colleagues)

Reference Harrigan, Nicholas. “ Exponential Rnadom Graph (ERG) models and their application to the study of corporate elites. Robins, Garry (manuscript). Exponential Random Graph (p*) models for social Networks, published in Melnet website. Robins, G., Pattison, P. Kalish, y. Lusher, D. (2007). “An introduction to exponential random graph (p*) models for social networks”. Social Networks, 29, Snijders, T.A.B., Pattison, P., Robins, G, Hancock M. (2006). “New specifications for exponential random graph models. Sociological Methodology, 36:

Thank you for your attention Any questions?