Human-centered Machine Learning

Slides:



Advertisements
Similar presentations
BURSTY SUBGRAPHS IN SOCIAL NETWORKS. Introduction 2.
Advertisements

Unità di Perugia e di Roma “Tor Vergata” "Uncertain production systems: optimal feedback control of the single site and extension to the multi-site case"
Taylor Collins 1 RECURSIVE MACROECONOMIC THEORY, LJUNGQVIST AND SARGENT, 3 RD EDITION, CHAPTER 19 DYNAMIC STACKELBER G PROBLEMS.
20 10 School of Electrical Engineering &Telecommunications UNSW UNSW 2. Laguerre Parameterised Hawkes Process To model spike train data,
Efficient Autoscaling in the Cloud using Predictive Models for Workload Forecasting Roy, N., A. Dubey, and A. Gokhale 4th IEEE International Conference.
COSC 878 Seminar on Large Scale Statistical Machine Learning 1.
Reinforcement Learning & Apprenticeship Learning Chenyi Chen.
1 1 Chenhao Tan, 1 Jie Tang, 2 Jimeng Sun, 3 Quan Lin, 4 Fengjiao Wang 1 Department of Computer Science and Technology, Tsinghua University, China 2 IBM.
Support Vector Machines
D Nagesh Kumar, IIScOptimization Methods: M1L1 1 Introduction and Basic Concepts (i) Historical Development and Model Building.
4/1 Agenda: Markov Decision Processes (& Decision Theoretic Planning)
1 Multiple Kernel Learning Naouel Baili MRL Seminar, Fall 2009.
Dimitrios Konstantas, Evangelos Grigoroudis, Vassilis S. Kouikoglou and Stratos Ioannidis Department of Production Engineering and Management Technical.
Diffusion in Social and Information Networks Part II W ORLD W IDE W EB 2015, F LORENCE MPI for Software SystemsGeorgia Institute of Technology Le Song.
1 Enabling Large Scale Network Simulation with 100 Million Nodes using Grid Infrastructure Hiroyuki Ohsaki Graduate School of Information Sci. & Tech.
Capacity analysis of complex materials handling systems.
Natural Gradient Works Efficiently in Learning S Amari (Fri) Computational Modeling of Intelligence Summarized by Joon Shik Kim.
Annealing Paths for the Evaluation of Topic Models James Foulds Padhraic Smyth Department of Computer Science University of California, Irvine* *James.
Influence Maximization in Dynamic Social Networks Honglei Zhuang, Yihan Sun, Jie Tang, Jialin Zhang, Xiaoming Sun.
Online Learning for Collaborative Filtering
Model-based Bayesian Reinforcement Learning in Partially Observable Domains by Pascal Poupart and Nikos Vlassis (2008 International Symposium on Artificial.
The Application of The Improved Hybrid Ant Colony Algorithm in Vehicle Routing Optimization Problem International Conference on Future Computer and Communication,
Prediction of Influencers from Word Use Chan Shing Hei.
RSVM: Reduced Support Vector Machines Y.-J. Lee & O. L. Mangasarian First SIAM International Conference on Data Mining Chicago, April 6, 2001 University.
Measuring Behavioral Trust in Social Networks
Forrelation: A Problem that Optimally Separates Quantum from Classical Computing.
Notes Over 5.6 Quadratic Formula
Nonlinear differential equation model for quantification of transcriptional regulation applied to microarray data of Saccharomyces cerevisiae Vu, T. T.,
Speaker : Yu-Hui Chen Authors : Dinuka A. Soysa, Denis Guangyin Chen, Oscar C. Au, and Amine Bermak From : 2013 IEEE Symposium on Computational Intelligence.
A comparative approach for gene network inference using time-series gene expression data Guillaume Bourque* and David Sankoff *Centre de Recherches Mathématiques,
Unsupervised Streaming Feature Selection in Social Media
Biao Wang 1, Ge Chen 1, Luoyi Fu 1, Li Song 1, Xinbing Wang 1, Xue Liu 2 1 Shanghai Jiao Tong University 2 McGill University
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
The Monte Carlo Method/ Markov Chains/ Metropolitan Algorithm from sec in “Adaptive Cooperative Systems” -summarized by Jinsan Yang.
Machine learning for Dynamic Social Network Analysis
Machine learning for Dynamic Social Network Analysis
Lecture 2: Performance Evaluation
Manuel Gomez Rodriguez
CEE 6410 Water Resources Systems Analysis
Machine learning for Dynamic Social Network Analysis
Variational filtering in generated coordinates of motion
Q-Learning for Policy Improvement in Network Routing
Depth from disparity (x´,y´)=(x+D(x,y), y)
CARP: Context-Aware Reliability Prediction of Black-Box Web Services
Manuel Gomez Rodriguez
Center for Complexity in Business, R. Smith School of Business
Reinforcement Learning
Policy Gradient in Continuous Time
"Playing Atari with deep reinforcement learning."
Degree and Eigenvector Centrality
Random WALK, BROWNIAN MOTION and SDEs
Unit 7 Day 4 the Quadratic Formula.
Statistical Learning Dong Liu Dept. EEIS, USTC.
Collaborative Filtering Matrix Factorization Approach
Filtering and State Estimation: Basic Concepts
Nonlinear Network Structures for Optimal Control
Unit 4: Dynamic Programming
Jiequn Han, Arnulf Jentzen, and Weinan E Presented By Aishwarya Singh
Identification of Wiener models using support vector regression
The loss function, the normal equation,
Deep Reinforcement Learning
Optimal stochastic control in continuous time with Wiener processes: - General results and applications to optimal wildlife management KEYNOTE Version.
Mathematical Foundations of BME Reza Shadmehr
Deep Reinforcement Learning: Learning how to act using a deep neural network Psych 209, Winter 2019 February 12, 2019.
Human-centered Machine Learning
Human-centered Machine Learning
Human-centered Machine Learning
Human-centered Machine Learning
Human-centered Machine Learning
Human-centered Machine Learning
Presentation transcript:

Human-centered Machine Learning Viral marketing with Stochastic optimal control of TPP Human-centered Machine Learning http://courses.mpi-sws.org/hcml-ws18/

Maximizing activity in a social network Can we steer users’ behavior to maximize activity in a social network?

Endogenous and exogeneous events Exogenous activity Users’ actions due to drives external to the network Endogenous activity Users’ responses to other users’ actions in the network

Multidimensional Hawkes process For each user u, actions as a counting process Nu(t) Nu(t) Intensities or rates (Actions per time unit) User influence matrix Non-negative kernel (memory) Exogenous actions Endogenous actions

Steering endogenous actions Organic actions + Directly incentivized actions Directly incentivized actions Intensities of directly incentivized actions [Zarezade et al., 2018]

Cost to go & Bellman’s principle of optimality Loss Optimization problem Dynamics defined by Jump SDEs To solve the problem, we first define the corresponding optimal cost-to-go: The cost-to-go, evaluated at t0, recovers the optimization problem! [Zarezade et al., 2018]

Cost to go & Bellman’s principle of optimality Loss Optimization problem This is a stochastic optimal control problem for jump SDEs (we know how to solve this!) Dynamics defined by Jump SDEs To solve the problem, we first define the corresponding optimal cost-to-go: The cost-to-go, evaluated at t0, recovers the optimization problem! [Zarezade et al., 2018]

Hamilton-Jacobi-Bellman (HJB) equation Lemma. The optimal cost-to-go satisfies Bellman’s Principle of Optimality t Hamilton-Jacobi-Bellman (HJB) equation Partial differential equation in J (with respect to λ and t) [Zarezade et al., 2018]

Solving the HJB equation Consider a quadratic loss Rewards organic actions Penalizes directly incentivizes actions We propose and then show that the optimal intensity is: Computed offline once! Closed form solution to a first order ODE Solution to a matrix Riccati differential equation [Zarezade et al., 2018]

The Cheshire algorithm Intuition Steering actions means sampling action user & times from u*(t) More in detail Since the intensity function u*(t) is stochastic, we sample from it using: Superposition principle Standard thinning Easy to implement It only requires sampling from inhomog. Poisson! 10 10 [Zarezade et al., 2018]

Experiments on real data Five Twitter datasets (users) where actions are tweets and retweets 1. Fit model parameters Network inference! exogeneous rate influence matrix 2. Simulate steering endogenous actions directly incentivized tweets chosen by each method [Zarezade et al., 2018]

Evaluation metrics & baselines Average number of not directly incentivized tweets Average time to reach 30,000 not directly incentivized tweets Baselines MSC [Farajtabar et al., NIPS ’16] OPL [Farajtabar et al., NIPS ’14] PRK (Pagerank) DEG (Out-degree) [Zarezade et al., 2018]

Performance vs. time Sports, M(tf) ≈ 5k Series, M(tf) ≈ 5k Cheshire (in red) triggers 100%-400% more posts than the second best performer. [Zarezade et al., 2018]

Performance vs. # of incentivized tweets Sports, M(tf) ≈ 5k Series, M(tf) ≈ 5k Cheshire (in red) reaches 30K tweets 20-50% faster than the second best performer [Zarezade et al., 2018]

Why Cheshire? “the Cheshire Cat has the ability to appear and disappear in any location” Alice’s Adventures in Wonderland, Lewis Carroll