CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

Similar documents
Lecture 17: More on Markov Decision Processes. Reinforcement learning

CS360 Homework 14 Solution

Maximizing the Spread of Influence through a Social Network Problem/Motivation: Suppose we want to market a product or promote an idea or behavior in

CEC login. Student Details Name SOLUTIONS

Markov Decision Processes

Q1. [?? pts] Search Traces

Opinion formation CS 224W. Cascades, Easley & Kleinberg Ch 19 1

CS 188: Artificial Intelligence

CPS 270: Artificial Intelligence Markov decision processes, POMDPs

CS 343: Artificial Intelligence

Integer Programming Models

COMP417 Introduction to Robotics and Intelligent Systems. Reinforcement Learning - 2

Essays on Some Combinatorial Optimization Problems with Interval Data

DRAFT. 1 exercise in state (S, t), π(s, t) = 0 do not exercise in state (S, t) Review of the Risk Neutral Stock Dynamics

Monte-Carlo Planning: Introduction and Bandit Basics. Alan Fern

Approximations of Stochastic Programs. Scenario Tree Reduction and Construction

CMSC 858F: Algorithmic Game Theory Fall 2010 Introduction to Algorithmic Game Theory

Lecture 7: Bayesian approach to MAB - Gittins index

Lecture 5. 1 Online Learning. 1.1 Learning Setup (Perspective of Universe) CSCI699: Topics in Learning & Game Theory

Monte-Carlo Planning: Introduction and Bandit Basics. Alan Fern

Reinforcement Learning. Slides based on those used in Berkeley's AI class taught by Dan Klein

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

Extending MCTS

CS 461: Machine Learning Lecture 8

Maximizing the Spread of Influence through a Social Network

CS599: Algorithm Design in Strategic Settings Fall 2012 Lecture 6: Prior-Free Single-Parameter Mechanism Design (Continued)

Fibonacci Heaps Y Y o o u u c c an an s s u u b b m miitt P P ro ro b blle e m m S S et et 3 3 iin n t t h h e e b b o o x x u u p p fro fro n n tt..

Yao s Minimax Principle

Reinforcement Learning Analysis, Grid World Applications

IEOR E4004: Introduction to OR: Deterministic Models

CS188 Spring 2012 Section 4: Games

Budget Management In GSP (2018)

Outline for today. Stat155 Game Theory Lecture 13: General-Sum Games. General-sum games. General-sum games. Dominated pure strategies

Efficient Estimation of Influence Functions for SIS Model on Social Networks

Quantitative Trading System For The E-mini S&P

Optimal Policies for Distributed Data Aggregation in Wireless Sensor Networks

On Bidding Algorithms for a Distributed Combinatorial Auction

Monte Carlo Methods (Estimators, On-policy/Off-policy Learning)

2D5362 Machine Learning

Action Selection for MDPs: Anytime AO* vs. UCT

Importance Sampling for Fair Policy Selection

Reinforcement Learning

CS221 / Spring 2018 / Sadigh. Lecture 7: MDPs I

Lecture 7: MDPs I. Question. Course plan. So far: search problems. Uncertainty in the real world

Issues. Senate (Total = 100) Senate Group 1 Y Y N N Y 32 Senate Group 2 Y Y D N D 16 Senate Group 3 N N Y Y Y 30 Senate Group 4 D Y N D Y 22

Making Decisions. CS 3793 Artificial Intelligence Making Decisions 1

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2017

June 11, Dynamic Programming( Weighted Interval Scheduling)

Large-Scale SVM Optimization: Taking a Machine Learning Perspective

Reinforcement learning and Markov Decision Processes (MDPs) (B) Avrim Blum

Algorithms and Networking for Computer Games

Worst-Case vs. Average Case. CSE 473: Artificial Intelligence Expectimax, Uncertainty, Utilities. Expectimax Search. Worst-Case vs.

CS221 / Spring 2018 / Sadigh. Lecture 9: Games I

Reinforcement Learning (1): Discrete MDP, Value Iteration, Policy Iteration

Non-Deterministic Search

Benchmarking. Club Fund. We like to think about being in an investment club as a group of people running a little business.

Reinforcement Learning (1): Discrete MDP, Value Iteration, Policy Iteration

Expectimax and other Games

Lecture 9: Games I. Course plan. A simple game. Roadmap. Machine learning. Example: game 1

PORTFOLIO OPTIMIZATION AND EXPECTED SHORTFALL MINIMIZATION FROM HISTORICAL DATA

CS364A: Algorithmic Game Theory Lecture #14: Robust Price-of-Anarchy Bounds in Smooth Games

15-451/651: Design & Analysis of Algorithms November 9 & 11, 2015 Lecture #19 & #20 last changed: November 10, 2015

Lecture 8 Feb 16, 2017

1 Appendix A: Definition of equilibrium

The Assumption(s) of Normality

6 -AL- ONE MACHINE SEQUENCING TO MINIMIZE MEAN FLOW TIME WITH MINIMUM NUMBER TARDY. Hamilton Emmons \,«* Technical Memorandum No. 2.

Basic Framework. About this class. Rewards Over Time. [This lecture adapted from Sutton & Barto and Russell & Norvig]

Introduction to Artificial Intelligence Spring 2019 Note 2

10703 Deep Reinforcement Learning and Control

G5212: Game Theory. Mark Dean. Spring 2017

Mechanism Design and Auctions

Uncertain Outcomes. CS 188: Artificial Intelligence Uncertainty and Utilities. Expectimax Search. Worst-Case vs. Average Case

Complex Decisions. Sequential Decision Making

HEMI: Hyperedge Majority Influence Maximization

Definition 4.1. In a stochastic process T is called a stopping time if you can tell when it happens.

Lecture 2: Making Good Sequences of Decisions Given a Model of World. CS234: RL Emma Brunskill Winter 2018

Probabilities. CSE 473: Artificial Intelligence Uncertainty, Utilities. Reminder: Expectations. Reminder: Probabilities

CS 343: Artificial Intelligence

Sublinear Time Algorithms Oct 19, Lecture 1

Pro Strategies Help Manual / User Guide: Last Updated March 2017

Lecture 9 Feb. 21, 2017

Markov Decision Processes

15-451/651: Design & Analysis of Algorithms October 23, 2018 Lecture #16: Online Algorithms last changed: October 22, 2018

> asympt( ln( n! ), n ); n 360n n

MATH 121 GAME THEORY REVIEW

Scenario reduction and scenario tree construction for power management problems

CS 5522: Artificial Intelligence II

CS 188: Artificial Intelligence. Outline

Introduction to Fall 2011 Artificial Intelligence Midterm Exam

Lessons from two case-studies: How to Build Accurate and models

Markov Decision Processes II

ORIGINALLY APPEARED IN ACTIVE TRADER M AGAZINE

THE TRAVELING SALESMAN PROBLEM FOR MOVING POINTS ON A LINE

Posted-Price Mechanisms and Prophet Inequalities

Homework #4. CMSC351 - Spring 2013 PRINT Name : Due: Thu Apr 16 th at the start of class

Lecture 12: Introduction to reasoning under uncertainty. Actions and Consequences

The Irrevocable Multi-Armed Bandit Problem

Introduction to Fall 2011 Artificial Intelligence Midterm Exam

CSE 417 Dynamic Programming (pt 2) Look at the Last Element

Recharging Bandits. Joint work with Nicole Immorlica.

Transcription:

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 2 (1) New problem: Outbreak detection (2) Develop an approximation algorithm It is a submodular opt. problem! (3) Speed-up greedy hill-climbing Valid for optimizing general submodular functions (i.e., also works for influence maximization) (4) Prove a new data dependent bound on the solution quality Valid for optimizing any submodular function (i.e., also works for influence maximization)

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 3 Given a real city water distribution network And data on how contaminants spread in the network Detect the contaminant as quickly as possible Problem posed by the US Environmental Protection Agency S

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 4 Posts Blogs Information cascade Time ordered hyperlinks Which blogs should one read to detect cascades as effectively as possible?

Want to read things before others do. Detect blue & yellow soon but miss red. Detect all stories but late. 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 5

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 6 Both of these two are an instance of the same underlying problem! Given a dynamic process spreading over a network we want to select a set of nodes to detect the process effectively Many other applications: Epidemics Influence propagation Network security

Utility of placing sensors: Water flow dynamics, demands of households, For each subset S V compute utility f(s) High impact outbreak Contamination S3 S1S2 Low impact outbreak Medium impact outbreak S3 S1 S4 Set V of all network junctions Sensor reduces impact through early detection! S1 S4 S2 High sensing quality f(s) = 0.9 Low sensing quality f(s)=0.01 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 7

Given: Graph G(V, E) Data on how outbreaks spread over the G: For each outbreak i we know the time T(u, i) when outbreak i contaminates node u Water distribution network (physical pipes and junctions) Simulator of water consumption&flow (built by Mech. Eng. people) We simulate the contamination spread for every possible location. 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 8

Given: Graph G(V, E) Data on how outbreaks spread over the G: For each outbreak i we know the time T(u, i) when outbreak i contaminates node u a b c b a c The network of the blogosphere Traces of the information flow and identify influence sets Collect lots of blogs posts and trace hyperlinks to obtain data about information flow from a given blog. 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 9

Given: Graph G(V, E) Data on how outbreaks spread over the G: For each outbreak i we know the time T(u, i) when outbreak i contaminates node u Goal: Select a subset of nodes S that maximizes the expected reward: max / 1 f S = 5 P i f 7 S subject to: cost(s) < B 7 Expected reward for detecting outbreak i P(i) probability of outbreak i occurring. f(i) reward for detecting outbreak i using sensors S. 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 10

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 11 Reward (one of the following three): (1) Minimize time to detection (2) Maximize number of detected propagations (3) Minimize number of infected people Cost (context dependent): Reading big blogs is more time consuming Placing a sensor in a remote location is expensive outbreak i Monitoring blue node saves more people than monitoring the green node f(s)

f i S is penalty reduction: f 7 S = π 7 π 7 (S) Objective functions: 1) Time to detection (DT) How long does it take to detect a contamination? Penalty for detecting at time t: π 7 (t) = min {t, T >?@ } 2) Detection likelihood (DL) How many contaminations do we detect? Penalty for detecting at time t: π 7 (t) = 0, π 7 ( ) = 1 Note, this is binary outcome: we either detect or not 3) Population affected (PA) How many people drank contaminated water? Penalty for detecting at time t: π 7 (t) = {# of infected nodes in outbreak i by time t}. Observation: In all cases detecting sooner does not hurt! 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 12

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 13 Observation: Diminishing returns S 1 New sensor: S s S 1 S 3 S 2 S 2 S 4 Placement S={s 1, s 2 } Adding s helps a lot Placement S ={s 1, s 2, s 3, s 4 } Adding s helps very little

Claim: For all A B V and sensors s V\B f A s f A f B s f B Proof: All our objectives are submodular Fix cascade/outbreak i Show f i A = π i π i (T(A, i)) is submodular Consider A B V and sensor s V\B When does node s detect cascade i? We analyze 3 cases based on when s detects outbreak i (1) T s, i T(A, i): s detects late, nobody benefits: f 7 A s = f 7 A, also f 7 B s = f 7 B and so f 7 A s f 7 A = 0 = f 7 B s f 7 B 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 14

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 15 Proof (contd.): (2) T B, i T s, i < T A, i : s detects after B but before A s detects sooner than any node in A but after all in B. So s only helps improve the solution A (but not B) f 7 A s f 7 A 0 = f 7 B s f 7 B (3) T s, i < T(B, i): s detects early f 7 A s f 7 A = π 7 π 7 T s, i f 7 (A) π 7 π 7 T s, i f 7 (B) = f 7 B s f 7 B Ineqaulity is due to non-decreasingness of f 7 ( ), i.e., f 7 A f 7 (B) So, f i ( ) is submodular! So, f( ) is also submodular Remember A B f S = 5 P i f 7 S 7

a b c d e Hill-climbing reward b Add sensor with highest marginal gain 10/27/16 c d a e What do we know about optimizing submodular functions? A hill-climbing (i.e., greedy) is near optimal: (1 1 e ) OPT But: (1) This only works for unit cost case! (each sensor costs the same) For us each sensor s has cost c(s) (2) Hill-climbing algorithm is slow At each iteration we need to re-evaluate marginal gains of all nodes Runtime O( V K) for placing K sensors Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu Part 2-16

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 17

Consider the following algorithm to solve the outbreak detection problem: Hill-climbing that ignores cost Ignore sensor cost c(s) Repeatedly select sensor with highest marginal gain Do this until the budget is exhausted Q: How well does this work? A: It can fail arbitrarily badly! L Next we come up with an example where Hillclimbing solution is arbitrarily away from OPT 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 18

Bad example when we ignore cost: n sensors, budget B s 1 : reward r, cost B, s 2 s n : reward r ε, All sensors have the same cost: c s i = 1 Hill-climbing always prefers more expensive sensor s 1 with reward r (and exhausts the budget). It never selects cheaper sensors with reward r ε For variable cost it can fail arbitrarily badly! Idea: What if we optimize benefit-cost ratio? s 7 = arg max j 1 f A 7kl {s} f(a 7kl ) c s Greedily pick sensor s i that maximizes benefit to cost ratio. 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 19

Benefit-cost ratio can also fail arbitrarily badly! Consider: budget B: 2 sensors s 1 and s 2 : Costs: c(s 1 ) = ε, c(s 2 ) = B Only 1 cascade: f(s 1 ) = 2ε, f(s 2 ) = B Then benefit-cost ratio is: B/c(s 1 ) = 2 and B/c(s 2 ) = 1 So, we first select s 1 and then can not afford s 2 We get reward 2ε instead of B! Now send ε 0 and we get arbitrarily bad solution! This algorithm incentivizes choosing nodes with very low cost, even when slightly more expensive ones can lead to much better global results. 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 20

CELF (Cost-Effective Lazy Forward-selection) A two pass greedy algorithm: Set (solution) S : Use benefit-cost greedy Set (solution) S : Use unit-cost greedy Final solution: S = arg max (f(s ),f(s )) How far is CELF from (unknown) optimal solution? Theorem: CELF is near optimal [Krause&Guestrin, 05] CELF achieves ½(1-1/e) factor approximation! This is surprising: We have two clearly suboptimal solutions, but taking best of the two is guaranteed to give a near-optimal solution. 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 21

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 22

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 23 a b c d e Hill-climbing reward b Add sensor with highest marginal gain c d a e What do we know about optimizing submodular functions? A hill-climbing (i.e., greedy) is near optimal (that is, (1 l v ) OPT) But: (2) Hill-climbing algorithm is slow! At each iteration we need to reevaluate marginal gains of all nodes Runtime O( V K) for placing K sensors

In round i + 1: So far we picked S i = {s 1,, s i } Now pick s i{1 = arg max u f(s i {u}) f(s i ) This our old friend greedy hill-climbing algorithm. It maximizes the marginal benefit δ i u = f(s i {u}) f(s i ) By submodularity property: f S 7 u f S 7 f S ƒ u f S ƒ for i < j Observation: By submodularity: For every u δ 7 (u) δ ƒ (u) for i < j since S i Sj δ i (u) Marginal benefits δ i (u) only shrink! (as i grows) 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 24 u δ j (u) Activating node u in step i helps more than activating it at step j (j>i)

Idea: Use δ i as upper-bound on δ j (j > i) Lazy hill-climbing: Keep an ordered list of marginal benefits δ i from previous iteration Re-evaluate δ i only for top node Re-sort and prune Marginal gain a b c d e S 1 ={a} f(s {u}) f(s) f(t {u}) f(t) S T 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 25

Idea: Use δ i as upper-bound on δ j (j > i) Lazy hill-climbing: Keep an ordered list of marginal benefits δ i from previous iteration Re-evaluate δ i only for top node Re-sort and prune Marginal gain a b c d e S 1 ={a} f(s {u}) f(s) f(t {u}) f(t) S T 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 26

Idea: Use δ i as upper-bound on δ j (j > i) Lazy hill-climbing: Keep an ordered list of marginal benefits δ i from previous iteration Re-evaluate δ i only for top node Re-sort and prune Marginal gain a d b e c S 1 ={a} S 2 ={a,b} f(s {u}) f(s) f(t {u}) f(t) S T 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 27

CELF (using Lazy evaluation) runs 700 times faster than greedy hillclimbing algorithm 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 28

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 29

Back to the solution quality! The (1-1/e) bound for submodular functions is the worst case bound (worst over all possible inputs) Data dependent bound: Value of the bound depends on the input data On easy data, hill climbing may do better than 63% Can we say something about the solution quality when we know the input data? 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 30

Suppose S is some solution to f(s) s.t. S k f(s) is monotone & submodular Let OPT = {t 1,, t k } be the OPT solution For each u let δ u = f S u f S Order δ u so that δ 1 δ 2 k iš1 Then: f OPT f S + δ i Note: This is a data dependent bound (δ i depends on input data) Bound holds for any algorithm Makes no assumption about how S was computed For some inputs it can be very loose (worse than 63%) 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 31

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 32 Claim: For each u let δ(u) = f(s {u}) f(s) Order δ u so that δ 1 δ 2 k Then: f OPT f S + iš1 δ(i) Proof: f OPT f OPT S = f S + f S t l t 7 f S t l t 7kl 7Šl 7Šl 7Šl f S + f S t 7 f S = f S + δ(t 7 ) Instead of taking t i OPT (of benefit δ(t 7 )), we take the best possible element (δ(i)) f S + 7Šl δ(i) f T f S + iš1 δ(i) k (we proved this last time)

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 33

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 34 Real metropolitan area water network V = 21,000 nodes E = 25,000 pipes Use a cluster of 50 machines for a month Simulate 3.6 million epidemic scenarios (random locations, random days, random time of the day)

Solution quality F(A) Higher is better 1.4 1.2 1 0.8 0.6 0.4 Offline the (1-1/e) bound Data-dependent bound Hill Climbing 0.2 0 0 5 10 15 20 Number of sensors placed Data-dependent bound is much tighter (gives more accurate estimate of alg. performance) 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 35

[w/ Ostfeld et al., J. of Water Resource Planning] Author Score Placement heuristics perform much worse CELF 26 Sandia 21 U Exter 20 Bentley systems 19 Technion (1) 14 Bordeaux 12 U Cyprus 11 U Guelph 7 U Michigan 4 Michigan Tech U 3 Malcolm 2 Proteo 2 Technion (2) 1 Battle of Water Sensor Networks competition 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 36

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 37 Different objective functions give different sensor placements Population affected Detection likelihood

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 38 CELF is 10 times faster than greedy hill-climbing!

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 39 = I have 10 minutes. Which blogs should I read to be? most up to date? = Who are the most influential bloggers?

Want to read things before others do. Detect blue & yellow soon but miss red. Detect all stories but late. 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 40

Crawled 45,000 blogs for 1 year Obtained 10 million posts And identified 350,000 cascades Cost of a blog is the number of posts it has 41

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 42 Online bound turns out to be much tighter! Based on the plot below: 87% instead of 32.5% Old bound vs. Our bound CELF

Heuristics perform much worse! One really needs to perform the optimization 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 43

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 44 CELF has 2 sub-algorithms. Which wins? Unit cost: CELF picks large popular blogs Cost-benefit: Cost proportional to the number of posts We can do much better when considering costs

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 45 Problem: Then CELF picks lots of small blogs that participate in few cascades We pick best solution that interpolates between the costs f(s)=0.3 Score f(s)=0.4 We can get good solutions with few blogs and few posts f(s)=0.2 Each curve represents a set of solutions S with the same final reward f(s)

10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu Part 2-46 We want to generalize well to future (unknown) cascades Limiting selection to bigger blogs improves generalization!

[Leskovec et al., KDD 07] 10/27/16 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 47 CELF runs 700 times faster than simple hillclimbing algorithm