Information Retrieval

Size: px
Start display at page:

Download "Information Retrieval"

Transcription

1 Information Retrieval Ranked Retrieval & the Vector Space Model Gintarė Grigonytė Department of Linguistics and Philology Uppsala University Slides based on IIR material Gintarė Grigonytė 1/57

2 Outline for today Ranked retrieval Efficient scoring & ranking IR system architecture Gintarė Grigonytė 2/57

3 Boolean Retrieval Model Advantages Easy for the system Users get transparency: it is easy to understand why a document was or was not retrieved Users get control: it easy to determine whether the query is too specific (few results) or too broad (many results) Disadvantages The burden is on the user to formulate a good boolean query Gintarė Grigonytė 3/57

4 Recap on Ranked Retrieval Boolean retrieval (= exact matches) is not enough! ranking of retrieved documents according to relevance vector-space model term weighting (tf-idf) documents as vectors (in space) queries as vectors (in same space) ranking according to vector similarity (proximity) How important is good ranking? Gintarė Grigonytė 4/57

5 (Slides from Dan Russell, Google)

6 To sum up: Good ranking is very important!

7 Recall: Boolean model (document = set of words) Anthony Julius The Hamlet Othello Macbeth... and Caesar Tempest Cleopatra ANTHONY BRUTUS CAESAR CALPURNIA CLEOPATRA MERCY WORSER Each document is represented by a binary vector {0, 1} V. Gintarė Grigonytė 7/57

8 Recall: Bag of Words Model Anthony Julius The Hamlet Othello Macbeth... and Caesar Tempest Cleopatra ANTHONY BRUTUS CAESAR CALPURNIA CLEOPATRA MERCY WORSER Each document is represented by a count vector N V (term frequency). Gintarė Grigonytė 8/57

9 Term frequency weight relevance as (non-linear) function of term frequency log frequency weight of term t in d is defined as follows w t,d = { 1 + log10 tf t,d if tf t,d > 0 0 otherwise Gintarė Grigonytė 9/57

10 Inverse Document Frequency weight some terms are more relevant than others the document frequency df t is defined as the number of documents that t occurs in. We define the idf weight of term t as follows: idf t = log 10 N df t idf is a measure of the informativeness of the term. term df t idf t calpurnia 1 6 animal sunday fly 10,000 2 under 100,000 1 the 1,000,000 0 Gintarė Grigonytė 10/57

11 Collection frequency vs. Document frequency We could also use collection frequency. (count multiple occurrences of terms in documents) Word Collection frequency Document frequency INSURANCE TRY Which one is better? Gintarė Grigonytė 11/57

12 tf-idf weighting product of term frequency (tf) and inverse document frequency (idf) w t,d = (1 + log tf t,d ) log N df t best known weighting scheme in IR increases with number of occurrences within a document rarity of the term in collection Gintarė Grigonytė 12/57

13 Binary count weight matrix Anthony Julius The Hamlet Othello Macbeth... and Caesar Tempest Cleopatra ANTHONY BRUTUS CAESAR CALPURNIA CLEOPATRA MERCY WORSER each document represented by a real-valued vector of tf-idf weights R V. How do we rank documents according to relevance? Gintarė Grigonytė 13/57

14 Best-match retrieval models So far, we ve discussed exact-match models Today, we start discussing best-match models Best-match models predict the degree to which a document is relevant to a query Ideally, this would be expressed as RELEVANT(q,d) In practice, it is expressed as SIMILAR(q,d) How can we compute the similarity between q and d? Gintarė Grigonytė 14/57

15 Vector Space Model A vector space is defined by a set of linearly independent basis vectors The basis vectors correspond to the dimensions or directions of the vector space (2D, 3D,...) A vector is a point in a vector space and has length (from the origin to the point) and direction The basis vectors are linearly independent because knowing a vector s value on one dimension doesn t say anything about its value along another dimension Gintarė Grigonytė 15/57

16 Vector Space Model The vector space model ranks documents based on the vector-space similarity between the query vector and the document vector There are many ways to compute the similarity between two vectors One way is to compute the inner product Gintarė Grigonytė 16/57

17 Vector Space Model (source: J. Arguello) Gintarė Grigonytė 17/57

18 Vector Space Model What is more relevant to a query? A 50-word document which contains 5 of the query- terms? A 100-word document which contains 5 of the query-terms? The inner-product doesn t account for the fact that documents have widely varying lengths All things being equal, longer documents are more likely to have the query-terms So, the inner-product favors long documents Gintarė Grigonytė 18/57

19 Documents and queries as vectors Measure proximity between query and documents! Gintarė Grigonytė 19/57

20 Cosine similarity between query and document cos( q, d) = q q d d = V i=1 q i V i=1 q2 i d i V i=1 d i 2 q i is the tf-idf weight of term i in the query. d i is the tf-idf weight of term i in the document. q and d are the lengths of q and d. q/ q and d/ d are length-1 vectors (= normalized). Gintarė Grigonytė 20/57

21 Cosine similarity between query and document Gintarė Grigonytė 21/57

22 Predicted and true probability of relevance A problem with cosine normalization: (source: Lillian Lee) Gintarė Grigonytė 22/57

23 Pivot normalization Linear adjustment to compensate for true relevance (source: Lillian Lee) Gintarė Grigonytė 23/57

24 Gintarė Grigonytė 24/57

25 (source: J. Arguello) Gintarė Grigonytė 25/57

26 (source: J. Arguello) Gintarė Grigonytė 26/57

27 Gintarė Grigonytė 27/57 (source: J. Arguello)

28 (source: J. Arguello) Gintarė Grigonytė 28/57

29 (source: J. Arguello) Gintarė Grigonytė 29/57

30 (source: J. Arguello) Gintarė Grigonytė 30/57

31 (source: J. Arguello) Gintarė Grigonytė 31/57

32 Components of tf-idf weighting Term frequency Document frequency Normalization n (natural) tf t,d n (no) 1 n (none) 1 l (logarithm) 1 + log(tf t,d ) t (idf) log df N c (cosine) 1 t w 2 1 +w w M 2 a (augmented) b (boolean) tf t,d max t (tf t,d ) { 1 if tft,d > 0 0 otherwise p (prob idf) max{0, log N dft df t } u (pivoted unique) αu d +(1 α)pivot b (byte size) 1/CharLength α, α < 1 L (log ave) 1+log(tf t,d ) 1+log(ave t d (tf t,d )) different weightings for queries & documents qqq.ddd example: ltn.lnc: query: logarithmic tf, idf, no normalization document: logarithmic tf, no df weighting, cosine normalization Gintarė Grigonytė 32/57

33 tf-idf Example: ltn.lnc Query: best car insurance. Document: car insurance auto insurance. word query document product tf-raw tf-wght df idf weight tf-raw tf-wght weight n lized auto best car insurance Key to columns: tf-raw: raw (unweighted) term frequency, tf-wght: logarithmically weighted term frequency, df: document frequency, idf: inverse document frequency, weight: the final weight of the term in the query or document, n lized: document weights after cosine normalization, product: the product of final query weight and final document weight Gintarė Grigonytė 33/57

34 tf-idf Example: ltn.lnc Query: best car insurance. Document: car insurance auto insurance. word query document product tf-raw tf-wght df idf weight tf-raw tf-wght weight n lized auto 0 best 1 car 1 insurance 1 Key to columns: tf-raw: raw (unweighted) term frequency, tf-wght: logarithmically weighted term frequency, df: document frequency, idf: inverse document frequency, weight: the final weight of the term in the query or document, n lized: document weights after cosine normalization, product: the product of final query weight and final document weight Gintarė Grigonytė 33/57

35 tf-idf Example: ltn.lnc Query: best car insurance. Document: car insurance auto insurance. word query document product tf-raw tf-wght df idf weight tf-raw tf-wght weight n lized auto 0 1 best 1 0 car 1 1 insurance 1 2 Key to columns: tf-raw: raw (unweighted) term frequency, tf-wght: logarithmically weighted term frequency, df: document frequency, idf: inverse document frequency, weight: the final weight of the term in the query or document, n lized: document weights after cosine normalization, product: the product of final query weight and final document weight Gintarė Grigonytė 33/57

36 tf-idf Example: ltn.lnc Query: best car insurance. Document: car insurance auto insurance. word query document product tf-raw tf-wght df idf weight tf-raw tf-wght weight n lized auto best car insurance Key to columns: tf-raw: raw (unweighted) term frequency, tf-wght: logarithmically weighted term frequency, df: document frequency, idf: inverse document frequency, weight: the final weight of the term in the query or document, n lized: document weights after cosine normalization, product: the product of final query weight and final document weight Gintarė Grigonytė 33/57

37 tf-idf Example: ltn.lnc Query: best car insurance. Document: car insurance auto insurance. word query document product tf-raw tf-wght df idf weight tf-raw tf-wght weight n lized auto best car insurance Key to columns: tf-raw: raw (unweighted) term frequency, tf-wght: logarithmically weighted term frequency, df: document frequency, idf: inverse document frequency, weight: the final weight of the term in the query or document, n lized: document weights after cosine normalization, product: the product of final query weight and final document weight Gintarė Grigonytė 33/57

38 tf-idf Example: ltn.lnc Query: best car insurance. Document: car insurance auto insurance. word query document product tf-raw tf-wght df idf weight tf-raw tf-wght weight n lized auto best car insurance Key to columns: tf-raw: raw (unweighted) term frequency, tf-wght: logarithmically weighted term frequency, df: document frequency, idf: inverse document frequency, weight: the final weight of the term in the query or document, n lized: document weights after cosine normalization, product: the product of final query weight and final document weight Gintarė Grigonytė 33/57

39 tf-idf Example: ltn.lnc Query: best car insurance. Document: car insurance auto insurance. word query document product tf-raw tf-wght df idf weight tf-raw tf-wght weight n lized auto best car insurance Key to columns: tf-raw: raw (unweighted) term frequency, tf-wght: logarithmically weighted term frequency, df: document frequency, idf: inverse document frequency, weight: the final weight of the term in the query or document, n lized: document weights after cosine normalization, product: the product of final query weight and final document weight Gintarė Grigonytė 33/57

40 tf-idf Example: ltn.lnc Query: best car insurance. Document: car insurance auto insurance. word query document product tf-raw tf-wght df idf weight tf-raw tf-wght weight n lized auto best car insurance Key to columns: tf-raw: raw (unweighted) term frequency, tf-wght: logarithmically weighted term frequency, df: document frequency, idf: inverse document frequency, weight: the final weight of the term in the query or document, n lized: document weights after cosine normalization, product: the product of final query weight and final document weight Gintarė Grigonytė 33/57

41 tf-idf Example: ltn.lnc Query: best car insurance. Document: car insurance auto insurance. word query document product tf-raw tf-wght df idf weight tf-raw tf-wght weight n lized auto best car insurance Key to columns: tf-raw: raw (unweighted) term frequency, tf-wght: logarithmically weighted term frequency, df: document frequency, idf: inverse document frequency, weight: the final weight of the term in the query or document, n lized: document weights after cosine normalization, product: the product of final query weight and final document weight Gintarė Grigonytė 33/57

42 tf-idf Example: ltn.lnc Query: best car insurance. Document: car insurance auto insurance. word query document product tf-raw tf-wght df idf weight tf-raw tf-wght weight n lized auto best car insurance Key to columns: tf-raw: raw (unweighted) term frequency, tf-wght: logarithmically weighted term frequency, df: document frequency, idf: inverse document frequency, weight: the final weight of the term in the query or document, n lized: document weights after cosine normalization, product: the product of final query weight and final document weight Gintarė Grigonytė 33/57

43 tf-idf Example: ltn.lnc Query: best car insurance. Document: car insurance auto insurance. word query document product tf-raw tf-wght df idf weight tf-raw tf-wght weight n lized auto best car insurance Key to columns: tf-raw: raw (unweighted) term frequency, tf-wght: logarithmically weighted term frequency, df: document frequency, idf: inverse document frequency, weight: the final weight of the term in the query or document, n lized: document weights after cosine normalization, product: the product of final query weight and final document weight / / Gintarė Grigonytė 33/57

44 tf-idf Example: ltn.lnc Query: best car insurance. Document: car insurance auto insurance. word query document product tf-raw tf-wght df idf weight tf-raw tf-wght weight n lized auto best car insurance Key to columns: tf-raw: raw (unweighted) term frequency, tf-wght: logarithmically weighted term frequency, df: document frequency, idf: inverse document frequency, weight: the final weight of the term in the query or document, n lized: document weights after cosine normalization, product: the product of final query weight and final document weight Gintarė Grigonytė 33/57

45 tf-idf Example: ltn.lnc Query: best car insurance. Document: car insurance auto insurance. word query document product tf-raw tf-wght df idf weight tf-raw tf-wght weight n lized auto best car insurance Key to columns: tf-raw: raw (unweighted) term frequency, tf-wght: logarithmically weighted term frequency, df: document frequency, idf: inverse document frequency, weight: the final weight of the term in the query or document, n lized: document weights after cosine normalization, product: the product of final query weight and final document weight Final similarity score between query and document: i w qi w di = = 3.08 Gintarė Grigonytė 33/57

46 Second part today efficient scoring and ranking a complete IR system components architecture implementational issues Gintarė Grigonytė 34/57

47 Efficient scoring Query stays the same: need only (relative) scores for each document no matching keywords no score use inverted index file (instead of comparing all d s with q) run through postings for each term accumulate scores per document in posting lists rank J documents with positive scores much less than N (documents in collection) Gintarė Grigonytė 35/57

48 Term frequencies in the inverted index BRUTUS 1,2 7,3 83,1 87,2... CAESAR 1,1 5,1 13,1 17,1... CALPURNIA 7,1 8,2 40,1 97,3 term frequencies are easier to store than log-weights compute term weight on-the-fly (store idf in dictionary) intersect as usual (iterative calculation of document scores) don t need complete ranking top k retrieval Gintarė Grigonytė 36/57

49 Term frequencies in the inverted index BRUTUS 1,2 7,3 83,1 87,2... CAESAR 1,1 5,1 13,1 17,1... CALPURNIA 7,1 8,2 40,1 97,3 term frequencies are easier to store than log-weights compute term weight on-the-fly (store idf in dictionary) intersect as usual (iterative calculation of document scores) don t need complete ranking top k retrieval Efficient selection of top k documents How? Gintarė Grigonytė 36/57

50 Binary min heap for selecting top k binary min heap = binary tree in which each node s value is less than the values of its children always keep top k documents seen so far takes O(J log k) operations to construct read off k winners in O(k log k) steps Gintarė Grigonytė 37/57

51 Inexact top k retrieval bottleneck: cosine computation (J can still be big) instead of retrieving top k: retrieve k documents with scores close to top k use search heuristics Is this a bad thing to do? Gintarė Grigonytė 38/57

52 Generic approach find a set A of contenders with K < A << N return top k documents in A selecting A = pruning non-contendors index elimination champion lists static quality scores impact ordering cluster pruning Gintarė Grigonytė 39/57

53 Index elimination only consider high-idf query terms query: catcher in the rye accumulate scores for catcher and rye only Gintarė Grigonytė 40/57

54 Index elimination only consider high-idf query terms query: catcher in the rye accumulate scores for catcher and rye only only consider docs containing many query terms multi-term queries: scores for docs that contain at least a fixed proportion pf query terms (e.g., 3 out of 4) soft conjunction (early Google) easy to implement in postings traversal Gintarė Grigonytė 40/57

55 Champion lists For each term in dictionary: precompute r documents of heighest weight (in the postings of the term) Champion lists! query: only compute scores for documents in union of champion lists for all query terms r is chosen at indexing time might use different r for each term (more for rare terms?!) Gintarė Grigonytė 41/57

56 Static quality scores Idea 2: reorder posting lists according to expected relevance query independent quality of documents (authority) examples: a paper with many citations many bookmarks (del.icio.us,...) PageRank (!) Gintarė Grigonytė 42/57

57 Static quality scores assign quality ( goodness ) score g(d) to each d combine with relevance score (cos(q, d)): net-score(q, d) = g(d) + cos(q, d) might use other type of combination return top k documents according to net-score Gintarė Grigonytė 43/57

58 Static quality scores How does this help? postings are ordered by g(d) (still consistent order!) traverse postings and compute scores early termination is possible stop if minimal score cannot be improved time threshold threshold for goodness score can be combined with champion lists Gintarė Grigonytė 44/57

59 Other ideas High and low lists : for each term: high list ( = champion list) low list (other documents) use high lists first use low list if < k documents found Gintarė Grigonytė 45/57

60 Impact ordering compute scores only for documents with high wf t,d sort each posting by wf t,d non-consistent order of postings! solution 1: early termination: for each term stop after a fixed number of r documents stop when wft,d <threshold score documents in union of retrieved postings solution 2: sort terms by idf stop if document scores don t change much Gintarė Grigonytė 46/57

61 Cluster Pruning Pre-processing (clustering): Pick N docs at random (= leaders ) (random = fast + reflects distribution well) For every other doc, pre-compute nearest leader attach them to leader each leader has N followers Query processing: given query q: find nearest leader L seek k nearest docs among L followers Gintarė Grigonytė 47/57

62 Cluster Pruning Variants: clustering: attach documents to x nearest leaders querying: find y nearest leaders and consider their followers Gintarė Grigonytė 48/57

63 Building an IR system Components of an IR system tiered indexes query term proximity parsing and scoring functions... Gintarė Grigonytė 49/57

64 Tiered indexes Basic idea: Create several tiers of indexes, corresponding to importance of indexing terms Query processing: start with highest-tier index Stop if highest-tier index returns at least k results Go to next tier if we ve only found < k hits (see high & low lists!) Gintarė Grigonytė 50/57

65 Tiered indexes Basic idea: Create several tiers of indexes, corresponding to importance of indexing terms Query processing: start with highest-tier index Stop if highest-tier index returns at least k results Go to next tier if we ve only found < k hits (see high & low lists!) Could also be related to zones: Tier 1: Index of all titles Tier 2: Index of the rest of documents (Pages containing the search words in the title are better hits than pages containing the search words in the body of the text) Gintarė Grigonytė 50/57

66 Tiered index Tier 1 auto best Doc2 car Doc1 Doc3 insurance Doc2 Doc3 auto Tier 2 best Doc1 Doc3 car insurance auto Doc1 Tier 3 best car Doc2 insurance Gintarė Grigonytė 51/57

67 Query term proximity Free text queries: just a set of terms typed into the query box common on the web Users prefer docs in which query terms occur within close proximity of each other Let w be the smallest window in a doc containing all query terms, e.g., For the query strained mercy the smallest window in the doc The quality of mercy is not strained is 4 (words) Gintarė Grigonytė 52/57

68 Query parsers Free text query from user may fire one or more queries to the indexes, e.g. query rising interest rates 1. Run the query as a phrase query 2. If < k docs contain the phrase rising interest rates, run the two phrase queries rising interest and interest rates 3. If we still have < k docs, run the vector space query rising interest rates 4. Rank matching docs by vector space scoring This sequence is issued by a query parser How to combine scores? Tuning! (experts/machine-learning) Gintarė Grigonytė 53/57

69 IR system architecture Gintarė Grigonytė 54/57

70 Watch online lecture watch online lecture by Jeff Dean: Gintarė Grigonytė 55/57

71 Summary IR includes many components Document preprocessing (linguistic and otherwise) Positional indexes Tiered indexes Spelling correction k-gram indexes for wildcard queries and spelling correction Query parsing & Query processing Document scoring (including proximity...) Gintarė Grigonytė 56/57

72 What is next? lab on boolean and ranked retrieval small test collection create inverted index test boolean queries and ranked retrieval Gintarė Grigonytė 57/57

CEC login. Student Details Name SOLUTIONS

CEC login. Student Details Name SOLUTIONS Student Details Name SOLUTIONS CEC login Instructions You have roughly 1 minute per point, so schedule your time accordingly. There is only one correct answer per question. Good luck! Question 1. Searching

More information

Methods for Retrieving Alternative Contract Language Using a Prototype

Methods for Retrieving Alternative Contract Language Using a Prototype ICAIL 2017 Presentation Methods for Retrieving Alternative Contract Language Using a Prototype Silviu Pitis spitis@gatech.edu Retrieval by Prototype 1 2 3 Given a prototype Retrieve similar provisions

More information

Algorithmic Game Theory and Applications. Lecture 11: Games of Perfect Information

Algorithmic Game Theory and Applications. Lecture 11: Games of Perfect Information Algorithmic Game Theory and Applications Lecture 11: Games of Perfect Information Kousha Etessami finite games of perfect information Recall, a perfect information (PI) game has only 1 node per information

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning Monte Carlo Methods Mark Schmidt University of British Columbia Winter 2018 Last Time: Markov Chains We can use Markov chains for density estimation, p(x) = p(x 1 ) }{{} d p(x

More information

COMP417 Introduction to Robotics and Intelligent Systems. Reinforcement Learning - 2

COMP417 Introduction to Robotics and Intelligent Systems. Reinforcement Learning - 2 COMP417 Introduction to Robotics and Intelligent Systems Reinforcement Learning - 2 Speaker: Sandeep Manjanna Acklowledgement: These slides use material from Pieter Abbeel s, Dan Klein s and John Schulman

More information

Reinforcement Learning. Slides based on those used in Berkeley's AI class taught by Dan Klein

Reinforcement Learning. Slides based on those used in Berkeley's AI class taught by Dan Klein Reinforcement Learning Slides based on those used in Berkeley's AI class taught by Dan Klein Reinforcement Learning Basic idea: Receive feedback in the form of rewards Agent s utility is defined by the

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning Monte Carlo Methods Mark Schmidt University of British Columbia Winter 2019 Last Time: Markov Chains We can use Markov chains for density estimation, d p(x) = p(x 1 ) p(x }{{}

More information

Yao s Minimax Principle

Yao s Minimax Principle Complexity of algorithms The complexity of an algorithm is usually measured with respect to the size of the input, where size may for example refer to the length of a binary word describing the input,

More information

Classifying Press Releases and Company Relationships Based on Stock Performance

Classifying Press Releases and Company Relationships Based on Stock Performance Classifying Press Releases and Company Relationships Based on Stock Performance Mike Mintz Stanford University mintz@stanford.edu Ruka Sakurai Stanford University ruka.sakurai@gmail.com Nick Briggs Stanford

More information

Markov Decision Process

Markov Decision Process Markov Decision Process Human-aware Robotics 2018/02/13 Chapter 17.3 in R&N 3rd Ø Announcement: q Slides for this lecture are here: http://www.public.asu.edu/~yzhan442/teaching/cse471/lectures/mdp-ii.pdf

More information

Sublinear Time Algorithms Oct 19, Lecture 1

Sublinear Time Algorithms Oct 19, Lecture 1 0368.416701 Sublinear Time Algorithms Oct 19, 2009 Lecturer: Ronitt Rubinfeld Lecture 1 Scribe: Daniel Shahaf 1 Sublinear-time algorithms: motivation Twenty years ago, there was practically no investigation

More information

Session 3. Life/Health Insurance technical session

Session 3. Life/Health Insurance technical session SOA Big Data Seminar 13 Nov. 2018 Jakarta, Indonesia Session 3 Life/Health Insurance technical session Anilraj Pazhety Life Health Technical Session ANILRAJ PAZHETY MS (BUSINESS ANALYTICS), MBA, BE (CS)

More information

CS 343: Artificial Intelligence

CS 343: Artificial Intelligence CS 343: Artificial Intelligence Markov Decision Processes II Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC

More information

CS 106X Lecture 11: Sorting

CS 106X Lecture 11: Sorting CS 106X Lecture 11: Sorting Friday, February 3, 2017 Programming Abstractions (Accelerated) Winter 2017 Stanford University Computer Science Department Lecturer: Chris Gregg reading: Programming Abstractions

More information

Decision Trees An Early Classifier

Decision Trees An Early Classifier An Early Classifier Jason Corso SUNY at Buffalo January 19, 2012 J. Corso (SUNY at Buffalo) Trees January 19, 2012 1 / 33 Introduction to Non-Metric Methods Introduction to Non-Metric Methods We cover

More information

CS 188: Artificial Intelligence

CS 188: Artificial Intelligence CS 188: Artificial Intelligence Markov Decision Processes Dan Klein, Pieter Abbeel University of California, Berkeley Non-Deterministic Search 1 Example: Grid World A maze-like problem The agent lives

More information

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2017

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2017 ECON 459 Game Theory Lecture Notes Auctions Luca Anderlini Spring 2017 These notes have been used and commented on before. If you can still spot any errors or have any suggestions for improvement, please

More information

Lecture outline W.B.Powell 1

Lecture outline W.B.Powell 1 Lecture outline What is a policy? Policy function approximations (PFAs) Cost function approximations (CFAs) alue function approximations (FAs) Lookahead policies Finding good policies Optimizing continuous

More information

Homework #4. CMSC351 - Spring 2013 PRINT Name : Due: Thu Apr 16 th at the start of class

Homework #4. CMSC351 - Spring 2013 PRINT Name : Due: Thu Apr 16 th at the start of class Homework #4 CMSC351 - Spring 2013 PRINT Name : Due: Thu Apr 16 th at the start of class o Grades depend on neatness and clarity. o Write your answers with enough detail about your approach and concepts

More information

Lecture 17: More on Markov Decision Processes. Reinforcement learning

Lecture 17: More on Markov Decision Processes. Reinforcement learning Lecture 17: More on Markov Decision Processes. Reinforcement learning Learning a model: maximum likelihood Learning a value function directly Monte Carlo Temporal-difference (TD) learning COMP-424, Lecture

More information

CSCE 750, Fall 2009 Quizzes with Answers

CSCE 750, Fall 2009 Quizzes with Answers CSCE 750, Fall 009 Quizzes with Answers Stephen A. Fenner September 4, 011 1. Give an exact closed form for Simplify your answer as much as possible. k 3 k+1. We reduce the expression to a form we ve already

More information

Reinforcement Learning

Reinforcement Learning Reinforcement Learning Basic idea: Receive feedback in the form of rewards Agent s utility is defined by the reward function Must (learn to) act so as to maximize expected rewards Grid World The agent

More information

Stochastic Games and Bayesian Games

Stochastic Games and Bayesian Games Stochastic Games and Bayesian Games CPSC 532L Lecture 10 Stochastic Games and Bayesian Games CPSC 532L Lecture 10, Slide 1 Lecture Overview 1 Recap 2 Stochastic Games 3 Bayesian Games Stochastic Games

More information

Lecture l(x) 1. (1) x X

Lecture l(x) 1. (1) x X Lecture 14 Agenda for the lecture Kraft s inequality Shannon codes The relation H(X) L u (X) = L p (X) H(X) + 1 14.1 Kraft s inequality While the definition of prefix-free codes is intuitively clear, we

More information

TTIC An Introduction to the Theory of Machine Learning. The Adversarial Multi-armed Bandit Problem Avrim Blum.

TTIC An Introduction to the Theory of Machine Learning. The Adversarial Multi-armed Bandit Problem Avrim Blum. TTIC 31250 An Introduction to the Theory of Machine Learning The Adversarial Multi-armed Bandit Problem Avrim Blum Start with recap 1 Algorithm Consider the following setting Each morning, you need to

More information

COMP Analysis of Algorithms & Data Structures

COMP Analysis of Algorithms & Data Structures COMP 3170 - Analysis of Algorithms & Data Structures Shahin Kamali Binomial Heaps CLRS 6.1, 6.2, 6.3 University of Manitoba Priority queues A priority queue is an abstract data type formed by a set S of

More information

Max Registers, Counters and Monotone Circuits

Max Registers, Counters and Monotone Circuits James Aspnes 1 Hagit Attiya 2 Keren Censor 2 1 Yale 2 Technion Counters Model Collects Our goal: build a cheap counter for an asynchronous shared-memory system. Two operations: increment and read. Read

More information

Final exam solutions

Final exam solutions EE365 Stochastic Control / MS&E251 Stochastic Decision Models Profs. S. Lall, S. Boyd June 5 6 or June 6 7, 2013 Final exam solutions This is a 24 hour take-home final. Please turn it in to one of the

More information

Outline. Objective. Previous Results Our Results Discussion Current Research. 1 Motivation. 2 Model. 3 Results

Outline. Objective. Previous Results Our Results Discussion Current Research. 1 Motivation. 2 Model. 3 Results On Threshold Esteban 1 Adam 2 Ravi 3 David 4 Sergei 1 1 Stanford University 2 Harvard University 3 Yahoo! Research 4 Carleton College The 8th ACM Conference on Electronic Commerce EC 07 Outline 1 2 3 Some

More information

Algorithms and Networking for Computer Games

Algorithms and Networking for Computer Games Algorithms and Networking for Computer Games Chapter 4: Game Trees http://www.wiley.com/go/smed Game types perfect information games no hidden information two-player, perfect information games Noughts

More information

Linear functions Increasing Linear Functions. Decreasing Linear Functions

Linear functions Increasing Linear Functions. Decreasing Linear Functions 3.5 Increasing, Decreasing, Max, and Min So far we have been describing graphs using quantitative information. That s just a fancy way to say that we ve been using numbers. Specifically, we have described

More information

Extending MCTS

Extending MCTS Extending MCTS 2-17-16 Reading Quiz (from Monday) What is the relationship between Monte Carlo tree search and upper confidence bound applied to trees? a) MCTS is a type of UCT b) UCT is a type of MCTS

More information

Heaps. Heap/Priority queue. Binomial heaps: Advanced Algorithmics (4AP) Heaps Binary heap. Binomial heap. Jaak Vilo 2009 Spring

Heaps. Heap/Priority queue. Binomial heaps: Advanced Algorithmics (4AP) Heaps Binary heap. Binomial heap. Jaak Vilo 2009 Spring .0.00 Heaps http://en.wikipedia.org/wiki/category:heaps_(structure) Advanced Algorithmics (4AP) Heaps Jaak Vilo 00 Spring Binary heap http://en.wikipedia.org/wiki/binary_heap Binomial heap http://en.wikipedia.org/wiki/binomial_heap

More information

Expectimax Search Trees. CS 188: Artificial Intelligence Fall Expectimax Quantities. Expectimax Pseudocode. Expectimax Pruning?

Expectimax Search Trees. CS 188: Artificial Intelligence Fall Expectimax Quantities. Expectimax Pseudocode. Expectimax Pruning? CS 188: Artificial Intelligence Fall 2010 Expectimax Search Trees What if we don t know what the result of an action will be? E.g., In solitaire, next card is unknown In minesweeper, mine locations In

More information

CS4311 Design and Analysis of Algorithms. Lecture 14: Amortized Analysis I

CS4311 Design and Analysis of Algorithms. Lecture 14: Amortized Analysis I CS43 Design and Analysis of Algorithms Lecture 4: Amortized Analysis I About this lecture Given a data structure, amortized analysis studies in a sequence of operations, the average time to perform an

More information

Do Media Sentiments Reflect Economic Indices?

Do Media Sentiments Reflect Economic Indices? Do Media Sentiments Reflect Economic Indices? Munich, September, 1, 2010 Paul Hofmarcher, Kurt Hornik, Stefan Theußl WU Wien Hofmarcher/Hornik/Theußl Sentiment Analysis 1/15 I I II Text Mining Sentiment

More information

Heaps

Heaps AdvancedAlgorithmics (4AP) Heaps Jaak Vilo 2009 Spring Jaak Vilo MTAT.03.190 Text Algorithms 1 Heaps http://en.wikipedia.org/wiki/category:heaps_(structure) Binary heap http://en.wikipedia.org/wiki/binary_heap

More information

Lecture 2: Making Good Sequences of Decisions Given a Model of World. CS234: RL Emma Brunskill Winter 2018

Lecture 2: Making Good Sequences of Decisions Given a Model of World. CS234: RL Emma Brunskill Winter 2018 Lecture 2: Making Good Sequences of Decisions Given a Model of World CS234: RL Emma Brunskill Winter 218 Human in the loop exoskeleton work from Steve Collins lab Class Structure Last Time: Introduction

More information

CSEP 573: Artificial Intelligence

CSEP 573: Artificial Intelligence CSEP 573: Artificial Intelligence Markov Decision Processes (MDP)! Ali Farhadi Many slides over the course adapted from Luke Zettlemoyer, Dan Klein, Pieter Abbeel, Stuart Russell or Andrew Moore 1 Outline

More information

CS360 Homework 14 Solution

CS360 Homework 14 Solution CS360 Homework 14 Solution Markov Decision Processes 1) Invent a simple Markov decision process (MDP) with the following properties: a) it has a goal state, b) its immediate action costs are all positive,

More information

Multiple regression - a brief introduction

Multiple regression - a brief introduction Multiple regression - a brief introduction Multiple regression is an extension to regular (simple) regression. Instead of one X, we now have several. Suppose, for example, that you are trying to predict

More information

CSE 473: Artificial Intelligence

CSE 473: Artificial Intelligence CSE 473: Artificial Intelligence Markov Decision Processes (MDPs) Luke Zettlemoyer Many slides over the course adapted from Dan Klein, Stuart Russell or Andrew Moore 1 Announcements PS2 online now Due

More information

Markov Decision Processes

Markov Decision Processes Markov Decision Processes Robert Platt Northeastern University Some images and slides are used from: 1. CS188 UC Berkeley 2. AIMA 3. Chris Amato Stochastic domains So far, we have studied search Can use

More information

The exam is closed book, closed calculator, and closed notes except your three crib sheets.

The exam is closed book, closed calculator, and closed notes except your three crib sheets. CS 188 Spring 2016 Introduction to Artificial Intelligence Final V2 You have approximately 2 hours and 50 minutes. The exam is closed book, closed calculator, and closed notes except your three crib sheets.

More information

CS599: Algorithm Design in Strategic Settings Fall 2012 Lecture 6: Prior-Free Single-Parameter Mechanism Design (Continued)

CS599: Algorithm Design in Strategic Settings Fall 2012 Lecture 6: Prior-Free Single-Parameter Mechanism Design (Continued) CS599: Algorithm Design in Strategic Settings Fall 2012 Lecture 6: Prior-Free Single-Parameter Mechanism Design (Continued) Instructor: Shaddin Dughmi Administrivia Homework 1 due today. Homework 2 out

More information

Is Greedy Coordinate Descent a Terrible Algorithm?

Is Greedy Coordinate Descent a Terrible Algorithm? Is Greedy Coordinate Descent a Terrible Algorithm? Julie Nutini, Mark Schmidt, Issam Laradji, Michael Friedlander, Hoyt Koepke University of British Columbia Optimization and Big Data, 2015 Context: Random

More information

Non-Deterministic Search

Non-Deterministic Search Non-Deterministic Search MDP s 1 Non-Deterministic Search How do you plan (search) when your actions might fail? In general case, how do you plan, when the actions have multiple possible outcomes? 2 Example:

More information

Advanced Algorithmics (4AP) Heaps

Advanced Algorithmics (4AP) Heaps Advanced Algorithmics (4AP) Heaps Jaak Vilo 2009 Spring Jaak Vilo MTAT.03.190 Text Algorithms 1 Heaps http://en.wikipedia.org/wiki/category:heaps_(structure) Binary heap http://en.wikipedia.org/wiki/binary

More information

Successor. CS 361, Lecture 19. Tree-Successor. Outline

Successor. CS 361, Lecture 19. Tree-Successor. Outline Successor CS 361, Lecture 19 Jared Saia University of New Mexico The successor of a node x is the node that comes after x in the sorted order determined by an in-order tree walk. If all keys are distinct,

More information

$tock Forecasting using Machine Learning

$tock Forecasting using Machine Learning $tock Forecasting using Machine Learning Greg Colvin, Garrett Hemann, and Simon Kalouche Abstract We present an implementation of 3 different machine learning algorithms gradient descent, support vector

More information

Decision Theory: Value Iteration

Decision Theory: Value Iteration Decision Theory: Value Iteration CPSC 322 Decision Theory 4 Textbook 9.5 Decision Theory: Value Iteration CPSC 322 Decision Theory 4, Slide 1 Lecture Overview 1 Recap 2 Policies 3 Value Iteration Decision

More information

MS-E2114 Investment Science Lecture 4: Applied interest rate analysis

MS-E2114 Investment Science Lecture 4: Applied interest rate analysis MS-E2114 Investment Science Lecture 4: Applied interest rate analysis A. Salo, T. Seeve Systems Analysis Laboratory Department of System Analysis and Mathematics Aalto University, School of Science Overview

More information

CS221 / Spring 2018 / Sadigh. Lecture 9: Games I

CS221 / Spring 2018 / Sadigh. Lecture 9: Games I CS221 / Spring 2018 / Sadigh Lecture 9: Games I Course plan Search problems Markov decision processes Adversarial games Constraint satisfaction problems Bayesian networks Reflex States Variables Logic

More information

Lecture 5. 1 Online Learning. 1.1 Learning Setup (Perspective of Universe) CSCI699: Topics in Learning & Game Theory

Lecture 5. 1 Online Learning. 1.1 Learning Setup (Perspective of Universe) CSCI699: Topics in Learning & Game Theory CSCI699: Topics in Learning & Game Theory Lecturer: Shaddin Dughmi Lecture 5 Scribes: Umang Gupta & Anastasia Voloshinov In this lecture, we will give a brief introduction to online learning and then go

More information

CSE 100: TREAPS AND RANDOMIZED SEARCH TREES

CSE 100: TREAPS AND RANDOMIZED SEARCH TREES CSE 100: TREAPS AND RANDOMIZED SEARCH TREES Midterm Review Practice Midterm covered during Sunday discussion Today Run time analysis of building the Huffman tree AVL rotations and treaps Huffman s algorithm

More information

Chapter 16. Binary Search Trees (BSTs)

Chapter 16. Binary Search Trees (BSTs) Chapter 16 Binary Search Trees (BSTs) Search trees are tree-based data structures that can be used to store and search for items that satisfy a total order. There are many types of search trees designed

More information

15-451/651: Design & Analysis of Algorithms November 9 & 11, 2015 Lecture #19 & #20 last changed: November 10, 2015

15-451/651: Design & Analysis of Algorithms November 9 & 11, 2015 Lecture #19 & #20 last changed: November 10, 2015 15-451/651: Design & Analysis of Algorithms November 9 & 11, 2015 Lecture #19 & #20 last changed: November 10, 2015 Last time we looked at algorithms for finding approximately-optimal solutions for NP-hard

More information

Tutorial. Morningstar DirectSM. Quick Start Guide

Tutorial. Morningstar DirectSM. Quick Start Guide April 2008 Software Tutorial Morningstar DirectSM Quick Start Guide Table of Contents Quick Start Guide Getting Started with Morningstar Direct Defining an Investment Lineup or Watch List Generating a

More information

Stochastic Games and Bayesian Games

Stochastic Games and Bayesian Games Stochastic Games and Bayesian Games CPSC 532l Lecture 10 Stochastic Games and Bayesian Games CPSC 532l Lecture 10, Slide 1 Lecture Overview 1 Recap 2 Stochastic Games 3 Bayesian Games 4 Analyzing Bayesian

More information

15-451/651: Design & Analysis of Algorithms October 23, 2018 Lecture #16: Online Algorithms last changed: October 22, 2018

15-451/651: Design & Analysis of Algorithms October 23, 2018 Lecture #16: Online Algorithms last changed: October 22, 2018 15-451/651: Design & Analysis of Algorithms October 23, 2018 Lecture #16: Online Algorithms last changed: October 22, 2018 Today we ll be looking at finding approximately-optimal solutions for problems

More information

Student Guide: RWC Simulation Lab. Free Market Educational Services: RWC Curriculum

Student Guide: RWC Simulation Lab. Free Market Educational Services: RWC Curriculum Free Market Educational Services: RWC Curriculum Student Guide: RWC Simulation Lab Table of Contents Getting Started... 4 Preferred Browsers... 4 Register for an Account:... 4 Course Key:... 4 The Student

More information

Some Characteristics of Data

Some Characteristics of Data Some Characteristics of Data Not all data is the same, and depending on some characteristics of a particular dataset, there are some limitations as to what can and cannot be done with that data. Some key

More information

Lecture 9: Classification and Regression Trees

Lecture 9: Classification and Regression Trees Lecture 9: Classification and Regression Trees Advanced Applied Multivariate Analysis STAT 2221, Spring 2015 Sungkyu Jung Department of Statistics, University of Pittsburgh Xingye Qiao Department of Mathematical

More information

Lattice Model of System Evolution. Outline

Lattice Model of System Evolution. Outline Lattice Model of System Evolution Richard de Neufville Professor of Engineering Systems and of Civil and Environmental Engineering MIT Massachusetts Institute of Technology Lattice Model Slide 1 of 48

More information

Multistage Stochastic Programming

Multistage Stochastic Programming IE 495 Lecture 21 Multistage Stochastic Programming Prof. Jeff Linderoth April 16, 2003 April 16, 2002 Stochastic Programming Lecture 21 Slide 1 Outline HW Fixes Multistage Stochastic Programming Modeling

More information

An Experimental Study of the Behaviour of the Proxel-Based Simulation Algorithm

An Experimental Study of the Behaviour of the Proxel-Based Simulation Algorithm An Experimental Study of the Behaviour of the Proxel-Based Simulation Algorithm Sanja Lazarova-Molnar, Graham Horton Otto-von-Guericke-Universität Magdeburg Abstract The paradigm of the proxel ("probability

More information

Finance 197. Simple One-time Interest

Finance 197. Simple One-time Interest Finance 197 Finance We have to work with money every day. While balancing your checkbook or calculating your monthly expenditures on espresso requires only arithmetic, when we start saving, planning for

More information

Risk-neutral Binomial Option Valuation

Risk-neutral Binomial Option Valuation Risk-neutral Binomial Option Valuation Main idea is that the option price now equals the expected value of the option price in the future, discounted back to the present at the risk free rate. Assumes

More information

Issues. Senate (Total = 100) Senate Group 1 Y Y N N Y 32 Senate Group 2 Y Y D N D 16 Senate Group 3 N N Y Y Y 30 Senate Group 4 D Y N D Y 22

Issues. Senate (Total = 100) Senate Group 1 Y Y N N Y 32 Senate Group 2 Y Y D N D 16 Senate Group 3 N N Y Y Y 30 Senate Group 4 D Y N D Y 22 1. Every year, the United States Congress must approve a budget for the country. In order to be approved, the budget must get a majority of the votes in the Senate, a majority of votes in the House, and

More information

Pattern Recognition Chapter 5: Decision Trees

Pattern Recognition Chapter 5: Decision Trees Pattern Recognition Chapter 5: Decision Trees Asst. Prof. Dr. Chumphol Bunkhumpornpat Department of Computer Science Faculty of Science Chiang Mai University Learning Objectives How decision trees are

More information

CSE 417 Algorithms. Huffman Codes: An Optimal Data Compression Method

CSE 417 Algorithms. Huffman Codes: An Optimal Data Compression Method CSE 417 Algorithms Huffman Codes: An Optimal Data Compression Method 1 Compression Example 100k file, 6 letter alphabet: a 45% b 13% c 12% d 16% e 9% f 5% File Size: ASCII, 8 bits/char: 800kbits 2 3 >

More information

Basic Framework. About this class. Rewards Over Time. [This lecture adapted from Sutton & Barto and Russell & Norvig]

Basic Framework. About this class. Rewards Over Time. [This lecture adapted from Sutton & Barto and Russell & Norvig] Basic Framework [This lecture adapted from Sutton & Barto and Russell & Norvig] About this class Markov Decision Processes The Bellman Equation Dynamic Programming for finding value functions and optimal

More information

SOLVING ROBUST SUPPLY CHAIN PROBLEMS

SOLVING ROBUST SUPPLY CHAIN PROBLEMS SOLVING ROBUST SUPPLY CHAIN PROBLEMS Daniel Bienstock Nuri Sercan Özbay Columbia University, New York November 13, 2005 Project with Lucent Technologies Optimize the inventory buffer levels in a complicated

More information

Health & Safety for Activities. Warning! Getting this wrong could cost YOU unbelievable amounts of money if you are sued for negligence

Health & Safety for Activities. Warning! Getting this wrong could cost YOU unbelievable amounts of money if you are sued for negligence Health & Safety for Activities Warning! Getting this wrong could cost YOU unbelievable amounts of money if you are sued for negligence Why??? Think of three good reasons why getting your H&S in order is

More information

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 15 Adaptive Huffman Coding Part I Huffman code are optimal for a

More information

A t S + b r t T B (h i + 1) (t S + t T ) C h i (t S + t T ) + t S + b t T D (h i + n) (t S + t T )

A t S + b r t T B (h i + 1) (t S + t T ) C h i (t S + t T ) + t S + b t T D (h i + n) (t S + t T ) Suppose we have a primary B+-tree index where the leaves contain search keys and RIDs, and the RIDs point to the records in a file that is ordered on the index search key. Assume that the blocks in the

More information

Expectimax Search Trees. CS 188: Artificial Intelligence Fall Expectimax Example. Expectimax Pseudocode. Expectimax Pruning?

Expectimax Search Trees. CS 188: Artificial Intelligence Fall Expectimax Example. Expectimax Pseudocode. Expectimax Pruning? CS 188: Artificial Intelligence Fall 2011 Expectimax Search Trees What if we don t know what the result of an action will be? E.g., In solitaire, next card is unknown In minesweeper, mine locations In

More information

Want more background and training tips?

Want more background and training tips? Learner Outcomes Outcome #1: Participants will be able to explain what credit is. Outcome #2: Participants will be able to explain why using a credit card is a form of borrowing. Outcome #3: Participants

More information

CS 188: Artificial Intelligence Fall 2011

CS 188: Artificial Intelligence Fall 2011 CS 188: Artificial Intelligence Fall 2011 Lecture 7: Expectimax Search 9/15/2011 Dan Klein UC Berkeley Many slides over the course adapted from either Stuart Russell or Andrew Moore 1 Expectimax Search

More information

CS 188: Artificial Intelligence

CS 188: Artificial Intelligence CS 188: Artificial Intelligence Markov Decision Processes Dan Klein, Pieter Abbeel University of California, Berkeley Non Deterministic Search Example: Grid World A maze like problem The agent lives in

More information

client user GUIDE 2011

client user GUIDE 2011 client user GUIDE 2011 STEP ACTION Accessing Risk Register 1. Type https://www.scm rms.ca/riskregister/login.aspx 2. Click in the Username field on the Risk Register home page. 3. Type your Username and

More information

Design and Analysis of Algorithms 演算法設計與分析. Lecture 8 November 16, 2016 洪國寶

Design and Analysis of Algorithms 演算法設計與分析. Lecture 8 November 16, 2016 洪國寶 Design and Analysis of Algorithms 演算法設計與分析 Lecture 8 November 6, 206 洪國寶 Outline Review Amortized analysis Advanced data structures Binary heaps Binomial heaps Fibonacci heaps Data structures for disjoint

More information

6.231 DYNAMIC PROGRAMMING LECTURE 8 LECTURE OUTLINE

6.231 DYNAMIC PROGRAMMING LECTURE 8 LECTURE OUTLINE 6.231 DYNAMIC PROGRAMMING LECTURE 8 LECTURE OUTLINE Suboptimal control Cost approximation methods: Classification Certainty equivalent control: An example Limited lookahead policies Performance bounds

More information

ORF 307: Lecture 12. Linear Programming: Chapter 11: Game Theory

ORF 307: Lecture 12. Linear Programming: Chapter 11: Game Theory ORF 307: Lecture 12 Linear Programming: Chapter 11: Game Theory Robert J. Vanderbei April 3, 2018 Slides last edited on April 3, 2018 http://www.princeton.edu/ rvdb Game Theory John Nash = A Beautiful

More information

Financial Market Models. Lecture 1. One-period model of financial markets & hedging problems. Imperial College Business School

Financial Market Models. Lecture 1. One-period model of financial markets & hedging problems. Imperial College Business School Financial Market Models Lecture One-period model of financial markets & hedging problems One-period model of financial markets a 4 2a 3 3a 3 a 3 -a 4 2 Aims of section Introduce one-period model with finite

More information

Lending Club Loan Portfolio Optimization Fred Robson (frobson), Chris Lucas (cflucas)

Lending Club Loan Portfolio Optimization Fred Robson (frobson), Chris Lucas (cflucas) CS22 Artificial Intelligence Stanford University Autumn 26-27 Lending Club Loan Portfolio Optimization Fred Robson (frobson), Chris Lucas (cflucas) Overview Lending Club is an online peer-to-peer lending

More information

StockFinder Workbook. Fast and flexible sorting and rule-based scanning. Charting with the largest selection of indicators available

StockFinder Workbook. Fast and flexible sorting and rule-based scanning. Charting with the largest selection of indicators available StockFinder Workbook revised Apr 23, 2009 Charting with the largest selection of indicators available Fast and flexible sorting and rule-based scanning Everything you need to make your own decisions StockFinder

More information

1 Solutions to Tute09

1 Solutions to Tute09 s to Tute0 Questions 4. - 4. are straight forward. Q. 4.4 Show that in a binary tree of N nodes, there are N + NULL pointers. Every node has outgoing pointers. Therefore there are N pointers. Each node,

More information

CSE 417 Dynamic Programming (pt 2) Look at the Last Element

CSE 417 Dynamic Programming (pt 2) Look at the Last Element CSE 417 Dynamic Programming (pt 2) Look at the Last Element Reminders > HW4 is due on Friday start early! if you run into problems loading data (date parsing), try running java with Duser.country=US Duser.language=en

More information

Basic Procedure for Histograms

Basic Procedure for Histograms Basic Procedure for Histograms 1. Compute the range of observations (min. & max. value) 2. Choose an initial # of classes (most likely based on the range of values, try and find a number of classes that

More information

Exercise List: Proving convergence of the (Stochastic) Gradient Descent Method for the Least Squares Problem.

Exercise List: Proving convergence of the (Stochastic) Gradient Descent Method for the Least Squares Problem. Exercise List: Proving convergence of the (Stochastic) Gradient Descent Method for the Least Squares Problem. Robert M. Gower. October 3, 07 Introduction This is an exercise in proving the convergence

More information

ECS171: Machine Learning

ECS171: Machine Learning ECS171: Machine Learning Lecture 15: Tree-based Algorithms Cho-Jui Hsieh UC Davis March 7, 2018 Outline Decision Tree Random Forest Gradient Boosted Decision Tree (GBDT) Decision Tree Each node checks

More information

COSC160: Data Structures Binary Trees. Jeremy Bolton, PhD Assistant Teaching Professor

COSC160: Data Structures Binary Trees. Jeremy Bolton, PhD Assistant Teaching Professor COSC160: Data Structures Binary Trees Jeremy Bolton, PhD Assistant Teaching Professor Outline I. Binary Trees I. Implementations I. Memory Management II. Binary Search Tree I. Operations Binary Trees A

More information

Introduction to Artificial Intelligence Spring 2019 Note 2

Introduction to Artificial Intelligence Spring 2019 Note 2 CS 188 Introduction to Artificial Intelligence Spring 2019 Note 2 These lecture notes are heavily based on notes originally written by Nikhil Sharma. Games In the first note, we talked about search problems

More information

Learning TradeStation. News, Time & Sales, Research, Browser, and Ticker Bar

Learning TradeStation. News, Time & Sales, Research, Browser, and Ticker Bar Learning TradeStation News, Time & Sales, Research, Browser, and Ticker Bar Important Information No offer or solicitation to buy or sell securities, securities derivative or futures products of any kind,

More information

On the Optimality of a Family of Binary Trees Techical Report TR

On the Optimality of a Family of Binary Trees Techical Report TR On the Optimality of a Family of Binary Trees Techical Report TR-011101-1 Dana Vrajitoru and William Knight Indiana University South Bend Department of Computer and Information Sciences Abstract In this

More information

Decision Theory: Sequential Decisions

Decision Theory: Sequential Decisions Decision Theory: CPSC 322 Decision Theory 2 Textbook 9.3 Decision Theory: CPSC 322 Decision Theory 2, Slide 1 Lecture Overview 1 Recap 2 Decision Theory: CPSC 322 Decision Theory 2, Slide 2 Decision Variables

More information

91.420/543: Artificial Intelligence UMass Lowell CS Fall 2010

91.420/543: Artificial Intelligence UMass Lowell CS Fall 2010 91.420/543: Artificial Intelligence UMass Lowell CS Fall 2010 Lecture 17 & 18: Markov Decision Processes Oct 12 13, 2010 A subset of Lecture 9 slides from Dan Klein UC Berkeley Many slides over the course

More information

yuimagui: A graphical user interface for the yuima package. User Guide yuimagui v1.0

yuimagui: A graphical user interface for the yuima package. User Guide yuimagui v1.0 yuimagui: A graphical user interface for the yuima package. User Guide yuimagui v1.0 Emanuele Guidotti, Stefano M. Iacus and Lorenzo Mercuri February 21, 2017 Contents 1 yuimagui: Home 3 2 yuimagui: Data

More information

Financial Optimization ISE 347/447. Lecture 15. Dr. Ted Ralphs

Financial Optimization ISE 347/447. Lecture 15. Dr. Ted Ralphs Financial Optimization ISE 347/447 Lecture 15 Dr. Ted Ralphs ISE 347/447 Lecture 15 1 Reading for This Lecture C&T Chapter 12 ISE 347/447 Lecture 15 2 Stock Market Indices A stock market index is a statistic

More information