STP Problem Set 2 Solutions
|
|
- Adrian Hall
- 5 years ago
- Views:
Transcription
1 STP Problem Set 2 Solutions 3.2.) Suppose that the inventory model is modified so that orders may be backlogged with a cost of b(u) when u units are backlogged for one period. We assume that revenue is received at the end of the period in which orders are placed and that backlogging costs are incurred at the beginning of each month. a.) The modified model can be formulated as a MDP as follows: Decision epochs: T = 1, 2,, N}. States: S =, 1, 0, 1,, M}, where s t < 0 if there are unfilled orders at the beginning of the t th period and s t > 0 if there is stock left over from the preceding period. Actions: A s = 0, 1,, M s}, assuming that there is no constraint on the amount of product that can be ordered to fill backlogged orders, but that the warehouse capacity limits the amount of stock that can be held until the end of the month. Transition probabilities: If p j, j 0 is the probability that j new orders are received during a period, then the transition probabilities are: p s+a j if j s + a p t (j s, a) = 0 if j > s + a. If we assume that newly ordered inventory is used to fill any backlogged orders as soon as it arrives at the beginning of each period, then r t (s, a) = f O(a) h([s + a] + ) b([s + a] ), where [x] + = x 0 is the positive part of x, [x] = x 0 is the negative part of x, and f = p j f(j). c.) If we assume that payment for an order is not received until the end of the month in which it is filled, then the rewards must be modified so that j=1 r t (s, a) = F (s, a) O(a) h([s + a] + ) b([s + a] ) where the expected revenue is equal to s+a 1 j=0 p j f(j) + f(s + a) j=s+a p j if s + a s 0 F (s, a) = s+a 1 j=0 p j f( s + j) + f(a) j=s+a p j if s + a 0 > s f(a) if s + a < 0, with f(u) equal to the actual revenue when u units are sold. 1
2 3.8) At each round of a game of chance, a player may either pay a fee of C units and play the game or else quit and depart with his current fortune. Upon paying the fee, he receives a random payoff with (discrete) distribution F ( ). a.) Here is one formulation as an optional stopping problem: Decision epochs: T = 1, 2,, N}. States: S = R }, where s t R are the player s cumulative winnings less costs at the start of the t th period if the player chooses to play the game and s t = if the player has chosen to quit. Here I am assuming that the gambler can run up an arbitrarily large debt. Actions: A s = P, Q} if s R Q} if s =, where P indicates that the player has chosen to play the game and Q indicates that the player has chosen to quit. Transition probabilities: p t (j s, a) = F (j s + C) if s, j R and a = P 1 if j = and a = Q. 0 if s R and a = P r t (s, a) = s if s R and a = Q 0 if s = and a = Q. This formulation assumes that the player does not receive the reward until he quits the game. However, other choices of the rewards are also possible, e.g., r t (s, a) = F C if a = P 0 if a = Q, where F is the mean of the payoff distribution. b.) If C = 1 and F (2) = p and F (0) = 1 p, where p [0, 1], then F = 2p and the transition probabilities are p if j = s + 1 R and a = P p t (j s, a) = 1 p if j = s 1 R and a = P 1 if j = and a = Q. 2
3 3.11) We reconsider the secretary problem when the aim is to minimize the expected rank of the candidate that receives the best offer. This can be formulated as an optional stopping problem. Decision epochs: T = 1, 2,, N}. States: S = 1, 2,, N} }. Here s t 1,, N} specifies the rank of the t th candidate relative to candidates 1,, t 1 and we require s 1 = 1. Actions: A s = C, Q} if 1 s N Q} if s =, where C indicates that the employer chooses to interview the next candidate and Q indicates that they choose to hire the current candidate. Transition probabilities: 1 t+1 if 1 s t, 1 j t + 1, and a = C p t (j s, a) = 1 if j = and a = Q 0 otherwise. where 0 if a = C r t (s, a) = Φ(s, t) if a = Q and s 1,, N} 0 if s =, Φ(s, t) = N i=1 it N ( i 1 )( N i s 1 t s ( N 1 t 1 is the expected absolute rank of the t th candidate given that their relative rank amongst the first t candidates is s 1,, t}. To derive the formula for Φ(s, t), let X t denote the relative rank of the t th candidate amongst the first t candidates and let Z t be the absolute rank of this candidate. Then, by using Bayes formula, the conditional probability mass function of Z t given X t = s 1,, t} is equal to since ) ). P(Z t = i X t = s) = P(Z t = i) P(X t = s Z t = k) P(Z t = i) ( i 1 )( N i ) / ( N 1 ) = 1 s 1 t s t 1 N 1/t P(Z t = i) = 1 N, 1 i N P(X t = s) = 1 t, 1 s t 3
4 P(X t = s Z t = i) = P ( the first t 1 objects sampled include s 1 objects with absolute rank < i and t s objects with absolute rank > i ) ) = ( i 1 )( N i s 1 t s ( N 1 t 1 ), the last expression being the probability that a hypergeometric random variable with parameters (N 1, s 1, t 1) is equal to i 1. It follows that by choosing a stopping rule that minimizes the total expected cost, we will also be minimizing the expected rank of the candidate that is offered the position. 3.18) At the start of each week a worker receives a wage offer of w units per week. He can either choose to accept employment at that wage for the entire week or he can seek alternative employment. If he decides to work during the current week, then with probability p he will have the same wage offer available at the start of the next week and with probability 1 p he will be laid off and be unable to seek new employment during that week. If he chooses to seek alternative employment, he will receive no wage during the current week, but will receive a wage offer of w during the next week, where w is chosen from the distribution p t (w w) and w is his most recent wage offer. The utility of wage w during week t is Ψ t (w). There is some ambiguity in the statement of the problem, as it is not entirely clear whether being laid off prevents the worker from seeking new employment for that week only or also prevents them from seeking alternative work for the next week. I will assume the latter, so that a worker that was laid off in week t 1 must seek alternative work during week t, which then leads to a job offer during week t + 1. I will also assume that the wage offer is w = 0 when the worker is laid off. Of course, other interpretations are possible and these will lead to somewhat different Markov decision processes. a.) This can be formulated as the following Markov decision process: Decision epochs: T = 1, 2,, N}, N. States: S = [0, ) e, l}, where a state s = (w, y) specifies the current wage offer w and y = e if the worker is either employed or seeking employment and y = l if the worker has just been laid off. Actions: A (w,e) = W, S} if w > 0, A (0,e) = S}, and A (w,l) = N} where a = W if the worker chooses to work, a = S if the worker chooses to seek alternative employment, and a = N if the worker has just been laid off, in which case they can neither work nor seek employment for the duration of that week. r t ((w, y), a) = Ψ t (w) if w > 0, y = e and a = W 0 otherwise 4
5 Transition probabilities: p if w = w, y = y = e and a = W ( p t (w, y ) (w, y), a ) 1 p if w = 0, w > 0, y = l, y = e and a = W = 1 if w = w = 0, y = e, y = l and a = N p t (w w) if y = y = e and a = S 0 otherwise. b.) Alternatively, this model can be formulated as a restless bandit. Let S 1 [0, ) l} and S 2 = [0, ) and suppose that the transition probabilities are given by p if s 1 = s 1 = s 2 = s 2 and a = 1 (1 p) if s 1 = l, s 2 = 0, s 1 = s 2 0 and a = 1 p t ((s 1, s 2) (s 1, s 2 ), a) = 1 if s 1 = s 2 = s 2 = 0, s 1 = l, and a 1, 2} p t (s 2 s 2) if s 1 = s 2, s 1 l, s 2 0 and a = 2 0 otherwise while the rewards are r t ((s 1, s 2 ), a) = Ψ t (s 1 ) if a = 1 and s 1 l 0 otherwise. As far as I can tell, this model cannot be formulated as a classical bandit problem, in which only the chosen process advances in each decision epoch. 3.26) The lazy, adaptable lion (Mangel and Clark, 1986). This problem is concerned with the effect of group size on the trade-off between hunting success and the amount of meat available to the members of a cooperatively hunting group. Larger group sizes increase the probability of a successful hunt, but result in less meat per individual. The following is a formulation of this problem as a Markov decision process. Decision epochs: T = 1, 2,, T }. States: S = [0, 30], where s specifies the energy reserves of an individual lion, measured in units of kilograms of meat. We assume that these are limited by gut capacity (30 kg) and thus we are ignoring fat reserves. We also assume that the lion is dead if s = 0. Actions: A s = 0, 1, 2, 3, 4, 5, 6} if s 0.5 and A s = 0} if s < 0.5. Here a = 0 means that the lion either chooses not to or is unable to hunt, a = 1 means that they hunt on their own, and a = n 2,, 6} means that they hunt in a group of size n. Since a hunt consumes 0.5 kg of energy, I assume that lions with energy reserves that are less than this are physically unable to hunt. On average, prey yield 164 kg of meat and this is shared equally between the members of a hunting party. Since the aim is to maximize the probability that the lion survives for T days, we set r t (s, a) = 0 for t = 1,, T 1 for all s and a, and r T (s) = 1 if s > 0 and r T (0) = 0 otherwise. 5
6 Transition probabilities: 1 if j = (s 6) + and a = 0 p(a) if j = min30, (s /a) + } and a > 0 p t (j s, a) = 1 p(a) if j = (s 6.5) + and a > 0 0 otherwise, where p(1) = 0.15, p(2) = 0.33, p(3) = 0.37, p(4) = 0.40, p(5) = 0.42 and p(6) =
STP Problem Set 3 Solutions
STP 425 - Problem Set 3 Solutions 4.4) Consider the separable sequential allocation problem introduced in Sections 3.3.3 and 4.6.3, where the goal is to maximize the sum subject to the constraints f(x
More informationProbability Theory and Simulation Methods. April 9th, Lecture 20: Special distributions
April 9th, 2018 Lecture 20: Special distributions Week 1 Chapter 1: Axioms of probability Week 2 Chapter 3: Conditional probability and independence Week 4 Chapters 4, 6: Random variables Week 9 Chapter
More informationDepartment of Statistics University of Warwick
Optimal ping Probability: the science of chance Department of Statistics University of Warwick Postgraduate Lunches 4th March 2013 In this talk... Optimal ping Optimal ping problems A simple example Every
More informationSection 7.5 The Normal Distribution. Section 7.6 Application of the Normal Distribution
Section 7.6 Application of the Normal Distribution A random variable that may take on infinitely many values is called a continuous random variable. A continuous probability distribution is defined by
More informationQ1. [?? pts] Search Traces
CS 188 Spring 2010 Introduction to Artificial Intelligence Midterm Exam Solutions Q1. [?? pts] Search Traces Each of the trees (G1 through G5) was generated by searching the graph (below, left) with a
More informationEE266 Homework 5 Solutions
EE, Spring 15-1 Professor S. Lall EE Homework 5 Solutions 1. A refined inventory model. In this problem we consider an inventory model that is more refined than the one you ve seen in the lectures. The
More informationPrediction Market Prices as Martingales: Theory and Analysis. David Klein Statistics 157
Prediction Market Prices as Martingales: Theory and Analysis David Klein Statistics 157 Introduction With prediction markets growing in number and in prominence in various domains, the construction of
More informationTABLE OF CONTENTS - VOLUME 2
TABLE OF CONTENTS - VOLUME 2 CREDIBILITY SECTION 1 - LIMITED FLUCTUATION CREDIBILITY PROBLEM SET 1 SECTION 2 - BAYESIAN ESTIMATION, DISCRETE PRIOR PROBLEM SET 2 SECTION 3 - BAYESIAN CREDIBILITY, DISCRETE
More informationBinomial Distributions
Binomial Distributions (aka Bernouli s Trials) Chapter 8 Binomial Distribution an important class of probability distributions, which occur under the following Binomial Setting (1) There is a number n
More informationPOMDPs: Partially Observable Markov Decision Processes Advanced AI
POMDPs: Partially Observable Markov Decision Processes Advanced AI Wolfram Burgard Types of Planning Problems Classical Planning State observable Action Model Deterministic, accurate MDPs observable stochastic
More informationCMSC 858F: Algorithmic Game Theory Fall 2010 Introduction to Algorithmic Game Theory
CMSC 858F: Algorithmic Game Theory Fall 2010 Introduction to Algorithmic Game Theory Instructor: Mohammad T. Hajiaghayi Scribe: Hyoungtae Cho October 13, 2010 1 Overview In this lecture, we introduce the
More informationLearning Goals: * Determining the expected value from a probability distribution. * Applying the expected value formula to solve problems.
Learning Goals: * Determining the expected value from a probability distribution. * Applying the expected value formula to solve problems. The following are marks from assignments and tests in a math class.
More informationNormal Cumulative Distribution Function (CDF)
The Normal Model 2 Solutions COR1-GB.1305 Statistics and Data Analysis Normal Cumulative Distribution Function (CDF 1. Suppose that an automobile muffler is designed so that its lifetime (in months is
More informationName. Answers Discussion Final Exam, Econ 171, March, 2012
Name Answers Discussion Final Exam, Econ 171, March, 2012 1) Consider the following strategic form game in which Player 1 chooses the row and Player 2 chooses the column. Both players know that this is
More informationFDPE Microeconomics 3 Spring 2017 Pauli Murto TA: Tsz-Ning Wong (These solution hints are based on Julia Salmi s solution hints for Spring 2015.
FDPE Microeconomics 3 Spring 2017 Pauli Murto TA: Tsz-Ning Wong (These solution hints are based on Julia Salmi s solution hints for Spring 2015.) Hints for Problem Set 2 1. Consider a zero-sum game, where
More informationGame Theory Tutorial 3 Answers
Game Theory Tutorial 3 Answers Exercise 1 (Duality Theory) Find the dual problem of the following L.P. problem: max x 0 = 3x 1 + 2x 2 s.t. 5x 1 + 2x 2 10 4x 1 + 6x 2 24 x 1 + x 2 1 (1) x 1 + 3x 2 = 9 x
More information**BEGINNING OF EXAMINATION**
Fall 2002 Society of Actuaries **BEGINNING OF EXAMINATION** 1. Given: The survival function s x sbxg = 1, 0 x < 1 b g x d i { } b g, where s x = 1 e / 100, 1 x < 45. b g = s x 0, 4.5 x Calculate µ b4g.
More informationBSc (Hons) Software Engineering BSc (Hons) Computer Science with Network Security
BSc (Hons) Software Engineering BSc (Hons) Computer Science with Network Security Cohorts BCNS/ 06 / Full Time & BSE/ 06 / Full Time Resit Examinations for 2008-2009 / Semester 1 Examinations for 2008-2009
More informationThe exam is closed book, closed calculator, and closed notes except your three crib sheets.
CS 188 Spring 2016 Introduction to Artificial Intelligence Final V2 You have approximately 2 hours and 50 minutes. The exam is closed book, closed calculator, and closed notes except your three crib sheets.
More information12 Math Chapter Review April 16 th, Multiple Choice Identify the choice that best completes the statement or answers the question.
Multiple Choice Identify the choice that best completes the statement or answers the question. 1. Which situation does not describe a discrete random variable? A The number of cell phones per household.
More informationChapter 5. Continuous Random Variables and Probability Distributions. 5.1 Continuous Random Variables
Chapter 5 Continuous Random Variables and Probability Distributions 5.1 Continuous Random Variables 1 2CHAPTER 5. CONTINUOUS RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Probability Distributions Probability
More informationECON 214 Elements of Statistics for Economists 2016/2017
ECON 214 Elements of Statistics for Economists 2016/2017 Topic The Normal Distribution Lecturer: Dr. Bernardin Senadza, Dept. of Economics bsenadza@ug.edu.gh College of Education School of Continuing and
More informationMATH1215: Mathematical Thinking Sec. 08 Spring Worksheet 9: Solution. x P(x)
N. Name: MATH: Mathematical Thinking Sec. 08 Spring 0 Worksheet 9: Solution Problem Compute the expected value of this probability distribution: x 3 8 0 3 P(x) 0. 0.0 0.3 0. Clearly, a value is missing
More informationFURTHER ASPECTS OF GAMBLING WITH THE KELLY CRITERION. We consider two aspects of gambling with the Kelly criterion. First, we show that for
FURTHER ASPECTS OF GAMBLING WITH THE KELLY CRITERION RAVI PHATARFOD *, Monash University Abstract We consider two aspects of gambling with the Kelly criterion. First, we show that for a wide range of final
More informationECON 214 Elements of Statistics for Economists
ECON 214 Elements of Statistics for Economists Session 7 The Normal Distribution Part 1 Lecturer: Dr. Bernardin Senadza, Dept. of Economics Contact Information: bsenadza@ug.edu.gh College of Education
More informationMonte-Carlo Planning: Introduction and Bandit Basics. Alan Fern
Monte-Carlo Planning: Introduction and Bandit Basics Alan Fern 1 Large Worlds We have considered basic model-based planning algorithms Model-based planning: assumes MDP model is available Methods we learned
More informationb) Consider the sample space S = {1, 2, 3}. Suppose that P({1, 2}) = 0.5 and P({2, 3}) = 0.7. Is P a valid probability measure? Justify your answer.
JARAMOGI OGINGA ODINGA UNIVERSITY OF SCIENCE AND TECHNOLOGY BACHELOR OF SCIENCE -ACTUARIAL SCIENCE YEAR ONE SEMESTER ONE SAS 103: INTRODUCTION TO PROBABILITY THEORY Instructions: Answer question 1 and
More informationMULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.
AP Stats: Test Review - Chapters 16-17 Name Period MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question. Find the expected value of the random variable.
More informationLecture 17: More on Markov Decision Processes. Reinforcement learning
Lecture 17: More on Markov Decision Processes. Reinforcement learning Learning a model: maximum likelihood Learning a value function directly Monte Carlo Temporal-difference (TD) learning COMP-424, Lecture
More information3 Stock under the risk-neutral measure
3 Stock under the risk-neutral measure 3 Adapted processes We have seen that the sampling space Ω = {H, T } N underlies the N-period binomial model for the stock-price process Elementary event ω = ω ω
More informationExpected Value of a Random Variable
Knowledge Article: Probability and Statistics Expected Value of a Random Variable Expected Value of a Discrete Random Variable You're familiar with a simple mean, or average, of a set. The mean value of
More informationChapter 8. Variables. Copyright 2004 Brooks/Cole, a division of Thomson Learning, Inc.
Chapter 8 Random Variables Copyright 2004 Brooks/Cole, a division of Thomson Learning, Inc. 8.1 What is a Random Variable? Random Variable: assigns a number to each outcome of a random circumstance, or,
More informationDynamic Pricing with Varying Cost
Dynamic Pricing with Varying Cost L. Jeff Hong College of Business City University of Hong Kong Joint work with Ying Zhong and Guangwu Liu Outline 1 Introduction 2 Problem Formulation 3 Pricing Policy
More informationCounting Basics. Venn diagrams
Counting Basics Sets Ways of specifying sets Union and intersection Universal set and complements Empty set and disjoint sets Venn diagrams Counting Inclusion-exclusion Multiplication principle Addition
More informationCEC login. Student Details Name SOLUTIONS
Student Details Name SOLUTIONS CEC login Instructions You have roughly 1 minute per point, so schedule your time accordingly. There is only one correct answer per question. Good luck! Question 1. Searching
More informationProbability Distribution Unit Review
Probability Distribution Unit Review Topics: Pascal's Triangle and Binomial Theorem Probability Distributions and Histograms Expected Values, Fair Games of chance Binomial Distributions Hypergeometric
More informationECE 302 Spring Ilya Pollak
ECE 302 Spring 202 Practice problems: Multiple discrete random variables, joint PMFs, conditional PMFs, conditional expectations, functions of random variables Ilya Pollak These problems have been constructed
More informationDepartment of Agricultural Economics PhD Qualifier Examination January 2005
Department of Agricultural Economics PhD Qualifier Examination January 2005 Instructions: The exam consists of six questions. You must answer all questions. If you need an assumption to complete a question,
More informationSECTION 4.4: Expected Value
15 SECTION 4.4: Expected Value This section tells you why most all gambling is a bad idea. And also why carnival or amusement park games are a bad idea. Random Variables Definition: Random Variable A random
More informationRollout Allocation Strategies for Classification-based Policy Iteration
Rollout Allocation Strategies for Classification-based Policy Iteration V. Gabillon, A. Lazaric & M. Ghavamzadeh firstname.lastname@inria.fr Workshop on Reinforcement Learning and Search in Very Large
More informationHeckmeck am Bratwurmeck or How to grill the maximum number of worms
Heckmeck am Bratwurmeck or How to grill the maximum number of worms Roland C. Seydel 24/05/22 (1) Heckmeck am Bratwurmeck 24/05/22 1 / 29 Overview 1 Introducing the dice game The basic rules Understanding
More informationDiscrete Probability Distributions
Chapter 5 Discrete Probability Distributions Goal: To become familiar with how to use Excel 2007/2010 for binomial distributions. Instructions: Open Excel and click on the Stat button in the Quick Access
More informationSolutions for practice questions: Chapter 15, Probability Distributions If you find any errors, please let me know at
Solutions for practice questions: Chapter 15, Probability Distributions If you find any errors, please let me know at mailto:msfrisbie@pfrisbie.com. 1. Let X represent the savings of a resident; X ~ N(3000,
More informationMonte-Carlo Planning: Introduction and Bandit Basics. Alan Fern
Monte-Carlo Planning: Introduction and Bandit Basics Alan Fern 1 Large Worlds We have considered basic model-based planning algorithms Model-based planning: assumes MDP model is available Methods we learned
More informationIntroduction to Game Theory
Introduction to Game Theory What is a Game? A game is a formal representation of a situation in which a number of individuals interact in a setting of strategic interdependence. By that, we mean that each
More informationDefinition 4.1. In a stochastic process T is called a stopping time if you can tell when it happens.
102 OPTIMAL STOPPING TIME 4. Optimal Stopping Time 4.1. Definitions. On the first day I explained the basic problem using one example in the book. On the second day I explained how the solution to the
More informationChapter 3 Discrete Random Variables and Probability Distributions
Chapter 3 Discrete Random Variables and Probability Distributions Part 2: Mean and Variance of a Discrete Random Variable Section 3.4 1 / 16 Discrete Random Variable - Expected Value In a random experiment,
More informationMULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question.
First Name: Last Name: SID: Class Time: M Tu W Th math10 - HW3 MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question. 1) Continuous random variables are
More informationModelling Anti-Terrorist Surveillance Systems from a Queueing Perspective
Systems from a Queueing Perspective September 7, 2012 Problem A surveillance resource must observe several areas, searching for potential adversaries. Problem A surveillance resource must observe several
More informationMonetary Economics Final Exam
316-466 Monetary Economics Final Exam 1. Flexible-price monetary economics (90 marks). Consider a stochastic flexibleprice money in the utility function model. Time is discrete and denoted t =0, 1,...
More informationReinforcement learning and Markov Decision Processes (MDPs) (B) Avrim Blum
Reinforcement learning and Markov Decision Processes (MDPs) 15-859(B) Avrim Blum RL and MDPs General scenario: We are an agent in some state. Have observations, perform actions, get rewards. (See lights,
More information2011 Pearson Education, Inc
Statistics for Business and Economics Chapter 4 Random Variables & Probability Distributions Content 1. Two Types of Random Variables 2. Probability Distributions for Discrete Random Variables 3. The Binomial
More informationOptimal Stopping. Nick Hay (presentation follows Thomas Ferguson s Optimal Stopping and Applications) November 6, 2008
(presentation follows Thomas Ferguson s and Applications) November 6, 2008 1 / 35 Contents: Introduction Problems Markov Models Monotone Stopping Problems Summary 2 / 35 The Secretary problem You have
More informationLecture 11: Bandits with Knapsacks
CMSC 858G: Bandits, Experts and Games 11/14/16 Lecture 11: Bandits with Knapsacks Instructor: Alex Slivkins Scribed by: Mahsa Derakhshan 1 Motivating Example: Dynamic Pricing The basic version of the dynamic
More informationMath-Stat-491-Fall2014-Notes-V
Math-Stat-491-Fall2014-Notes-V Hariharan Narayanan December 7, 2014 Martingales 1 Introduction Martingales were originally introduced into probability theory as a model for fair betting games. Essentially
More informationPh.D. Preliminary Examination MICROECONOMIC THEORY Applied Economics Graduate Program June 2015
Ph.D. Preliminary Examination MICROECONOMIC THEORY Applied Economics Graduate Program June 2015 The time limit for this exam is four hours. The exam has four sections. Each section includes two questions.
More informationThe exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
CS 188 Summer 2015 Introduction to Artificial Intelligence Midterm 2 You have approximately 80 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. Mark
More informationGame Theory with Applications to Finance and Marketing, I
Game Theory with Applications to Finance and Marketing, I Homework 1, due in recitation on 10/18/2018. 1. Consider the following strategic game: player 1/player 2 L R U 1,1 0,0 D 0,0 3,2 Any NE can be
More informationIntroduction to Fall 2007 Artificial Intelligence Final Exam
NAME: SID#: Login: Sec: 1 CS 188 Introduction to Fall 2007 Artificial Intelligence Final Exam You have 180 minutes. The exam is closed book, closed notes except a two-page crib sheet, basic calculators
More informationTUFTS UNIVERSITY DEPARTMENT OF CIVIL AND ENVIRONMENTAL ENGINEERING ES 152 ENGINEERING SYSTEMS Spring Lesson 16 Introduction to Game Theory
TUFTS UNIVERSITY DEPARTMENT OF CIVIL AND ENVIRONMENTAL ENGINEERING ES 52 ENGINEERING SYSTEMS Spring 20 Introduction: Lesson 6 Introduction to Game Theory We will look at the basic ideas of game theory.
More informationMidTerm 1) Find the following (round off to one decimal place):
MidTerm 1) 68 49 21 55 57 61 70 42 59 50 66 99 Find the following (round off to one decimal place): Mean = 58:083, round off to 58.1 Median = 58 Range = max min = 99 21 = 78 St. Deviation = s = 8:535,
More informationFinal Examination CS540: Introduction to Artificial Intelligence
Final Examination CS540: Introduction to Artificial Intelligence December 2008 LAST NAME: FIRST NAME: Problem Score Max Score 1 15 2 15 3 10 4 20 5 10 6 20 7 10 Total 100 Question 1. [15] Probabilistic
More informationBayes s Rule Example. defective. An MP3 player is selected at random and found to be defective. What is the probability it came from Factory I?
Bayes s Rule Example A company manufactures MP3 players at two factories. Factory I produces 60% of the MP3 players and Factory II produces 40%. Two percent of the MP3 players produced at Factory I are
More informationChapter 2: Random Variables (Cont d)
Chapter : Random Variables (Cont d) Section.4: The Variance of a Random Variable Problem (1): Suppose that the random variable X takes the values, 1, 4, and 6 with probability values 1/, 1/6, 1/, and 1/6,
More information16 MAKING SIMPLE DECISIONS
253 16 MAKING SIMPLE DECISIONS Let us associate each state S with a numeric utility U(S), which expresses the desirability of the state A nondeterministic action a will have possible outcome states Result(a)
More informationCS360 Homework 14 Solution
CS360 Homework 14 Solution Markov Decision Processes 1) Invent a simple Markov decision process (MDP) with the following properties: a) it has a goal state, b) its immediate action costs are all positive,
More informationChapter 3 Discrete Random Variables and Probability Distributions
Chapter 3 Discrete Random Variables and Probability Distributions Part 4: Special Discrete Random Variable Distributions Sections 3.7 & 3.8 Geometric, Negative Binomial, Hypergeometric NOTE: The discrete
More informationOverview: Representation Techniques
1 Overview: Representation Techniques Week 6 Representations for classical planning problems deterministic environment; complete information Week 7 Logic programs for problem representations including
More informationChapter 2 Linear programming... 2 Chapter 3 Simplex... 4 Chapter 4 Sensitivity Analysis and duality... 5 Chapter 5 Network... 8 Chapter 6 Integer
目录 Chapter 2 Linear programming... 2 Chapter 3 Simplex... 4 Chapter 4 Sensitivity Analysis and duality... 5 Chapter 5 Network... 8 Chapter 6 Integer Programming... 10 Chapter 7 Nonlinear Programming...
More informationMulti-armed bandit problems
Multi-armed bandit problems Stochastic Decision Theory (2WB12) Arnoud den Boer 13 March 2013 Set-up 13 and 14 March: Lectures. 20 and 21 March: Paper presentations (Four groups, 45 min per group). Before
More informationBinomal and Geometric Distributions
Binomal and Geometric Distributions Sections 3.2 & 3.3 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Department of Mathematics University of Houston Lecture 7-2311 Cathy Poliak, Ph.D. cathy@math.uh.edu
More information16 MAKING SIMPLE DECISIONS
247 16 MAKING SIMPLE DECISIONS Let us associate each state S with a numeric utility U(S), which expresses the desirability of the state A nondeterministic action A will have possible outcome states Result
More informationProblem Set: Contract Theory
Problem Set: Contract Theory Problem 1 A risk-neutral principal P hires an agent A, who chooses an effort a 0, which results in gross profit x = a + ε for P, where ε is uniformly distributed on [0, 1].
More informationThe exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
CS 188 Spring 2015 Introduction to Artificial Intelligence Midterm 1 You have approximately 2 hours and 50 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib
More informationHW Consider the following game:
HW 1 1. Consider the following game: 2. HW 2 Suppose a parent and child play the following game, first analyzed by Becker (1974). First child takes the action, A 0, that produces income for the child,
More informationSI 563 Homework 3 Oct 5, Determine the set of rationalizable strategies for each of the following games. a) X Y X Y Z
SI 563 Homework 3 Oct 5, 06 Chapter 7 Exercise : ( points) Determine the set of rationalizable strategies for each of the following games. a) U (0,4) (4,0) M (3,3) (3,3) D (4,0) (0,4) X Y U (0,4) (4,0)
More informationRecord on a ScanTron, your choosen response for each question. You may write on this form. One page of notes and a calculator are allowed.
Ch 16, 17 Math 240 Exam 4 v1 Good SAMPLE No Book, Yes 1 Page Notes, Yes Calculator, 120 Minutes Dressler Record on a ScanTron, your choosen response for each question. You may write on this form. One page
More informationChapter 4 Discrete Random variables
Chapter 4 Discrete Random variables A is a variable that assumes numerical values associated with the random outcomes of an experiment, where only one numerical value is assigned to each sample point.
More informationChapter 3 Discrete Random Variables and Probability Distributions
Chapter 3 Discrete Random Variables and Probability Distributions Part 3: Special Discrete Random Variable Distributions Section 3.5 Discrete Uniform Section 3.6 Bernoulli and Binomial Others sections
More informationCS188 Spring 2012 Section 4: Games
CS188 Spring 2012 Section 4: Games 1 Minimax Search In this problem, we will explore adversarial search. Consider the zero-sum game tree shown below. Trapezoids that point up, such as at the root, represent
More information2D5362 Machine Learning
2D5362 Machine Learning Reinforcement Learning MIT GALib Available at http://lancet.mit.edu/ga/ download galib245.tar.gz gunzip galib245.tar.gz tar xvf galib245.tar cd galib245 make or access my files
More informationPROBABILITY AND STATISTICS, A16, TEST 1
PROBABILITY AND STATISTICS, A16, TEST 1 Name: Student number (1) (1.5 marks) i) Let A and B be mutually exclusive events with p(a) = 0.7 and p(b) = 0.2. Determine p(a B ) and also p(a B). ii) Let C and
More informationECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2017
ECON 459 Game Theory Lecture Notes Auctions Luca Anderlini Spring 2017 These notes have been used and commented on before. If you can still spot any errors or have any suggestions for improvement, please
More informationConsider the following (true) preference orderings of 4 agents on 4 candidates.
Part 1: Voting Systems Consider the following (true) preference orderings of 4 agents on 4 candidates. Agent #1: A > B > C > D Agent #2: B > C > D > A Agent #3: C > B > D > A Agent #4: D > C > A > B Assume
More informationThe Binomial Distribution
The Binomial Distribution Properties of a Binomial Experiment 1. It consists of a fixed number of observations called trials. 2. Each trial can result in one of only two mutually exclusive outcomes labeled
More informationLecture 23. STAT 225 Introduction to Probability Models April 4, Whitney Huang Purdue University. Normal approximation to Binomial
Lecture 23 STAT 225 Introduction to Probability Models April 4, 2014 approximation Whitney Huang Purdue University 23.1 Agenda 1 approximation 2 approximation 23.2 Characteristics of the random variable:
More informationAP Statistics Section 6.1 Day 1 Multiple Choice Practice. a) a random variable. b) a parameter. c) biased. d) a random sample. e) a statistic.
A Statistics Section 6.1 Day 1 ultiple Choice ractice Name: 1. A variable whose value is a numerical outcome of a random phenomenon is called a) a random variable. b) a parameter. c) biased. d) a random
More informationMultinomial Coefficient : A Generalization of the Binomial Coefficient
Multinomial Coefficient : A Generalization of the Binomial Coefficient Example: A team plays 16 games in a season. At the end of the season, the team has 8 wins, 3 ties and 5 losses. How many different
More informationLecture 22. Survey Sampling: an Overview
Math 408 - Mathematical Statistics Lecture 22. Survey Sampling: an Overview March 25, 2013 Konstantin Zuev (USC) Math 408, Lecture 22 March 25, 2013 1 / 16 Survey Sampling: What and Why In surveys sampling
More informationCUR 412: Game Theory and its Applications, Lecture 9
CUR 412: Game Theory and its Applications, Lecture 9 Prof. Ronaldo CARPIO May 22, 2015 Announcements HW #3 is due next week. Ch. 6.1: Ultimatum Game This is a simple game that can model a very simplified
More informationStatistics, Their Distributions, and the Central Limit Theorem
Statistics, Their Distributions, and the Central Limit Theorem MATH 3342 Sections 5.3 and 5.4 Sample Means Suppose you sample from a popula0on 10 0mes. You record the following sample means: 10.1 9.5 9.6
More informationCPS 270: Artificial Intelligence Markov decision processes, POMDPs
CPS 270: Artificial Intelligence http://www.cs.duke.edu/courses/fall08/cps270/ Markov decision processes, POMDPs Instructor: Vincent Conitzer Warmup: a Markov process with rewards We derive some reward
More informationChapter 10 Inventory Theory
Chapter 10 Inventory Theory 10.1. (a) Find the smallest n such that g(n) 0. g(1) = 3 g(2) =2 n = 2 (b) Find the smallest n such that g(n) 0. g(1) = 1 25 1 64 g(2) = 1 4 1 25 g(3) =1 1 4 g(4) = 1 16 1
More informationMATH3075/3975 FINANCIAL MATHEMATICS TUTORIAL PROBLEMS
MATH307/37 FINANCIAL MATHEMATICS TUTORIAL PROBLEMS School of Mathematics and Statistics Semester, 04 Tutorial problems should be used to test your mathematical skills and understanding of the lecture material.
More informationSubject : Computer Science. Paper: Machine Learning. Module: Decision Theory and Bayesian Decision Theory. Module No: CS/ML/10.
e-pg Pathshala Subject : Computer Science Paper: Machine Learning Module: Decision Theory and Bayesian Decision Theory Module No: CS/ML/0 Quadrant I e-text Welcome to the e-pg Pathshala Lecture Series
More informationHomework: Due Wed, Nov 3 rd Chapter 8, # 48a, 55c and 56 (count as 1), 67a
Homework: Due Wed, Nov 3 rd Chapter 8, # 48a, 55c and 56 (count as 1), 67a Announcements: There are some office hour changes for Nov 5, 8, 9 on website Week 5 quiz begins after class today and ends at
More informationHonors Statistics. Aug 23-8:26 PM. 1. Collect folders and materials. 2. Continue Binomial Probability. 3. Review OTL C6#11 homework
Honors Statistics Aug 23-8:26 PM 1. Collect folders and materials 2. Continue Binomial Probability 3. Review OTL C6#11 homework 4. Binomial mean and standard deviation 5. Past Homework discussion 6. Return
More informationMixed Strategies. Samuel Alizon and Daniel Cownden February 4, 2009
Mixed Strategies Samuel Alizon and Daniel Cownden February 4, 009 1 What are Mixed Strategies In the previous sections we have looked at games where players face uncertainty, and concluded that they choose
More informationThis assignment is due on Tuesday, September 15, at the beginning of class (or sooner).
Econ 434 Professor Ickes Homework Assignment #1: Answer Sheet Fall 2009 This assignment is due on Tuesday, September 15, at the beginning of class (or sooner). 1. Consider the following returns data for
More informationDiscrete Random Variables and Their Probability Distributions
58 Chapter 5 Discrete Random Variables and Their Probability Distributions Discrete Random Variables and Their Probability Distributions Chapter 5 Section 5.6 Example 5-18, pg. 213 Calculating a Binomial
More information