BMI/CS 776 Lecture #15: Multiple Alignment - ProbCons. Colin Dewey
|
|
- Brett Banks
- 5 years ago
- Views:
Transcription
1 BMI/CS 776 Lecture #15: Multiple Alignment - ProbCons Colin Dewey
2 Probabilistic multiple alignment Like Needleman-Wunsch, pair HMMs can be generalized to n > 2 sequences Unfortunately, the complexity of such algorithms is exponential in n However, we can still use pair HMMs to our advantage in multiple alignment ProbCons & AMAP: use pairwise statistics as part of objective function 2
3 ProbCons Do et al., 2005 Progressive multiple alignment with refinement Key ideas: Objective function based on PHMM probabilities - maximum expected accuracy Probabilistic consistency - transitivity of homology 3
4 ProbCons algorithm 1. Compute PHMM posterior probabilities for all pairs of sequences 2. Compute maximum expected accuracy (MEA) between all pairs of sequences 3. Apply probabilistic consistency transformation to posterior probabilities 4. Compute guide tree from MEA values 5. Progressively align sequences using guide tree 6. Iteratively refine multiple alignment 4
5 PHMM posteriors Forward and backward algorithms for PHMMs - use to calculate posterior probabilities of alignment properties Specifically, interested in calculating the posterior probability of being aligned to σj 2 P[σ 1 i σ 2 j σ 1, σ 2 ] = P[σ1, σ 2, σ 1 i σ2 j ] P[σ 1, σ 2 ] σ 1 i = P[σ1 1...i, σ2 1...j, σ1 i σ2 j ]P[σ1 i+1...n, σ2 j+1...m σ1 i σ2 j ] P[σ 1, σ 2 ] = f H (i, j)b H (i, j) f E (n, m) 5
6 Expected accuracy Assume P[h r σ 1, σ 2 ] (according to the PHMM) is the probability of alignment being the true alignment E h r[accuracy(h, h r ) σ 1, σ 2 ] = accuracy(h, h r ) = h H h r H min(n, m) = = = h P[h r σ 1, σ 2 ](h r H h r H ) min(n, m) h r P[h r σ 1, σ 2 ] σ 1 i σ2 j h I h r (i, j) min(n, m) σ 1i σ2j h h P[h r σ 1, σ 2 ]I r h r(i, j) min(n, m) σi 1 σ2 j h P[σ1 i σ2 j σ1, σ 2 ] min(n, m) h r 6
7 Maximum expected accuracy (MEA) Goal: Find h such that expected accuracy is maximized E h r[accuracy(h, h r ) σ 1, σ 2 ] = σ 1 i σ2 j h P[σ1 i σ2 j σ1, σ 2 ] min(n, m) Algorithm: Run Needleman-Wunsch σ 1 i σ 2 j alignment with match score of equal to P[σi 1 σj 2 σ 1, σ 2 ] and with zero gap penalties 7
8 Probabilistic consistency transformation Take into account third sequence in calculating P[σ 1 i σ 2 j σ 1, σ 2, σ 3 ] = P[σ 1 i σ 2 j σ 1, σ 2, σ 3 ] + k P[σ 1 i σ 2 j σ 3 k σ 1, σ 2, σ 3 ] k P[σ 1 i σ 2 j σ 3 k σ 1, σ 2, σ 3 ] = k P[σ 1 i σ 3 k and σ 3 k σ 2 j σ 1, σ 2, σ 3 ] k P[σ 1 i σ 3 k σ 1, σ 3 ]P[σ 3 k σ 2 j σ 2, σ 3 ] where are all heuristic approximations 8
9 Probabilistic consistency transformation Over all possible third sequences P [σ 1 i σ 2 j σ 1, σ 2 ] = σ 3 S k P[σ 1 i σ 3 k σ 1, σ 3 ]P[σ 3 k σ 2 j σ 2, σ 3 ] Implemented as sparse matrix multiplication 9
10 Guide tree construction UPGMA using maximum expected accuracy distances Sequences start out as individual clusters Join clusters x and y with highest E(x, y) into new cluster xy Compute E(xy, z) = (E(x,z) + E(y,z)) / 2 Repeat until all clusters joined 10
11 Progressive alignment Profile alignment Sum-of-pairs with maximal expected accuracy scoring No gap penalities Score for aligning two columns: score(c 1, c 2 ) = P [σ s i σ t j S] σ s i c 1 σ t j c 2 11
12 Iterative refinement Randomly partition sequences into two sets Extract multiple alignments for both sets from current multiple alignment Re-align two multiple alignments to make one No gap penalty - sum-of-pairs scoring guaranteed to increase or stay the same 12
13 Other features Unsupervised learning of PHMM parameters Emission probabilities fixed (BLOSUM62) Learn transitions probs via EM Column reliability ψ(c) = ( ) 1 c 2 σ s i,σt j c P [σ s i σ t j S] 13
14 Performance Table 1. Aligner Performance of aligners on the BAliBASE benchmark alignments database Ref 1 (82) Ref 2 (23) Ref 3 (12) Ref 4 (12) Ref 5 (12) Overall (141) SP CS SP CS SP CS SP CS SP CS SP CS Time (mm:ss) Align-m 76.6 n/a 88.4 n/a 68.4 n/a 91.1 n/a 91.7 n/a 80.4 n/a 19:25 DIALIGN :53 CLUSTALW :07 MAFFT :18 T-Coffee :31 MUSCLE :05 ProbCons :32 ProbCons-ext :02 Columns show the average sum-of-pairs (SP) and column scores (CS) achieved by each aligner for each of the five BAliBASE references. All scores have been multiplied by 100. The number of sequences in each reference is given in parentheses. Overall numbers for the entire database are reported in addition to the total running time of each aligner for all 141 alignments. The best results in each column are shown in bold. Do et al.,
15 Friedman rank test k treatments b blocks Each treatment is ranked within each block Tests whether treatments have identical effects Multiple alignment case: treatment = alignment method block = alignment reference database rank methods based on some score (e.g., sensitivity) 15
16 Comparing ProbCons Table 2. Significance test for differences in BAliBASE performance Align-M DIALIGN CLUSTALW MAFFT T-Coffee MUSCLE ProbCons ProbCons-ext Align-M (0.61) <10 10 <10 10 <10 10 <10 10 <10 10 DIALIGN <10 10 <10 10 <10 10 <10 10 <10 10 CLUSTALW <10 10 MAFFT (0.65) T-Coffee +< (0.92) MUSCLE +< ProbCons +< < ProbCons-ext +< < (0.092) (0.088) Entries show the p-value indicating the significance of a difference in performance between two alignment methods as measured using a Friedman rank test. Nonitalicized values above the diagonal were calculated using SP scores on all alignments, whereas italicized values were computed using CS scores. (+) Method on the left had lower average rank (better performance); ( ) Method on the left had higher average rank (worse performance); parentheses denote (nonsignificant) p-values >0.05. Methods compared pairwise via Friedman rank test Do et al.,
Exact Inference (9/30/13) 2 A brief review of Forward-Backward and EM for HMMs
STA561: Probabilistic machine learning Exact Inference (9/30/13) Lecturer: Barbara Engelhardt Scribes: Jiawei Liang, He Jiang, Brittany Cohen 1 Validation for Clustering If we have two centroids, η 1 and
More informationAdvanced Sequence Alignment. Problem Set #4 is posted.
Advanced Sequence Alignment Problem Set #4 is posted. 1 Recall Local Alignment The zero is our free ride that allows the node to restart with a score of 0 at any point What does this imply? After solving
More informationThe EM algorithm for HMMs
The EM algorithm for HMMs Michael Collins February 22, 2012 Maximum-Likelihood Estimation for Fully Observed Data (Recap from earlier) We have fully observed data, x i,1... x i,m, s i,1... s i,m for i
More informationMolecular Phylogenetics
Mole_Oce Lecture # 16: Molecular Phylogenetics Maximum Likelihood & Bahesian Statistics Optimality criterion: a rule used to decide which of two trees is best. Four optimality criteria are currently widely
More informationA Learning Theory of Ranking Aggregation
A Learning Theory of Ranking Aggregation France/Japan Machine Learning Workshop Anna Korba, Stephan Clémençon, Eric Sibony November 14, 2017 Télécom ParisTech Outline 1. The Ranking Aggregation Problem
More informationNotes on the EM Algorithm Michael Collins, September 24th 2005
Notes on the EM Algorithm Michael Collins, September 24th 2005 1 Hidden Markov Models A hidden Markov model (N, Σ, Θ) consists of the following elements: N is a positive integer specifying the number of
More informationCS 188: Artificial Intelligence
CS 188: Artificial Intelligence Markov Decision Processes Dan Klein, Pieter Abbeel University of California, Berkeley Non Deterministic Search Example: Grid World A maze like problem The agent lives in
More informationCS188 Spring 2012 Section 4: Games
CS188 Spring 2012 Section 4: Games 1 Minimax Search In this problem, we will explore adversarial search. Consider the zero-sum game tree shown below. Trapezoids that point up, such as at the root, represent
More informationRandom Variables and Applications OPRE 6301
Random Variables and Applications OPRE 6301 Random Variables... As noted earlier, variability is omnipresent in the business world. To model variability probabilistically, we need the concept of a random
More informationComputer Vision Group Prof. Daniel Cremers. 7. Sequential Data
Group Prof. Daniel Cremers 7. Sequential Data Bayes Filter (Rep.) We can describe the overall process using a Dynamic Bayes Network: This incorporates the following Markov assumptions: (measurement) (state)!2
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning Monte Carlo Methods Mark Schmidt University of British Columbia Winter 2019 Last Time: Markov Chains We can use Markov chains for density estimation, d p(x) = p(x 1 ) p(x }{{}
More informationMarkov Decision Processes
Markov Decision Processes Robert Platt Northeastern University Some images and slides are used from: 1. CS188 UC Berkeley 2. AIMA 3. Chris Amato Stochastic domains So far, we have studied search Can use
More informationAcademic Research Review. Classifying Market Conditions Using Hidden Markov Model
Academic Research Review Classifying Market Conditions Using Hidden Markov Model INTRODUCTION Best known for their applications in speech recognition, Hidden Markov Models (HMMs) are able to discern and
More informationBasic Procedure for Histograms
Basic Procedure for Histograms 1. Compute the range of observations (min. & max. value) 2. Choose an initial # of classes (most likely based on the range of values, try and find a number of classes that
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning Monte Carlo Methods Mark Schmidt University of British Columbia Winter 2018 Last Time: Markov Chains We can use Markov chains for density estimation, p(x) = p(x 1 ) }{{} d p(x
More informationSupplementary Material for Combinatorial Partial Monitoring Game with Linear Feedback and Its Application. A. Full proof for Theorems 4.1 and 4.
Supplementary Material for Combinatorial Partial Monitoring Game with Linear Feedback and Its Application. A. Full proof for Theorems 4.1 and 4. If the reader will recall, we have the following problem-specific
More informationAlgorithmic Trading using Reinforcement Learning augmented with Hidden Markov Model
Algorithmic Trading using Reinforcement Learning augmented with Hidden Markov Model Simerjot Kaur (sk3391) Stanford University Abstract This work presents a novel algorithmic trading system based on reinforcement
More informationGotoh Scan Algorithm for matching RNA sequences. By Hila Abukasis & Shai Kerer
otoh Scan lgorithm for matching RN sequences By Hila bukasis & Shai Kerer ontents What is RN? Matching RN Needleman-Wunsch lgorithm lobal lignment VS Local lignment Smith-Waterman lgorithm otoh Scan lgorithm
More informationReasoning with Uncertainty
Reasoning with Uncertainty Markov Decision Models Manfred Huber 2015 1 Markov Decision Process Models Markov models represent the behavior of a random process, including its internal state and the externally
More informationStatistical and Computational Inverse Problems with Applications Part 5B: Electrical impedance tomography
Statistical and Computational Inverse Problems with Applications Part 5B: Electrical impedance tomography Aku Seppänen Inverse Problems Group Department of Applied Physics University of Eastern Finland
More informationMarkov Decision Processes
Markov Decision Processes Robert Platt Northeastern University Some images and slides are used from: 1. CS188 UC Berkeley 2. RN, AIMA Stochastic domains Image: Berkeley CS188 course notes (downloaded Summer
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2011 Lecture 9: MDPs 2/16/2011 Pieter Abbeel UC Berkeley Many slides over the course adapted from either Dan Klein, Stuart Russell or Andrew Moore 1 Announcements
More informationThreshold Accepting for Credit Risk Assessment and Validation
Threshold Accepting for Credit Risk Assessment and Validation M. Lyra 1 A. Onwunta P. Winker COMPSTAT 2010 August 24, 2010 1 Financial support from the EU Commission through COMISEF is gratefully acknowledged
More informationMultiple Regression and Logistic Regression II. Dajiang 525 Apr
Multiple Regression and Logistic Regression II Dajiang Liu @PHS 525 Apr-19-2016 Materials from Last Time Multiple regression model: Include multiple predictors in the model = + + + + How to interpret the
More informationSublinear Time Algorithms Oct 19, Lecture 1
0368.416701 Sublinear Time Algorithms Oct 19, 2009 Lecturer: Ronitt Rubinfeld Lecture 1 Scribe: Daniel Shahaf 1 Sublinear-time algorithms: motivation Twenty years ago, there was practically no investigation
More informationChapter 10: Mixed strategies Nash equilibria, reaction curves and the equality of payoffs theorem
Chapter 10: Mixed strategies Nash equilibria reaction curves and the equality of payoffs theorem Nash equilibrium: The concept of Nash equilibrium can be extended in a natural manner to the mixed strategies
More informationTopics in Computational Sustainability CS 325 Spring 2016
Topics in Computational Sustainability CS 325 Spring 2016 Note to other teachers and users of these slides. Andrew would be delighted if you found this source material useful in giving your own lectures.
More informationComparing alternatives using multiple criteria
Comparing alternatives using multiple criteria Denns L. Bricker Dept of Mechanical & Industrial Engineering The University of Iowa AHP 2/4/2003 page 1 of 22 When a decision-maker has multiple objectives,
More informationA Hidden Markov Model Approach to Information-Based Trading: Theory and Applications
A Hidden Markov Model Approach to Information-Based Trading: Theory and Applications Online Supplementary Appendix Xiangkang Yin and Jing Zhao La Trobe University Corresponding author, Department of Finance,
More informationLecture 10: Alignments with Affine Gaps. The Local Alignment Recurrence
Lecture 10: Alignments with Affine Gaps Study Chapter 6.9-6.10 1 The Local Alignment Recurrence The largest value of s i,j over the whole edit graph is the score of the best local alignment. Smith-Waterman
More informationModeling Private Firm Default: PFirm
Modeling Private Firm Default: PFirm Grigoris Karakoulas Business Analytic Solutions May 30 th, 2002 Outline Problem Statement Modelling Approaches Private Firm Data Mining Model Development Model Evaluation
More information4 Martingales in Discrete-Time
4 Martingales in Discrete-Time Suppose that (Ω, F, P is a probability space. Definition 4.1. A sequence F = {F n, n = 0, 1,...} is called a filtration if each F n is a sub-σ-algebra of F, and F n F n+1
More informationNon-Deterministic Search
Non-Deterministic Search MDP s 1 Non-Deterministic Search How do you plan (search) when your actions might fail? In general case, how do you plan, when the actions have multiple possible outcomes? 2 Example:
More informationIntroduction to Fall 2007 Artificial Intelligence Final Exam
NAME: SID#: Login: Sec: 1 CS 188 Introduction to Fall 2007 Artificial Intelligence Final Exam You have 180 minutes. The exam is closed book, closed notes except a two-page crib sheet, basic calculators
More informationEpistemic Game Theory
Epistemic Game Theory Lecture 1 ESSLLI 12, Opole Eric Pacuit Olivier Roy TiLPS, Tilburg University MCMP, LMU Munich ai.stanford.edu/~epacuit http://olivier.amonbofis.net August 6, 2012 Eric Pacuit and
More informationEmil Tesliuc and Phillippe Leite November 23, 2009
Emil Tesliuc and Phillippe Leite November 23, 2009 ADePT SP (developed by HDNSP-SSN SSN team and Development Research Group -Poverty Team ) ADePT SP is a Stata routine built as a special module in ADePT.
More informationMDP Algorithms. Thomas Keller. June 20, University of Basel
MDP Algorithms Thomas Keller University of Basel June 20, 208 Outline of this lecture Markov decision processes Planning via determinization Monte-Carlo methods Monte-Carlo Tree Search Heuristic Search
More informationCISC 889 Bioinformatics (Spring 2004) Phylogenetic Trees (II)
CISC 889 ioinformatics (Spring 004) Phylogenetic Trees (II) Character-based methods CISC889, S04, Lec13, Liao 1 Parsimony ased on sequence alignment. ssign a cost to a given tree Search through the topological
More informationReinforcement Learning
Reinforcement Learning Basic idea: Receive feedback in the form of rewards Agent s utility is defined by the reward function Must (learn to) act so as to maximize expected rewards Grid World The agent
More informationComputational Finance Least Squares Monte Carlo
Computational Finance Least Squares Monte Carlo School of Mathematics 2019 Monte Carlo and Binomial Methods In the last two lectures we discussed the binomial tree method and convergence problems. One
More informationDecision Analysis. Introduction. Job Counseling
Decision Analysis Max, min, minimax, maximin, maximax, minimin All good cat names! 1 Introduction Models provide insight and understanding We make decisions Decision making is difficult because: future
More informationLecture 3: Factor models in modern portfolio choice
Lecture 3: Factor models in modern portfolio choice Prof. Massimo Guidolin Portfolio Management Spring 2016 Overview The inputs of portfolio problems Using the single index model Multi-index models Portfolio
More informationIt is used when neither the TX nor RX knows anything about the statistics of the source sequence at the start of the transmission
It is used when neither the TX nor RX knows anything about the statistics of the source sequence at the start of the transmission -The code can be described in terms of a binary tree -0 corresponds to
More informationReinforcement Learning. Slides based on those used in Berkeley's AI class taught by Dan Klein
Reinforcement Learning Slides based on those used in Berkeley's AI class taught by Dan Klein Reinforcement Learning Basic idea: Receive feedback in the form of rewards Agent s utility is defined by the
More informationPredicting Electricity Pool Prices Using Hidden Markov Models
Preprints of the 9th International Symposium on Advanced Control of Chemical Processes The International Federation of Automatic Control June 7-1, 215, Whistler, British Columbia, Canada MoPoster2.7 Predicting
More informationChapter 2 Uncertainty Analysis and Sampling Techniques
Chapter 2 Uncertainty Analysis and Sampling Techniques The probabilistic or stochastic modeling (Fig. 2.) iterative loop in the stochastic optimization procedure (Fig..4 in Chap. ) involves:. Specifying
More informationGamma Distribution Fitting
Chapter 552 Gamma Distribution Fitting Introduction This module fits the gamma probability distributions to a complete or censored set of individual or grouped data values. It outputs various statistics
More information91.420/543: Artificial Intelligence UMass Lowell CS Fall 2010
91.420/543: Artificial Intelligence UMass Lowell CS Fall 2010 Lecture 17 & 18: Markov Decision Processes Oct 12 13, 2010 A subset of Lecture 9 slides from Dan Klein UC Berkeley Many slides over the course
More information(High Dividend) Maximum Upside Volatility Indices. Financial Index Engineering for Structured Products
(High Dividend) Maximum Upside Volatility Indices Financial Index Engineering for Structured Products White Paper April 2018 Introduction This report provides a detailed and technical look under the hood
More informationFINITE DIFFERENCE METHODS
FINITE DIFFERENCE METHODS School of Mathematics 2013 OUTLINE Review 1 REVIEW Last time Today s Lecture OUTLINE Review 1 REVIEW Last time Today s Lecture 2 DISCRETISING THE PROBLEM Finite-difference approximations
More informationLecture notes on risk management, public policy, and the financial system. Credit portfolios. Allan M. Malz. Columbia University
Lecture notes on risk management, public policy, and the financial system Allan M. Malz Columbia University 2018 Allan M. Malz Last updated: June 8, 2018 2 / 23 Outline Overview of credit portfolio risk
More informationComputational Finance Finite Difference Methods
Explicit finite difference method Computational Finance Finite Difference Methods School of Mathematics 2018 Today s Lecture We now introduce the final numerical scheme which is related to the PDE solution.
More informationCS360 Homework 14 Solution
CS360 Homework 14 Solution Markov Decision Processes 1) Invent a simple Markov decision process (MDP) with the following properties: a) it has a goal state, b) its immediate action costs are all positive,
More informationStandard Normal, Inverse Normal and Sampling Distributions
Standard Normal, Inverse Normal and Sampling Distributions Section 5.5 & 6.6 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Department of Mathematics University of Houston Lecture 9-3339 Cathy
More informationLecture 22: Dynamic Filtering
ECE 830 Fall 2011 Statistical Signal Processing instructor: R. Nowak Lecture 22: Dynamic Filtering 1 Dynamic Filtering In many applications we want to track a time-varying (dynamic) phenomenon. Example
More informationWEALTH MANAGEMENT AND INVESTMENT ADVISORS
Portfolio Strategies Our investment strategies are built specifically to match your financial plan. Starting with your goals, cash flow needs and time-horizon, we create a customized portfolio that seeks
More informationAppendix A. Additional estimation results for section 5.
Appendix A. Additional estimation results for section 5. This appendix presents detailed estimation results discussed in section 5. Table A.1 shows coefficient estimates for the regression of the probability
More informationComparison of design-based sample mean estimate with an estimate under re-sampling-based multiple imputations
Comparison of design-based sample mean estimate with an estimate under re-sampling-based multiple imputations Recai Yucel 1 Introduction This section introduces the general notation used throughout this
More informationEngineering Risk Benefit Analysis
Engineering Risk Benefit Analysis 1.155, 2.943, 3.577, 6.938, 10.816, 13.621, 16.862, 22.82, ES.72, ES.721 A 1. The Multistage ecision Model George E. Apostolakis Massachusetts Institute of Technology
More informationCounting Basics. Venn diagrams
Counting Basics Sets Ways of specifying sets Union and intersection Universal set and complements Empty set and disjoint sets Venn diagrams Counting Inclusion-exclusion Multiplication principle Addition
More informationMath 167: Mathematical Game Theory Instructor: Alpár R. Mészáros
Math 167: Mathematical Game Theory Instructor: Alpár R. Mészáros Midterm #1, February 3, 2017 Name (use a pen): Student ID (use a pen): Signature (use a pen): Rules: Duration of the exam: 50 minutes. By
More informationOntological Automation of Strategic Information System Planning
Ontological Automation of Strategic Information System Planning Ontological Automation of Strategic Information System Planning Kasiphan Masakul Faculty of Informatics Sripatum University Bangkok, Thailand
More informationCS 188: Artificial Intelligence
CS 188: Artificial Intelligence Markov Decision Processes Dan Klein, Pieter Abbeel University of California, Berkeley Non-Deterministic Search 1 Example: Grid World A maze-like problem The agent lives
More information56:171 Operations Research Midterm Exam Solutions October 22, 1993
56:171 O.R. Midterm Exam Solutions page 1 56:171 Operations Research Midterm Exam Solutions October 22, 1993 (A.) /: Indicate by "+" ="true" or "o" ="false" : 1. A "dummy" activity in CPM has duration
More informationThe change of correlation structure across industries: an analysis in the regime-switching framework
Kyoto University, Graduate School of Economics Research Project Center Discussion Paper Series The change of correlation structure across industries: an analysis in the regime-switching framework Masahiko
More informationFinancial Optimization ISE 347/447. Lecture 15. Dr. Ted Ralphs
Financial Optimization ISE 347/447 Lecture 15 Dr. Ted Ralphs ISE 347/447 Lecture 15 1 Reading for This Lecture C&T Chapter 12 ISE 347/447 Lecture 15 2 Stock Market Indices A stock market index is a statistic
More informationECS171: Machine Learning
ECS171: Machine Learning Lecture 15: Tree-based Algorithms Cho-Jui Hsieh UC Davis March 7, 2018 Outline Decision Tree Random Forest Gradient Boosted Decision Tree (GBDT) Decision Tree Each node checks
More informationNumerical investigation on multiclass probabilistic classification of damage location in a plate structure
Numerical investigation on multiclass probabilistic classification of damage location in a plate structure Rims Janeliukstis *, Sandris Rucevskis, Andrejs Kovalovs and Andris Chate Institute of Materials
More informationExtended Model: Posterior Distributions
APPENDIX A Extended Model: Posterior Distributions A. Homoskedastic errors Consider the basic contingent claim model b extended by the vector of observables x : log C i = β log b σ, x i + β x i + i, i
More informationEquity correlations implied by index options: estimation and model uncertainty analysis
1/18 : estimation and model analysis, EDHEC Business School (joint work with Rama COT) Modeling and managing financial risks Paris, 10 13 January 2011 2/18 Outline 1 2 of multi-asset models Solution to
More informationAccelerated Option Pricing Multiple Scenarios
Accelerated Option Pricing in Multiple Scenarios 04.07.2008 Stefan Dirnstorfer (stefan@thetaris.com) Andreas J. Grau (grau@thetaris.com) 1 Abstract This paper covers a massive acceleration of Monte-Carlo
More informationHidden Markov Models. Slides by Carl Kingsford. Based on Chapter 11 of Jones & Pevzner, An Introduction to Bioinformatics Algorithms
Hidden Markov Models Slides by Carl Kingsford Based on Chapter 11 of Jones & Pevzner, An Introduction to Bioinformatics Algorithms Eukaryotic Genes & Exon Splicing Prokaryotic (bacterial) genes look like
More informationCSEP 573: Artificial Intelligence
CSEP 573: Artificial Intelligence Markov Decision Processes (MDP)! Ali Farhadi Many slides over the course adapted from Luke Zettlemoyer, Dan Klein, Pieter Abbeel, Stuart Russell or Andrew Moore 1 Outline
More informationPakes (1986): Patents as Options: Some Estimates of the Value of Holding European Patent Stocks
Pakes (1986): Patents as Options: Some Estimates of the Value of Holding European Patent Stocks Spring 2009 Main question: How much are patents worth? Answering this question is important, because it helps
More information2.1 Mathematical Basis: Risk-Neutral Pricing
Chapter Monte-Carlo Simulation.1 Mathematical Basis: Risk-Neutral Pricing Suppose that F T is the payoff at T for a European-type derivative f. Then the price at times t before T is given by f t = e r(t
More informationDynamic Programming cont. We repeat: The Dynamic Programming Template has three parts.
Page 1 Dynamic Programming cont. We repeat: The Dynamic Programming Template has three parts. Subproblems Sometimes this is enough if the algorithm and its complexity is obvious. Recursion Algorithm Must
More informationECO220Y Continuous Probability Distributions: Normal Readings: Chapter 9, section 9.10
ECO220Y Continuous Probability Distributions: Normal Readings: Chapter 9, section 9.10 Fall 2011 Lecture 8 Part 2 (Fall 2011) Probability Distributions Lecture 8 Part 2 1 / 23 Normal Density Function f
More informationTHE investment in stock market is a common way of
PROJECT REPORT, MACHINE LEARNING (COMP-652 AND ECSE-608) MCGILL UNIVERSITY, FALL 2018 1 Comparison of Different Algorithmic Trading Strategies on Tesla Stock Price Tawfiq Jawhar, McGill University, Montreal,
More information[D7] PROBABILITY DISTRIBUTION OF OUTSTANDING LIABILITY FROM INDIVIDUAL PAYMENTS DATA Contributed by T S Wright
Faculty and Institute of Actuaries Claims Reserving Manual v.2 (09/1997) Section D7 [D7] PROBABILITY DISTRIBUTION OF OUTSTANDING LIABILITY FROM INDIVIDUAL PAYMENTS DATA Contributed by T S Wright 1. Introduction
More informationAdvanced Micro 1 Lecture 14: Dynamic Games Equilibrium Concepts
Advanced Micro 1 Lecture 14: Dynamic Games quilibrium Concepts Nicolas Schutz Nicolas Schutz Dynamic Games: quilibrium Concepts 1 / 79 Plan 1 Nash equilibrium and the normal form 2 Subgame-perfect equilibrium
More informationRobust Dual Dynamic Programming
1 / 18 Robust Dual Dynamic Programming Angelos Georghiou, Angelos Tsoukalas, Wolfram Wiesemann American University of Beirut Olayan School of Business 31 May 217 2 / 18 Inspired by SDDP Stochastic optimization
More informationDATA MINING FOR OPTIMAL GAMBLING.
DATA MINING FOR OPTIMAL GAMBLING. Gabriele Torre 1 and Fabrizio Malfanti 2 1 Dipartimento di Matematica, Università degli Studi di Genova, via Dodecaneso 35, 16146, Genova, Italy. (e-mail: torre@dima.unige.it)
More informationStochastic Dual Dynamic Programming
1 / 43 Stochastic Dual Dynamic Programming Operations Research Anthony Papavasiliou 2 / 43 Contents [ 10.4 of BL], [Pereira, 1991] 1 Recalling the Nested L-Shaped Decomposition 2 Drawbacks of Nested Decomposition
More informationFE570 Financial Markets and Trading. Stevens Institute of Technology
FE570 Financial Markets and Trading Lecture 6. Volatility Models and (Ref. Joel Hasbrouck - Empirical Market Microstructure ) Steve Yang Stevens Institute of Technology 10/02/2012 Outline 1 Volatility
More informationOption-Implied Correlations, Factor Models, and Market Risk
Option-Implied Correlations, Factor Models, and Market Risk Adrian Buss Lorenzo Schönleber Grigory Vilkov INSEAD Frankfurt School Frankfurt School of Finance & Management of Finance & Management 17th November
More informationReal Business Cycles (Solution)
Real Business Cycles (Solution) Exercise: A two-period real business cycle model Consider a representative household of a closed economy. The household has a planning horizon of two periods and is endowed
More informationEnforcing monotonicity of decision models: algorithm and performance
Enforcing monotonicity of decision models: algorithm and performance Marina Velikova 1 and Hennie Daniels 1,2 A case study of hedonic price model 1 Tilburg University, CentER for Economic Research,Tilburg,
More informationLecture 3: Return vs Risk: Mean-Variance Analysis
Lecture 3: Return vs Risk: Mean-Variance Analysis 3.1 Basics We will discuss an important trade-off between return (or reward) as measured by expected return or mean of the return and risk as measured
More information6.254 : Game Theory with Engineering Applications Lecture 3: Strategic Form Games - Solution Concepts
6.254 : Game Theory with Engineering Applications Lecture 3: Strategic Form Games - Solution Concepts Asu Ozdaglar MIT February 9, 2010 1 Introduction Outline Review Examples of Pure Strategy Nash Equilibria
More informationCSE 473: Artificial Intelligence
CSE 473: Artificial Intelligence Markov Decision Processes (MDPs) Luke Zettlemoyer Many slides over the course adapted from Dan Klein, Stuart Russell or Andrew Moore 1 Announcements PS2 online now Due
More informationProject Trainee : Abhinav Yellanki 5 th year Integrated M.Sc. Student Mathematics and Computing Indian Institute of Technology, Kharagpur
SIMULATION MODELLING OF ASSETS AND LIABILITI ES OF A BANK Project Trainee : Abhinav Yellanki 5 th year Integrated M.Sc. Student Mathematics and Computing Indian Institute of Technology, Kharagpur Project
More informationCourse Information and Introduction
August 20, 2015 Course Information 1 Instructor : Email : arash.rafiey@indstate.edu Office : Root Hall A-127 Office Hours : Tuesdays 12:00 pm to 1:00 pm in my office (A-127) 2 Course Webpage : http://cs.indstate.edu/
More informationFinal Examination CS540: Introduction to Artificial Intelligence
Final Examination CS540: Introduction to Artificial Intelligence December 2008 LAST NAME: FIRST NAME: Problem Score Max Score 1 15 2 15 3 10 4 20 5 10 6 20 7 10 Total 100 Question 1. [15] Probabilistic
More informationRichardson Extrapolation Techniques for the Pricing of American-style Options
Richardson Extrapolation Techniques for the Pricing of American-style Options June 1, 2005 Abstract Richardson Extrapolation Techniques for the Pricing of American-style Options In this paper we re-examine
More informationCS 461: Machine Learning Lecture 8
CS 461: Machine Learning Lecture 8 Dr. Kiri Wagstaff kiri.wagstaff@calstatela.edu 2/23/08 CS 461, Winter 2008 1 Plan for Today Review Clustering Reinforcement Learning How different from supervised, unsupervised?
More informationThe exam is closed book, closed notes except a two-page crib sheet. Non-programmable calculators only.
CS 188 Spring 2011 Introduction to Artificial Intelligence Final Exam INSTRUCTIONS You have 3 hours. The exam is closed book, closed notes except a two-page crib sheet. Non-programmable calculators only.
More informationRisk and Return and Portfolio Theory
Risk and Return and Portfolio Theory Intro: Last week we learned how to calculate cash flows, now we want to learn how to discount these cash flows. This will take the next several weeks. We know discount
More informationMatrix Factorization 1
Matrix Factorization 1 Recovering latent factors in a matrix m columns v11 n rows vij vnm 2 Recovering latent factors in a matrix K * m x1 y1 x2 y2.... a1 a2.. am b1 b2 bm ~ v11 n * K vij xn yn vnm 3 What
More informationLog-linear Modeling Under Generalized Inverse Sampling Scheme
Log-linear Modeling Under Generalized Inverse Sampling Scheme Soumi Lahiri (1) and Sunil Dhar (2) (1) Department of Mathematical Sciences New Jersey Institute of Technology University Heights, Newark,
More informationChapter 13 Decision Analysis
Problem Formulation Chapter 13 Decision Analysis Decision Making without Probabilities Decision Making with Probabilities Risk Analysis and Sensitivity Analysis Decision Analysis with Sample Information
More information