Huber smooth M-estimator. Mâra Vçliòa, Jânis Valeinis. University of Latvia. Sigulda,

Similar documents
Much of what appears here comes from ideas presented in the book:

Chapter 4: Asymptotic Properties of MLE (Part 3)

Statistical analysis and bootstrapping

Chapter 8: Sampling distributions of estimators Sections

Window Width Selection for L 2 Adjusted Quantile Regression

Chapter 7 - Lecture 1 General concepts and criteria

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

IEOR E4602: Quantitative Risk Management

Jackknife Empirical Likelihood Inferences for the Skewness and Kurtosis

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

1 Bayesian Bias Correction Model

Applied Statistics I

STRESS-STRENGTH RELIABILITY ESTIMATION

Lecture 23. STAT 225 Introduction to Probability Models April 4, Whitney Huang Purdue University. Normal approximation to Binomial

EE641 Digital Image Processing II: Purdue University VISE - October 29,

Definition 9.1 A point estimate is any function T (X 1,..., X n ) of a random sample. We often write an estimator of the parameter θ as ˆθ.

ECE 295: Lecture 03 Estimation and Confidence Interval

Lecture Notes 6. Assume F belongs to a family of distributions, (e.g. F is Normal), indexed by some parameter θ.

Financial Time Series and Their Characteristics

MATH 3200 Exam 3 Dr. Syring

Financial Risk Management

A New Hybrid Estimation Method for the Generalized Pareto Distribution

Likelihood Methods of Inference. Toss coin 6 times and get Heads twice.

Robust X control chart for monitoring the skewed and contaminated process

Lecture 10: Point Estimation

PROBABILITY AND STATISTICS

Normal Distribution. Notes. Normal Distribution. Standard Normal. Sums of Normal Random Variables. Normal. approximation of Binomial.

Chapter 7: Estimation Sections

Probability & Statistics

EVA Tutorial #1 BLOCK MAXIMA APPROACH IN HYDROLOGIC/CLIMATE APPLICATIONS. Rick Katz

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

Business Statistics 41000: Probability 3

Random Variables and Probability Distributions

Back to estimators...

Modeling Co-movements and Tail Dependency in the International Stock Market via Copulae

Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing Examples

Chapter 7: Estimation Sections

Analysis of truncated data with application to the operational risk estimation

1. You are given the following information about a stationary AR(2) model:

Chapter 7: Point Estimation and Sampling Distributions

Point Estimation. Some General Concepts of Point Estimation. Example. Estimator quality

Bivariate Birnbaum-Saunders Distribution

Modelling Returns: the CER and the CAPM

STAT/MATH 395 PROBABILITY II

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

Chapter 7 Sampling Distributions and Point Estimation of Parameters

Qualifying Exam Solutions: Theoretical Statistics

Internet Appendix for Asymmetry in Stock Comovements: An Entropy Approach

Chapter 4 Continuous Random Variables and Probability Distributions

Statistical Tables Compiled by Alan J. Terry

Estimation after Model Selection

MVE051/MSG Lecture 7

Point Estimation. Principle of Unbiased Estimation. When choosing among several different estimators of θ, select one that is unbiased.

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is:

Chapter 6: Point Estimation

Computational Finance. Computational Finance p. 1

A potentially useful approach to model nonlinearities in time series is to assume different behavior (structural break) in different subsamples

Vladimir Spokoiny (joint with J.Polzehl) Varying coefficient GARCH versus local constant volatility modeling.

NEWCASTLE UNIVERSITY SCHOOL OF MATHEMATICS, STATISTICS & PHYSICS SEMESTER 1 SPECIMEN 2 MAS3904. Stochastic Financial Modelling. Time allowed: 2 hours

Bias Reduction Using the Bootstrap

Why Are Big Banks Getting Bigger?

Course information FN3142 Quantitative finance

M.I.T Fall Practice Problems

The Bernoulli distribution

Chapter 8. Introduction to Statistical Inference

Optimally Thresholded Realized Power Variations for Lévy Jump Diffusion Models

continuous rv Note for a legitimate pdf, we have f (x) 0 and f (x)dx = 1. For a continuous rv, P(X = c) = c f (x)dx = 0, hence

Practice Exam 1. Loss Amount Number of Losses

Financial Econometrics

SYSM 6304 Risk and Decision Analysis Lecture 2: Fitting Distributions to Data

Stochastic Differential Equations in Finance and Monte Carlo Simulations

1. Covariance between two variables X and Y is denoted by Cov(X, Y) and defined by. Cov(X, Y ) = E(X E(X))(Y E(Y ))

Chapter 8: Sampling distributions of estimators Sections

Commonly Used Distributions

Which GARCH Model for Option Valuation? By Peter Christoffersen and Kris Jacobs

Chapter 4 Continuous Random Variables and Probability Distributions

Continuous random variables

An Approach for Comparison of Methodologies for Estimation of the Financial Risk of a Bond, Using the Bootstrapping Method

12. Conditional heteroscedastic models (ARCH) MA6622, Ernesto Mordecki, CityU, HK, 2006.

Chapter 7: SAMPLING DISTRIBUTIONS & POINT ESTIMATION OF PARAMETERS

Chapter 5 Univariate time-series analysis. () Chapter 5 Univariate time-series analysis 1 / 29

PARAMETRIC AND NON-PARAMETRIC BOOTSTRAP: A SIMULATION STUDY FOR A LINEAR REGRESSION WITH RESIDUALS FROM A MIXTURE OF LAPLACE DISTRIBUTIONS

Booth School of Business, University of Chicago Business 41202, Spring Quarter 2014, Mr. Ruey S. Tsay. Solutions to Midterm

Parameterized Expectations

Asymmetric Price Transmission: A Copula Approach

STA 532: Theory of Statistical Inference

(a) Is it possible for the rate of exit from OLF into E tobethesameastherateof exit from U into E? Why or why not?

Review for Final Exam Spring 2014 Jeremy Orloff and Jonathan Bloom

Modeling the dependence between a Poisson process and a continuous semimartingale

Statistical Inference and Methods

IEOR 165 Lecture 1 Probability Review

A Robust Test for Normality

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی

Comparing the Means of. Two Log-Normal Distributions: A Likelihood Approach

Homework Problems Stat 479

10/1/2012. PSY 511: Advanced Statistics for Psychological and Behavioral Research 1

Alexander Marianski August IFRS 9: Probably Weighted and Biased?

An Improved Skewness Measure

Financial Risk Forecasting Chapter 9 Extreme Value Theory

Lecture Note 8 of Bus 41202, Spring 2017: Stochastic Diffusion Equation & Option Pricing

Transcription:

University of Latvia Sigulda, 28.05.2011

Contents M-estimators Huber estimator Smooth M-estimator Empirical likelihood method for M-estimators

Introduction Aim: robust estimation of location parameter Huber M-estimator (1964) - well known robust location estimator Owen (1988) introduced empirical likelihood method, also applicable to M-estimators Hampel (2011) proposed a smoothed version of Huber estimator Work in progress Two sample problem: empirical likelihood based method for a difference of smoothed Huber estimators (Valeinis, Velina, Luta: abstract for ICORS 2011 conference)

M-estimators M-estimator Let X 1, X 2,..., X n iid, X 1 F. An M-estimator T n is defined as a solution of n ρ(x i, t) = ρ(x, t)df n (x), (1) i=1 for a specific funtion ρ where F n is the empirical CDF. If ρ is differentiable in t, then (1) is minimized by the solution of where ψ(x, t) = tρ(x, t). n ψ(x i, t) = 0, i=1

M-estimators Examples Mean. ψ(x, t) = x t gives T n = X. ML estimator. ψ(x, θ) = d dθ log f(x, θ) for a class of density functions f(x, θ), gives T n is the root of likelihood equation ( n ) d dθ log f(x i, θ) = 0. i=1 Median. ψ(x, t) = ψ 0 (x t), ψ 0 (z) = k sgn(z), k > 0.

Huber estimator Huber estimator for location parameter µ Huber (1964) combined examples of mean and median. Let F have a symmetric density f µ,σ (x) = 1 σ f ( x µ σ assume σ = 1. Then M-estimator for the location parameter µ is defined as n ( ) Xi t ψ = 0. (2) σ i=1 ), Huber M-estimator is defined by the function ψ in (2): k, x k ψ k (x) = x, k x k k, x k. (3)

Huber estimator Huber's motivaton: Unrestricted ψ-functions have undesired properties (unstable to outliers); Cosider the limiting values of k in ψ k and their respective M-estimators: If k, then ψ k is mean; If k 0, then ψ k is median. k is a tuning constant determining the degree of robustness. Huber estimator has minimax assymptotic variance for class of distribution functions (1 ɛ)φ(x) + ɛh(x), where φ is pdf of N(0, 1) and h is a symmetric density.

Huber estimator Scaled estimator of location In reality σ is not known, thus a robust estimate of σ should be used. A common choice is MAD. MAD S n = MAD = median( X i median(x i ) ). Robust estimator is acquired, even in presence of outliers (up to 50% of the sample).

Smooth M-estimator Smoothed M-estimator (Hampel, 2011) For a general ψ-function of an M-estimator define ψ(x) = ψ(x + u)dq n (u), (4) where Q n may be chosen as a the distribution of the initial M-estimator Q n can be approximated by N(0, V/n), where V is assymptotic variance of the M-estimator. Need to specify distribution under which the assymptotic variance is computed. The smoothing prinicple can be applied to ψ functions already smooth.

Smooth M-estimator Smoothed Huber estimator The ψ-function of the smoothed Huber estimator defined by ψ = ψ k can be written in closed form as ( ) ( ( )) x k x + k ψ k (x) = kφ k 1 Φ σ n σ n ( ) ( )) x + k x k +x Φ Φ σ n σ n ( ) ( )) x + k x k +σ n (φ φ, (5) where σ n = V/n, and Φ and φ denote the cdf and pdf of N(0, 1). σ n σ n

Smooth M-estimator Example 1.0 0.5 0.0 0.5 1.0 1.0 0.5 0.0 0.5 1.0 4 2 0 2 4 4 2 0 2 4 (a) (a) ψ function of Huber M-estimate; (b) ψ function of smoothed Huber M-estimate. k=1.35. (b)

Empirical likelihood method for M-estimators Empirical likelihood method for M-estimators Owen (1988) showed that EL method can be applied to certain M-estimators, including Huber estimator. Nonparametric Wilk's theorem applies thus EL based confidence intervals for Huber estimate can be obtained. Tsao, Zhu (2001) showed that EL based confidence intervals preserves robustness.

Empirical likelihood method for M-estimators EL confidence bands for Huber estimator Empirical likelihood ratio for parameter t R(t) = sup{ n ω i n n ω i ψ(x i, t) = 0, ω i 0, ω i = 1} i=1 i=1 i=1 is maximized by ω i (λ), where ω i (λ) = {n(1 + λz i )} 1, and Z i = ψ(x i, t) and λ follows from n 1 Zi /(1 + λz i ) = 0.

Empirical likelihood method for M-estimators 2lnL 0 5 10 15 20 EL vid.vert EL Huber 2lnL 0 5 10 15 20 EL vid.vert EL Huber 2 1 0 1 2 2 1 0 1 2 3 4 5 Figure: EL -2*ln,(a) N(0, 3) (b) 0.95 N(0, 3) + 0.05 N(20, 3)

Empirical likelihood method for M-estimators Simulation results for one sample problem Table: Huber estimation for location parameter and its EL confidence bands, alpha=0.05 N(0, 3) 0.95 N(0, 3) + 0.05 N(20, 3) sample len estimate len estimate n=50 EL.huber 0.494 EL.huber -0.055 EL.huber 1.706 EL.huber 0.159 EL.mean 0.492 EL.mean -0.064 EL.mean 3.14 EL.mean 1.008 t-test 0.506 mean -0.064 t-test 3.117 mean 1.008 z-test 0.554 huber -0.076 z-test 0.554 huber 0.159 Bootstrap 0.497 Bootstrap 3.057 n=20 EL.huber 0.667 EL.huber -0.167 EL.huber 2.478 EL.huber -0.441 EL.mean 0.667 EL.mean -0.167 EL.mean 4.894 EL.mean 0.498 t-test 0.732 mean -0.167 t-test 4.938 mean 0.498 z-test 0.877 huber -0.643 z-test 0.877 huber -0.441 Bootstrap 0.699 Bootstrap 4.583 n=10 EL.huber 1.001 EL.huber -0.067 EL.huber 4.303 EL.huber -0.189 EL.mean 1.001 EL.mean -0.067 EL.mean 9.68 EL.mean 1.008 t-test 1.239 mean -0.067 t-test 11.494 mean 1.799 z-test 1.24 huber -0.201 z-test 1.24 huber -0.189 Bootstrap 1.039 Bootstrap 9.74

Empirical likelihood method for M-estimators Two sample EL problem Consider empirical likelihood-based method for the difference of smoothed Huber estimators. Given two independent samples X and Y with distribution functions F 1 and F 2, respectively, we have two unbiased estimating functions: E F1 w 1 (X, θ 0, ) = 0, E F2 w 2 (Y, θ 0, ) = 0, where is the parameter of interest and θ 0 is a nuisance parameter. Specifically, = θ 1 θ 0 and ( ) ( ) X θ0 Y + θ0 w 1 (X, θ 0, ) = ψ w 2 (Y, θ 0, ) = ψ, ˆσ 1 ˆσ 2 where ˆσ 1 and ˆσ 2 are scale estimators, and ψ corresponds to the smoothed Huber estimator.

Empirical likelihood method for M-estimators Simulation results for two sample problem Conisder two models: Y 1 (1 ɛ)gamma(α = 5; σ = 1) + ɛuniform[0; 50] Y 2 Gamma(α = 1; σ = 5) Table: Coverage accuray and average confidence interval lengths based on 1000 replicates, n 1 = n 2 = 50 t.int EL.hub1 EL.hub2 Boot1 Boot2 acc ave acc len acc len acc len acc len σ = 5 0.62 3.05 0.66 2.99 0.56 2.83 0.36 2.98 0.36 2.98 σ = 6 0.69 3.56 0.73 3.51 0.65 3.34 0.38 3.46 0.38 3.47 σ = 7 0.74 4.09 0.77 4.04 0.72 3.85 0.44 3.97 0.45 3.99 σ = 8 0.78 4.62 0.81 4.56 0.76 4.39 0.48 4.49 0.48 4.50 σ = 9 0.81 5.19 0.84 5.13 0.80 4.95 0.50 5.00 0.50 5.02

Empirical likelihood method for M-estimators Thank you for your attention!