Wrong Skewness and Finite Sample Correction in Parametric Stochastic Frontier Models

Similar documents
Wrong Skewness and Finite Sample Correction in Parametric Stochastic Frontier Models

On the Distributional Assumptions in the StoNED model

The Stochastic Approach for Estimating Technical Efficiency: The Case of the Greek Public Power Corporation ( )

A Monte Carlo Study of Ranked Efficiency Estimates from Frontier Models

Volume 37, Issue 2. Handling Endogeneity in Stochastic Frontier Analysis

Statistical Evidence and Inference

Alternative Technical Efficiency Measures: Skew, Bias and Scale

WORKING PAPERS IN ECONOMICS. No 449. Pursuing the Wrong Options? Adjustment Costs and the Relationship between Uncertainty and Capital Accumulation

Mean-Variance Analysis

Pseudolikelihood estimation of the stochastic frontier model SFB 823. Discussion Paper. Mark Andor, Christopher Parmeter

Empirical Tests of Information Aggregation

Conditional Investment-Cash Flow Sensitivities and Financing Constraints

Solving dynamic portfolio choice problems by recursing on optimized portfolio weights or on the value function?

Chapter 5 Univariate time-series analysis. () Chapter 5 Univariate time-series analysis 1 / 29

Mossin s Theorem for Upper-Limit Insurance Policies

Effective Tax Rates and the User Cost of Capital when Interest Rates are Low

Human capital and the ambiguity of the Mankiw-Romer-Weil model

Window Width Selection for L 2 Adjusted Quantile Regression

Consistent estimators for multilevel generalised linear models using an iterated bootstrap

Real Wage Rigidities and Disin ation Dynamics: Calvo vs. Rotemberg Pricing

OPTIMAL INCENTIVES IN A PRINCIPAL-AGENT MODEL WITH ENDOGENOUS TECHNOLOGY. WP-EMS Working Papers Series in Economics, Mathematics and Statistics

Asset Pricing under Information-processing Constraints

Time Invariant and Time Varying Inefficiency: Airlines Panel Data

Supply-side effects of monetary policy and the central bank s objective function. Eurilton Araújo

1. Money in the utility function (continued)

Behavioral Finance and Asset Pricing

Principles of Econometrics Mid-Term

Package semsfa. April 21, 2018

Multivariate Statistics Lecture Notes. Stephen Ansolabehere

Research of the impact of agricultural policies on the efficiency of farms

Volume 30, Issue 1. Samih A Azar Haigazian University

Fuel-Switching Capability

FS January, A CROSS-COUNTRY COMPARISON OF EFFICIENCY OF FIRMS IN THE FOOD INDUSTRY. Yvonne J. Acheampong Michael E.

Heterogeneity in Returns to Wealth and the Measurement of Wealth Inequality 1

ECON Micro Foundations

Published: 14 October 2014

Richardson Extrapolation Techniques for the Pricing of American-style Options

Subject CS1 Actuarial Statistics 1 Core Principles. Syllabus. for the 2019 exams. 1 June 2018

Statistical Analysis of Data from the Stock Markets. UiO-STK4510 Autumn 2015

Omitted Variables Bias in Regime-Switching Models with Slope-Constrained Estimators: Evidence from Monte Carlo Simulations

Appendix for The Long-Run Risks Model and Aggregate Asset Prices: An Empirical Assessment

Nonparametric Estimation of a Hedonic Price Function

STOCK RETURNS AND INFLATION: THE IMPACT OF INFLATION TARGETING

MEASURING PORTFOLIO RISKS USING CONDITIONAL COPULA-AR-GARCH MODEL

Introductory Econometrics for Finance

Equity, Vacancy, and Time to Sale in Real Estate.

Government expenditure and Economic Growth in MENA Region

Expected Utility and Risk Aversion

Wealth E ects and Countercyclical Net Exports

Measuring the Wealth of Nations: Income, Welfare and Sustainability in Representative-Agent Economies

An Improved Saddlepoint Approximation Based on the Negative Binomial Distribution for the General Birth Process

Upward pricing pressure of mergers weakening vertical relationships

Sequential Decision-making and Asymmetric Equilibria: An Application to Takeovers

The quantile regression approach to efficiency measurement: insights from Monte Carlo Simulations

On the Distribution and Its Properties of the Sum of a Normal and a Doubly Truncated Normal

Growth and Welfare Maximization in Models of Public Finance and Endogenous Growth

Uncertainty and Capital Accumulation: Empirical Evidence for African and Asian Firms

Statistical Models and Methods for Financial Markets

Investment is one of the most important and volatile components of macroeconomic activity. In the short-run, the relationship between uncertainty and

A Test of the Normality Assumption in the Ordered Probit Model *

Introduction to the Maximum Likelihood Estimation Technique. September 24, 2015

The E ects of Adjustment Costs and Uncertainty on Investment Dynamics and Capital Accumulation

SOLUTION PROBLEM SET 3 LABOR ECONOMICS

Copula-Based Pairs Trading Strategy

Are Financial Markets Stable? New Evidence from An Improved Test of Financial Market Stability and the U.S. Subprime Crisis

Booms and Busts in Asset Prices. May 2010

Performance of Statistical Arbitrage in Future Markets

E cient Minimum Wages

Journal of Economics and Financial Analysis, Vol:1, No:1 (2017) 1-13

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی

Efficiency Measurement with the Weibull Stochastic Frontier*

Extreme Return-Volume Dependence in East-Asian. Stock Markets: A Copula Approach

Conditional Investment-Cash Flow Sensitivities and Financing Constraints

Quantitative Techniques Term 2

Faster solutions for Black zero lower bound term structure models

E ects of di erences in risk aversion on the. distribution of wealth

Applying regression quantiles to farm efficiency estimation

STATE UNIVERSITY OF NEW YORK AT ALBANY Department of Economics. Ph. D. Comprehensive Examination: Macroeconomics Spring, 2013

Banking Concentration and Fragility in the United States

Estimation of a Ramsay-Curve IRT Model using the Metropolis-Hastings Robbins-Monro Algorithm

Modelling the Sharpe ratio for investment strategies

Online Appendix B. Assessing Sale Strategies in Online Markets using Matched Listings. By Einav, Kuchler, Levin, and Sundaresan

Carmen M. Reinhart b. Received 9 February 1998; accepted 7 May 1998

Estimation of a parametric function associated with the lognormal distribution 1

PARAMETRIC AND NON-PARAMETRIC BOOTSTRAP: A SIMULATION STUDY FOR A LINEAR REGRESSION WITH RESIDUALS FROM A MIXTURE OF LAPLACE DISTRIBUTIONS

Introduction to Sequential Monte Carlo Methods

Returns to Education and Wage Differentials in Brazil: A Quantile Approach. Abstract

Small Sample Performance of Instrumental Variables Probit Estimators: A Monte Carlo Investigation

Market Risk Analysis Volume I

Liquidity skewness premium

The Economics of State Capacity. Ely Lectures. Johns Hopkins University. April 14th-18th Tim Besley LSE

Monotonicity in Asset Returns: New Tests with Applications to the Term Structure, the CAPM and Portfolio Sorts

2018 outlook and analysis letter

Consumption and Portfolio Choice under Uncertainty

For Online Publication Only. ONLINE APPENDIX for. Corporate Strategy, Conformism, and the Stock Market

Questions of Statistical Analysis and Discrete Choice Models

The Duration Derby: A Comparison of Duration Based Strategies in Asset Liability Management

Idiosyncratic risk, insurance, and aggregate consumption dynamics: a likelihood perspective

Web Appendix. Are the effects of monetary policy shocks big or small? Olivier Coibion

Amath 546/Econ 589 Univariate GARCH Models: Advanced Topics

Transcription:

Wrong Skewness and Finite Sample Correction in Parametric Stochastic Frontier Models Qu Feng y, William C. Horrace z, Guiying Laura Wu x October, 05 Abstract In parametric stochastic frontier models, the composed error is speci ed as the sum of a twosided noise component and a one-sided ine ciency component, which is usually assumed to be half-normal, implying that the error distribution is skewed in one direction. In practice, however, estimation residuals may display skewness in the wrong direction. Model re-speci cation or pulling a new sample is often prescribed. Since wrong skewness is considered a nite sample problem, this paper proposes a nite sample adjustment to existing estimators to obtain the desired direction of residual skewness. This provides another empirical approach to dealing with the so-called wrong skewness problem. JEL Classi cations: C3, C3, D4 Keywords: Stochastic frontier model, skewness, MLE, constrained estimators, BIC We thank William Greene for providing the airlines dataset. The comments of Peter Schmidt, Robin Sickles, Daniel Henderson and the particiants of the 0 Conference in Honor of Peter Schmidt, Houston TX, and Singapore Economic Review Conference 05 are appreciated. Financial support from the MOE AcRF Tier research grant at anyang Technological University is gratefully acknowledged. y Email: qfeng@ntu.edu.sg, Tel: +65 659 543. Division of Economics, School of Humanities and Social Sciences, 4 anyang Drive, Singapore 63733. z Email: whorrace@maxwell.syr.edu, Tel: 35-443-906. Center for Policy Research, 46 Eggers Hall, Syracuse, Y 344-00 x Email: guiying.wu@ntu.edu.sg, Tel: +65 659 553. Division of Economics, School of Humanities and Social Sciences, 4 anyang Drive, Singapore 63733.

Introduction In parametric stochastic frontier models, the error term is composed as the sum of a two-sided noise component and a one-sided ine ciency component. For cross-sectional models, the noise distribution is assumed normal, while the ine ciency distribution is usually assumed to be halfnormal (Aigner, Lovell and Schmidt, 977), exponential (Meeusen and van den Broeck, 977; Aigner, Lovell and Schmidt, 977), or truncated normal (Stevenson, 980). It is sometimes gamma (Stevenson, 980; Greene, 980). For surveys, see Greene (007) and Kumbhakar and Lovell (000). In the widely used normal-half normal speci cation of the stochastic frontier production function model, the skewness of the composed error is negative, and parameters can be estimated by maximum likelihood estimation (MLE) or corrected ordinary least squares (COLS). Waldman (98) shows that when the skewness of the ordinary least squares (OLS) residuals is positive, OLS is a local maximum of the likelihood function, and estimated ine ciency is zero in the sample. This "wrong skewness" phenomenon is widely documented in the literature and is often regarded as an estimation failure. 3 When it occurs, researchers are advised to either obtain a new sample or respecify the model (Li, 996; Carree, 00; Almanidis and Sickles, 0; Almanidis, Qian and Sickles, 04; Hafner, Manner and Simar, 03). Simar and Wilson (00) argue that "wrong skewness" is not an estimation or modelling failure, but a nite sample problem that is most likely to occur when the signal-to-noise ratio (the variance ratio of the ine ciency component to the composite error) is small. That is, wrong skewness may not be an indication that the model is wrong or that ine ciency does not exist in the population. They propose a bootstrap method (called "bagging") to construct con dence intervals for model parameters and expected ine ciency which have higher coverage than traditional intervals, regardless of residual skewness direction. The sample under study can still be used to infer the model parameters. We follow Simar and Wilson s (00) view that wrong skewness is a consequence of a small signal-to-noise ratio in nite samples. However, instead of the bagging approach of Simar and Wilson (00), this paper provides a nite sample adjustment to existing estimators in the presence The skewness of the composed error is positive in the stochastic frontier cost function model. We use the terminology COLS following Olson, Schmidt and Waldman (98). COLS is also called MOLS. See Greene (007). Greene (007, p.3) claims "In this instance, the OLS results are the MLEs, and consequently, one must estimate the one-sided terms as 0." 3 For example, estimating the variance parameters in COLS is invalid in this case. However, as emphasized by Greene (007, note 9), this problem does not carry over to other model speci cations. Several sources can lead to wrong skewness.

of wrong skewness. That is, we impose a negative residual skewness constraint in the MLE (or COLS) algorithm. A natural candidate for this constraint is the upper bound of the population skew, which is a monotonic function of the positive lower bound of the signal-to-noise ratio in the half-normal model. However, the constraint is non-linear in the parameters of interest, complicating computation of the optimum. Therefore, a linearization approximation of the constraint is proposed. Additionally, a model selection approach is proposed to determine the lower bound of the signalto-noise ratio used in the constraint. A shortcoming of the approach is that in nite samples the linear approximation may not be accurate enough to guarantee a negative sign of residual skewness. In this case, additional nite sample adjustment is required. Monte Carlo experiments suggest that our correction becomes more reliable when the true signal-to-noise ratio increases. The possible failure of correcting the sign of residual skewness using the linearized constraint illustrates a trade-o between computational complexity and accuracy. Using the original non-linear constraint avoids this issue, but the computational convenience of our approach, shown in the Monte Carlo experiments and empirical example below, would be lost. The proposed nite sample adjustment provides a point estimate with a correct sign of residual skewness that can be used in applied research. Since wrong skewness can occur fairly regularly (even when e ciency may exist in the population under study), the nite sample adjustment is attractive particularly in cases where the ine ciency distribution is half-normal. It is worthwhile to note that the proposed adjustment is only needed in nite samples, for as the sample size increases wrong skewness is less likely to be an issue when the signal-to-noise ratio is sizable. This rest of this paper is organized as follows. The next section discusses the wrong skewness issue in the literature. In Section 3, we propose a nite sample correction approach. To simplify computation of the proposed constrained estimation, a linearized version of the constraint is used, so that constrained MLE (or COLS) can be easily implemented in most software packages. The constrained estimators are discussed in Section 4. In Section 5, Monte Carlo experiments are conducted to study the properties of constrained COLS. An empirical example is used to illustrate the proposed approach in Section 6. The last section concludes.

Wrong Skewness Issue A stochastic production frontier (SPF) model for a cross-sectional sample of size is: with composed error " i = v i y i = x 0 i + " i ; i = ; ; ; () u i : The disturbance v i is assumed iid(0; v). Ine ciency of rm i is characterized by u i 0. In the SPF literature, u i is usually assumed half-normal jiid(0; u)j (Aigner, Lovell and Schmidt, 977; Wang and Schmidt, 009), and independent of v i, with variance V ar(u i ) = u. The rst component of the p vector x i is, so the intercept term is contained into the p slope parameter vector. As in Aigner, Lovell and Schmidt (977) and Simar and Wilson (00), let = u + v and = u = v. The parameters to be estimated are = (; ; ). There are two primary estimators suggested in the literature: the maximum likelihood estimator and corrected least squares (Aigner, Lovell and Schmidt, 977; Olson, Schmidt and Waldman, 980). Under the normal-half normal speci cation, the MLE of (; ; ) is the set of parameters values maximizing the likelihood function: ln L(; ; j(y i ; x i ); i = ; :::; ) () = ln( ) ln P + ln p (y i x 0 i) P (y i x 0 i) ; where () is the standard normal cumulative distribution function. is simply the least squares slope estimate in the regression of y i on x i. The COLS estimate of However, the mean of " i = v i u i is negative due to the term u i, so the COLS estimate needs to be adjusted by adding the bias, p u=, back into the intercept estimator. The bias can be consistently estimated using the variance estimates: ^ u = r 4 =3 ^ 0 3 ; ^ v = ^ 0 ^ u; (3) where ^ 0 and ^ 0 3 are the second and third sample central moments of the least squares residuals. Both MLE and COLS are consistent. The Monte Carlo experiments in Olson, Schmidt and Waldman (980) show that there is little di erence between MLE and COLS for the slope coe cients in nite samples. For the intercept and variance parameters, however, MLE and COLS di er. In addition to MLE and COLS, Olson, Schmidt and Waldman (980) also consider a third consistent 3

estimator, the two-step ewton-raphson estimator, which has di erent nite sample properties than MLE and COLS. Waldman (98) discovers an important property of MLE: for the likelihood function () above, the point (b; 0; s ) is a stationary point, where b and s are the OLS estimates of and. Intuitively, when = 0, the term u i disappears, so the likelihood function of the SPF model () boils down to one of a linear model with u i = 0. A salient result in Waldman (98) is that when the skewness of the OLS residuals is positive, i.e., ^ 0 3 > 0, then (b; 0; s ) is a local maximum in the parameter space of the likelihood function. 4 This is the so-called "wrong skewness issue" in the literature, because 0 3 < 0 in the normal-half normal model. Olson, Schmidt and Waldman (980) refer to this phenomenon as "Type I failure" since the COLS estimator de ned in (3) does not exist when ^ 0 3 > 0. The Monte Carlo studies in Simar and Wilson (00) show that the wrong skewness issue is not rare, even when the signal-to-noise ratio is considerably large. For example, the frequency of wrong skewness could be 30% for a sample of size of 00 when = u = v =. Wrong skewness casts doubt on the speci cation of the SPF model (Greene, 007). Moreover, it invalidates the calculation of standard errors of parameter estimates (Simar and Wilson, 00). Greene (007) considers OLS residual skewness a useful diagnostic tool for the normal-half normal model. Wrong skewness suggests there is little evidence of ine ciency in the sample, implying that rms in the sample are "super e cient". Thus, and u are assumed to be zero, and the stochastic frontier model reduces to a production function without the ine ciency term. 5 Another interpretation of the wrong skewness issue is that the normal-half normal model is not the correct speci cation. Other speci cations may well reveal the presence of ine ciency and reconcile the distribution of one-sided ine ciency with the data. The binomial distribution considered by Carree (00) and doubly truncated normal distribution proposed by Almanidis and Sickles (0) and Almanidis, Qian and Sickles (04) could have either negative or positive skewness. argue that models with ambiguous skewness may be more appropriate in applied research. They Simar and Wilson (00) argue that wrong skewness is a nite sample problem, even when 4 Waldman (98, p. 78) also suggests that (b; 0; s ) may be a global maximum. There are two roots in this normal-half normal model: OLS (b; 0; s ) and one at the MLE with positive. When the residual skewness is positive, the rst is superior to the second (Greene, 007, note 8). 5 Kumbhakar, Parmeter and Tsionas (03) propose a stochastic frontier model to accommodate the presence of both e cient and ine cient rms in the sample. 4

the model is correctly speci ed. 6 They show that a bootstrap aggregating method provides useful information about ine ciency and the model parameters, regardless of whether residuals are skewed in the desired direction. We also consider wrong skewness to be a consequence of estimation in nite samples when the signal-to-noise ratio V ar(u i )=V ar(" i ) is small. 7 Since the OLS residuals of a production function regression with u i = 0 display skewness in either direction with probability of 50%, a sample drawn from an SPF model with small signal-to-noise ratio could generate positively skewed residuals with high probability. 8 3 Finite Sample Correction As illustrated by Simar and Wilson (00), wrong skewness may occur when the signal-to-noise ratio is sizable, so simply setting u = 0 when the skewness is positive could be a mistake. Instead of improved interval estimates proposed by Simar and Wilson (00), this paper proposes a nite sample adjustment to existing estimators in the presence of wrong skewness. For MLE, a constraint with non-positive residuals skewness is imposed: s:t: max ln L(; ; j(y i ; x i ); i = ; :::; ) X 4q 3 y i y x 0 i x0 5 P (y i x 0 i y x0 ) 3 0; (4) where y = P y i and x = P x i. Unfortunately, when implementing maximum likelihood estimation with the inequality constraint de ned by (4), there is a practical issue. As pointed out by Waldman (98), in the case of positive skewness of residuals, OLS (b; 0; s ) is a local maximum and the unconstrained MLE is equal to (b; 0; s ). Since, OLS is a local maximum in the parameter space of unconstrained MLE, the constraint (4) is always binding at the maximum, leading to zero skewness of the constrained MLE residuals. 9 6 Waldman (98, p.78) notes that for u > 0 "as the sample size increases the probability that P e 3 t > 0 and hence that (b; 0; s ) locates a local maximum goes to zero." 7 Badunenko, Henderson and Kumbhakar (0) nd that the estimation of e ciency scores depends on the estimated ratio of the variation in e ciency to the variation in noise. As discussed by Kim, Kim and Schmidt (007) and Feng and Horrace (0) in xed e eccts stochastic frontier models, small signal-to-noise ratio leads to inaccurate inference. 8 As pointed out by Simar and Wilson (00, p.7), this problem could happen in other one-sided speci cations. In a previous version of this paper, our Monte Carlo experiments suggest that wrong skewness could also occur with high probability in exponential and binomial SPF models, when the signal-to-noise ratio is small. 9 This stems from the fact that Waldman (98) shows that OLS is local maximum in the parameter space of MLE when the OLS residuals are positively skewed. In fact, the non-positivity contraint will bind globally (when the OLS 5

If we regard the sign of residual skewness as an important indicator of model speci cation, the constrained MLE above seems unsatisfactory. We, therefore, propose a (negative) upper bound of skewness instead of zero in (4). This is relevant for empirical modeling. As in the empirical example below, when there is evidence of technical ine ciency in the data (Greene 007, p.0), its variance can not be too small, relative to that of the composite error " i. Denote the signal-to-noise ratio by instead of = u = v. 0 k = V ar(u i )=V ar(" i ); That is, a lower bound on the signal-to-noise ratio is implicitly imposed, k k 0. To develop the relationship between the upper bound of skewness and the lower bound of the signal-to-noise ratio, consider the second and third moment of " i. Under the normal-half normal speci cation, Olson, Schmidt and Waldman (980) show that and It follows that the skewness of " i is " 3 " i E(" i ) E p = V ar("i )# where u is replaced with V ar(" i ) = v + u (5) E[" i E(" i )] 3 = 3 up =[( 4)=]: (6) u V ar(" i ) V ar(u i). Reparameterizing the skewness in terms of the signal-to- q q = k3=, with a constant = 4 ' 0:9953. noise ratio, we have g(k) = k 3= 4 3= p =[( 4)=] = V ar(ui ) V ar(" i ) 3= r 4 ; Since > 0, g(k) < 0 (e.g., g(0:) 0:035, g(0:) 0:0890 and g(0:3) 0:635) and g 0 (k) = 3 k= < 0. An important property of g(k) is that it is a monotonically decreasing function of k. This implies that any upper bound, say g 0, of the population skewness, g(k) g 0, is equivalent to a lower bound, denoted by k 0, of the signal-to-noise ratio, k k 0, i.e., g 0 = g(k 0 ) < 0. We impose this upper bound on the sample skewness, by replacing 0 in the constraint (4) with the negative upper bound of the population skewness, g(k 0 ). Consequently, a modi ed constraint 3 X 4 y i y x 0 3 i x0 q 5 P g(k 0 ) (y i x 0 i y x0 ) residuals are positively skewed), if OLS is a global maximum, as the Monte Carlo studies of Olsen, Waldman and Schmidt (980) suggest. 0 Coelli (995) also uses this signal-to-noise ratio measure, denoted by, in his Monte Carlo experiments. 6

is used in the constrained MLE in the event of wrong skewness of the OLS residuals. Based on Waldman s (98) argument, the constraint above will also be binding at a maximum in the neighborhood of OLS. The constraint becomes 3 X 4 y i y x 0 i x0 q 5 P (y i x 0 i y x0 ) 3 = g(k 0 ) (7) This nite sample adjustment gives a constrained estimator of parameter vector (; ; ). The constrained COLS slope coe cients can be similarly de ned. We use constraint (7), but replace the likelihood () with the sum of squared residuals as the objective function of a minimization problem. Since COLS reduces to OLS in the presence of wrong skewness and OLS is a local maximum of likelihood, as a nite sample adjustment to OLS, the constrained COLS slope coe cients are expected be close to their constrained MLE counterparts. 3. Linearizing the constraint The non-linearity of in the constraint (7) creates computational di culties in calculating the constrained MLE. To simplify computation, a linearized version of the constraint (7) is considered. Given that OLS is a local maximum of likelihood in the presence of wrong skewness, empiricists normally start by estimating OLS with u i = 0. This is the rst step in LIMDEP (Greene, 995) and FROTIER (Coelli, 996). If the skewness of the OLS residuals is positive, then OLS is the optimum and the point of departure for our linearization concept. Since the primary concern is skewness correction, we impose the additional restriction that the MLE residual variance P (y i x 0 i y x0 ) is equal to that of OLS residuals, ^ 0. Thus, the linearized constraint becomes: P [y i y (x i x) 0 ] 3 = g(k 0 ) (^ 0 ) 3= : Denote f() = P [y i y (x i x) 0 ] 3. The rst-order Taylor expansion of f() at the OLS estimate ^ OLS is: f() f(^ OLS ) + " # 0 @f() ( ^OLS); @ j^ols where @f() @ j^ols is the derivative of f() with respect to evaluated at ^ OLS. f(^ OLS ) is the 3 rd central moment of OLS residuals, i.e., ^ 0 3. ow, @f() @ = 3 P [y i y (x i x) 0 ] (x i x); 7

and @f() @ j^ols = 3 P e i (x i x); where e i denotes the OLS residual y i x 0 i^ OLS with a sample mean equal to zero. Hence, an approximation of the constraint (7) is ^ 0 3 3 P e i (x i x) 0 ( ^OLS ) = g(k 0 ) (^ 0 ) 3= ; (8) or [ P e i (x i x)] 0 ( ^OLS ) = ^0 3 3 g(k 0 ) 3 (^0 ) 3= : (9) Letting the vector ~e be the squared OLS residual vector (e ; :::; e )0, the constraint above can be written in matrix form as ~e0 M 0 X( ^OLS ) = ^0 3 3 g(k 0 ) 3 (^0 ) 3= ; where M 0 = I 0 and = (; :::; ) 0. Thus, the linear constraint above can be written as R = q(k 0 ) (0) with R = ~e0 M 0 X and q(k 0 ) = R^ OLS + ^0 3 3 + 3 k3= 0 (^ 0 ) 3=, depending on the value of k 0. Therefore, the proposed nite sample correction for MLE of (; ; ), i.e., the constrained MLE, is de ned as the solution to maximizing the likelihood () subject to the linear constraint (0). The corresponding estimators of u and v can be obtained by using the relationship = u + v and = u = v. Similarly, the constrained COLS of is de ned to minimize the sum of squared residuals subject to (0). As in the unconstrained estimation, the constrained estimators of u and v can be obtained by formula (3). If k 0 = 0, then g(k 0 ) = 0 and the constraint above becomes R( ^OLS ) = ^ 0 3=3. This implies that the constrained and unconstrained estimators would be similar, since ^ 0 3 is usually very small in the presence of wrong skewness. In the extreme case of ^ 0 3 = 0, the constrained estimator reduces to OLS, which is a local maximum of the likelihood. It is worth noting that (0) is not a direct linearization of (7). Alternatively, a full linearization of (7) can be similarly obtained by replacing R = ~e0 M 0X with R = (~e0 0 M 0 p^ ^ 0 3e 0 )X. The additional term 0 p^ ^ 0 3e 0 X is from the e ect of the denominator of the constraint in (7). Monte Carlo simulations suggest that the estimation results are robust to this choice. Details are available upon request. 8

Using the linearized constraint (0), the estimates, standard errors and con dence intervals of the constrained MLE and constrained COLS can be easily obtained using Stata or other existing software. However, since (0) does not guarantee a negative residual skewness in nite samples, there is a possibility that wrong skewness could still occur after our correction. The Monte Carlo experiments below show that this may only be a concern when the underlying signal-to-noise ratio is very small. 3. Choosing the value of k 0 The idea of the proposed constrained estimators is to adjust the slope coe cients to obtain a correct sign of residual skewness using the constraint (0), which is a function of k 0. It is expected that when the chosen value of k 0 is small, a slight adjustment results in the constrained MLE (or constrained COLS), and its value will be close to the unconstrained MLE. Choosing a speci c value of k 0 is an empirical issue. On the one hand, when there is a priori evidence of ine ciency, the signal-to-noise ratio cannot be too small. On the other hand, as illustrated by the Monte Carlo study in Simar and Wilson (00), wrong skewness is less likely to occur as the signal-to-noise ratio increases. 3 In the spirit of this trade-o we develop model selection criteria to choose k 0. The idea is to incorporate a penalty function, so that as k 0 increases the penalty decreases. Hence, the t of the model and e ect of the constraint on the optimum can be balanced. For the constrained MLE we propose a Bayesian information criterion (BIC) via the likelihood to choose the value of k 0 : BIC(k 0 ) = l r (k 0 ) k 0 ln ; where l r (k 0 ) is the log-likelihood evaluated at the constrained MLE of (; ; ), depending on k 0. Since OLS (b; 0; s ) is a local maximum of the log-likelihood function in the presence of positive skewness with a restriction on k 0, the value of l r (k 0 ) decreases with k 0 in the neighborhood of (b; 0; s ). 4 Di erent from the usual BIC, here we use a negative sign in front of the penalty term In the empirical example below, the command Frontier in Stata, which allows for a linear constraint, is employed. 3 Table in Simar and Wilson (00) provides some guidance. On the one hand, when 0: (i.e., k = =( + ) < 0:035) for samples with size less than 00, the proportion of wrong skewness is close to 50%, implying that the ine ciency term is hard to distinguish from noise. On the other hand, when (k 0:67), the wrong skewness probability decreases dramatically. For example, only 6% of samples display wrong residual skewness for = (k = 0:4) and = 00. We have a similar nding both for Simar and Wilson s design and the design in Section 5 of this paper. Results are available upon request. 4 The constraint k k 0 is always binding in the neighborhood of OLS. And a restriction on k is equivalent on, 9

k 0 ln so that l r (k 0 ) and k 0 ln move in opposite directions with k 0. An optimal value of k 0 is chosen to minimize BIC(k 0 ): ~k 0 = arg min k 0 [0;) BIC(k 0): Similarly, for the constrained COLS, a criterion based on sum of squared residuals is proposed to select the value of k 0 : C(k 0 ) = SSR r(k 0 ) k 0^ ln " ; where SSR r (k 0 ) is the sum of squared residuals of OLS with the constraint (0). C(k 0 ) is a Mallows C p -type criterion, similar to the expression proposed by Bai and g (00) to choose the number of factors in the approximate factor models, except that the penalty term takes a negative sign. By applying the properties of the usual restricted least squares, it can be shown that SSR r (k 0 ) increases with k 0. (See the appendix.) Hence, the e ect of increasing k 0 on the model t can be balanced by the penalty term, thus an appropriate value of k 0 is chosen to minimize C(k 0 ): ^k 0 = arg min C(k 0): k 0 [0;) The estimated error variance ^ " provides an appropriate scaling to the penalty term. Here, we use ^ " = SSR, where SSR is the sum of squared residuals of OLS without constraint. In practice, to nd the value of ~ k 0 (or ^k 0 ) a grid search can be applied to BIC(k 0 ) (or C(k 0 )) starting from a small positive value, e.g., 0:05. Since the measures of the model t in the constrained MLE and COLS, i.e., the objective functions in the penalized least squares and penalized maximum likelihood, are di erent, ~ k 0 is not necessarily equal to ^k 0. However, in the neighborhood of OLS (b; 0; s ) with a small value of, P when the term h i ln p (y i x 0 i ) in l(; ; ) has small values of partial derivatives in the rst-order conditions, ~ k 0 should be close to ^k 0. It is worthwhile to note that k 0 is not a model parameter here, and is selected by the proposed selection criteria only for nite sample correction. Thus, choosing k 0 is inherently di erent from model selection in the literature, such as, choosing the number of model parameters, where consistency is a primary requirement for the penalty term. Therefore, we could use di erent penalty which is a monotonic increasing function of k in the half-normal model, s s s s = u V ar(u i) k = = = v ( k) (=k ) : v 0

terms in BIC(k 0 ) or C(k 0 ) above as long as a unique value of k 0 can be chosen. The Monte Carlo experiments and empirical example below suggest that the proposed selection criteria work well. 4 Constrained Estimators With the proposed nite sample adjustment, the sample can still be used to construct a point estimate for inferring population parameters in the presence of wrong skewness. This is similar in spirit to Simar and Wilson (00), who still rely on the MLE estimation results, but provide more accurate interval estimates using improved inference (bagging) methods. As previously mentioned, any negative constraint on sample skewness is binding in the presence of wrong skewness. This result implies that estimated (or k) is implicitly determined by the constraint (0). Consequently, it is biased when the selected value of k 0, the lower bound of k, is not equal to the true value of k. Inconsistency of the proposed constrained estimators might be a concern. However, this concern may be overstated. The wrong skewness is a nite sample issue under the true speci cation. As the sample size increases, wrong skewness is less likely to appear, so the proposed nite sample adjustment becomes unnecessary. Thus, asymptotics are less of a concern here. In addition, with the nature of nite sample adjustment, the proposed method is regarded as an adjustment to existing estimators, rather than a new estimator. 5 In the next subsection, properties of constrained estimators are studied. Since the constrained COLS is essentially restricted least squares, which has an analytical solution, we mainly focus on it. 4. Constrained COLS The proposed constrained COLS, denoted by ^ r, is a -step estimator. In the rst step, for a given k 0, the constrained COLS ^ r (k 0 ) is de ned as the solution of min SSR() = min (Y X) 0 (Y X) s:t: R = q(k 0 ): In the second step, k 0 is selected such that ^k 0 = arg min k0 C(k 0 ), where C(k 0 ) = (Y X ^ r (k 0 )) 0 (Y X ^ r (k 0 )) k 0^ " ln. The proposed constrained COLS is de ned as ^ r = ^ r (^k 0 ). 5 In this sense, our approach is di erent from the literature on models with moment conditions, e.g., Moon and Schorfheide (009).

This -step estimator is equivalent to a -step penalized least squares with the linear constraint: min ;k 0 (Y X)0 (Y X) k 0^ ln " s:t: R = q(k 0 ): This equivalence comes from the fact that in the objective function k 0 only appears in the penalty term k 0^ " ln. Thus, can be concentrated out for a given k 0. For a given k 0, ^ r (k 0 ) is the restricted least square. By Amemiya (985) or Greene (0), ^ r (k 0 ) = ^ OLS (X 0 X) R 0 [R(X 0 X) R 0 ] [R^ OLS q(k 0 )]; and SSR r (k 0 ) = SSR + [R^ OLS q(k 0 )] 0 [R(X 0 X) R 0 ] [R^ OLS q(k 0 )]: Thus, the criterion is C(k 0 ) = SSR + [R^ OLS q(k 0 )] 0 [R(X 0 X) R 0 ] [R^ OLS q(k 0 )] k 0^ ln " : Minimizing C(k 0 ) de nes ^k 0. The follow proposition proves the existence and uniqueness of ^k 0. Proposition In the presence of positive skewness of OLS residuals, i.e., ^ 0 3 > 0, (i) dssrr(k 0) dk 0 > 0; (ii) for a reasonable sample size, there exists a solution for ^k 0 such that ^k 0 minimizes C(k 0 ); (iii) d C(k 0 ) dk 0 > 0, implying that ^k 0 is the unique solution. The proof is in Appendix. Since ln! 0, when!, compared with the rst term SSR r(k 0 ), which converges to a non-zero constant, the penalty term k 0^ " ln in C(k 0) can be ignored asymptotically. This implies that ^k 0! 0 as!. Hence, when is large the proposed constrained COLS approaches the OLS with constraint R( skewness. ^OLS ) = ^ 0 3=3, which is very close to OLS in the presence of wrong This property also implies that in a sample with a large number of rms, the selected ^k 0 could be 0. In this case, to avoid wrong skewness, a small positive value, say, 0:05, is suggested. For a given sample, the di erence between OLS and the constrained COLS ^ OLS ^r = (X 0 X) R 0 [R(X 0 X) R 0 ] [R^ OLS q(^k 0 )]

depends on ^k 0, and d[^ OLS ^r ] d^k 0 = (X 0 X) R 0 [R(X 0 X) R 0 ] ^k = 0 (^ 0 ) 3= implying that the magnitude of this di erence is positively correlated with the chosen value ^k 0. 4. Constrained MLE For a given k 0, the constrained MLE (^ CMLE (k 0 ); ^ CMLE (k 0 ); ^ CMLE (k 0)) depends on k 0. Minimizing BIC(k 0 ) determines the value of k 0, i.e., ~ k 0 = arg min k0 [0;) BIC(k 0 ). Similar to the constrained COLS, (^ CMLE ; ^ CMLE ; ^ CMLE ) is de ned as (^ CMLE ( ~ k 0 ); ^ CMLE ( ~ k 0 ); ^ CMLE (~ k 0 )). It can also be written as a penalized maximum likelihood estimator with a constraint, where l(; ; ) = ln( ) in (). min l(; ; ) k 0 ln ;; ;k 0 ln P + s:t:r = q(k 0 ); h ln p (y i x 0 i ) i P (y i x 0 i ) de ned Since there is no analytical solution to the constrained optimization problem above, it is di cult to derive the properties of constrained MLE. However, dividing by, BIC(k 0) = l r(k 0 ) k 0 ln, compared with l r(k 0 ), which does not converge to zero, the penalty term k 0 ln can be asymptotically ignored as!, implying that ~ k 0 tends to 0 as!. Since ~ k 0 is small when is large, the proposed constrained MLE is expected be close to MLE. Since the MLE of slope parameters is very close to OLS, the constrained MLE and constrained COLS are expected to be close. Similar to the constrained COLS, the selected ~ k 0 could be 0 in a sample with a large. In this case, we also impose a lower bound of, say, 0:05, to avoid wrong skewness. We now consider the di erence between constrained MLE and OLS by examining the rst-order conditions of (). Aigner, Lovell and Schmidt (977) show that: @ ln L @ = @ ln L @ @ ln L @ + P 4 (y i x 0 i) + P () 3 () (y i x 0 i) = 0; () P () = () (y i x 0 i) = 0; () = P i x (y 0 i)x i + P () () x i = 0; (3) 3

where () is the standard normal density function. () and () are evaluated at (y i x 0 i ) = " i. Waldman (98) shows that in the presence of wrong skewness = 0 and OLS is a local maximum of the log-likelihood. For our constrained MLE, the constraint (7) or (9) involves the value of k 0, not directly. Since is a monotonic increasing function of k, k k 0 implies s (=k 0 ) : (4) To show how restricting a ects the estimation result and how the constrained MLE of is di erent from the OLS, consider equation (3). 6 Taking the rst-order Taylor expansion at = 0 gives Thus, (3) becomes ( " r i) ( " i) + " i: That is, P 0 = (y i x 0 P i)x i + = ( + P ) (y i x 0 i)x i + ( " r i) ( " i) x P i (y i x 0 P i)x i + ( + r P x i : " i)x i q P (y i x 0 i)x i + p P ( + x i = 0: (5) ) In matrix form, the equation (5) above can be written as where '() = X 0 y X 0 X + '() p X 0 = 0 (6) q =( + ) and is the vector of ones. Equivalently, ^ CMLE ' (X 0 X) X 0 y + '() p (X 0 X) X 0 : (7) In the presence of wrong skewness, OLS (i.e., = ' = 0) is a local maximum of the log-likelihood. Under the constraint (4), the estimator of is adjusted by the second term in equation (7). Given the fact that '() is monotonically increasing in in the range [0, p = :533], the di erence between the constrained MLE and the OLS of is positively related to the value 6 Strictly speaking, restricting as a constraint yields a di erent result from constraint (7). Though the population skewness is equal to g(k 0) and thus a monotonic function of, the sample skewness is not a function of. However, the insights derived here on the e ect of the chosen value of k 0 on estimation still apply. 4

of. 7 The larger (or k 0 ) is imposed, the bigger is the di erence between the OLS and the constrained MLE. Furthermore, in a given sample this di erence depends not only on '(), but also on the sample value of the regressors and jointly determined by rst-order equations. We conjecture that constraint (0) with a small value of k 0, slightly adjusts the estimators of and v, but has a much larger e ect on the estimated u and. This point is con rmed in the Monte Carlo experiments and empirical example below. 5 Monte Carlo Experiments In this section, Monte Carlo experiments are conducted to study how the proposed constraints a ect the estimates, and how the chosen value of k 0, the imposed lower bound of k, is a ected by the sample size. Di erent from the constrained MLE, the constrained COLS has an analytical solution and established results are in the previous section. Thus, it is computationally convenient to use the constrained COLS in Monte Carlo experiments, and the main results can be carried over to the constrained MLE. We consider a speci cation y i = 0 + x i + x i + " i ; " i = u i + v i ; i = ; ; ; where 0 = ; = 0:8; = 0:, x i log(j(4; 00)j), x i log(j(; 60)j), v i (0; v) and u i j(0; u)j. k = V ar(u i )=V ar(" i ) is the signal-to-noise ratio. 8 u = V ar(u i) = kv ar(" i) and v = ( k)v ar(" i ). We set V ar(" i ) = v + V ar(u i ) = 0:06, so the variance of x i and V ar(" i ) are comparable to those in the empirical example below. Since the focus is the proposed correction for samples with wrong residual skewness, we drop the samples with correct skewness. The number of replications is 4000 after dropping the samples with correct skewness. We conduct experiments with k = 0:, 0:, 0:3, 0:5; 0:7 and = 50, 00, 00. 9 Table reports the simulation results. Column () gives the average value of ^k 0. To obtain ^k 0 for each sample, a grid search is conducted to minimize C(k 0 ) on the interval [0:05; 0:9]. As expected, the average value of ^k 0 decreases with. Column (3) shows that there is still a possibility of wrong 7 For a small value of k 0, e.g., k 0 [0:; 0:3], lies in the interval [0:5530; :0860]. 8 Coelli (995) also uses this signal-to-noise ratio measure, denoted by, in his Monte Carlo experiments. 9 Our Monte carlo experiments show that it is very unlikely to have wrong skewness when k 0:7 and 00. Results are available upon request. 5

skewness after the proposed nite sample correction. The frequency depends on the signal-to-noise ratio and sample size, varying from 6:3% to 39:9%. For example, for k = 0:5, = 00, our nite sample correction approach could fail with a possibility of 8:4%. This failure is a cost of the linearization approximation (8). When k 0 is small, g(k 0 )(^ 0 ) 3= could be a small negative value close to zero. Consequently, due to approximation error, a linearized constraint does not guarantee a negative third moment of residuals or skewness. However, as k increases, the failure frequency can be greatly reduced, e.g., to 6:3% for k = 0:7, = 00. p For parameter estimators, columns (4)-(7) indicate that with the correction of ^ u =, constrained COLS of 0 is less biased than OLS, but with a much larger root mean squared errors (RMSE). But when k and increase, the RMSE of constrained COLS is comparable to that of OLS. (Bias and RMSE of OLS of 0 (and ) are included in columns (5), (7) (and (9), ()) for comparison). In addition, compared with OLS, the constrained COLS of is slightly upward biased with bigger RMSE, and the bias and RMSE decrease with k and. In the presence of wrong skewness, u is typically assumed to be zero. Using our correction, column () shows that the estimated u tends to be overestimated for a small value of k and underestimated for a big value of k. Compared with u, v can be estimated more accurately in terms of bias, as indicated in column (3). Column (4) of Table shows that k is generally underestimated. This is due to the fact that a relatively small value of ^k 0 is often chosen when is large, and that the estimated k is implicitly determined by ^k 0 suggested by the linear constraint (0). 0 Finally, column (5) reports the bias of the mean technical e ciency E[exp( u i )] = exp( u=)[ ( u )]. In the presence wrong skewness, traditional practice suggests that the estimated u is 0, implying that the estimated mean technical e ciency is. This practice obviously overestimates the true mean technical e ciency. Column (5) shows that the mean technical e ciency estimator using the proposed correction could be unbiased for a sizable value of k, say, 0: here under the current design. It is downward biased for a small value of k, and upward biased for k > 0:. 0 But this is not a big concern since k is not a parameter of interest in this model. 6

6 Empirical Example: the US Airline Industry In this section, an airlines example is used to illustrate our approach. This is an unbalanced panel data set with 56 observations. See Greene (007) for detailed information of this data set. In this example, the dependent variable is the logarithm of output and the independent variables include the logarithm of fuel, materials, equipment, labor and property. Here, the unbalanced panel is treated as a cross section for 56 rms to ensure that the wrong skewness issue arises. Column () of Table presents the OLS estimates along with standard errors (column 3). Except for the constant term, the slope coe cients are consistent with Table. in Greene (007). The OLS residual skewness (0:067) is in the wrong direction for the estimated normal-half normal model. Thus, the estimates of and u are set to zero and rms are considered to be "super e cient". However, Greene (007, footnote 84) does suggest that there is evidence of technical ine ciency in the data. The second root of the likelihood with positive is reported in the second section of Table. This MLE yields a small positive residual skewness 0:0093. Usually, in the presence of "wrong" skewness, researchers are advised to obtained a new sample or respecify the model. Instead, we use the constrained MLE (and constrained COLS), a nite sample adjustment to the existing MLE (and COLS). The optimal value of k 0 can be chosen by BIC(k 0 ) (and C(k 0 ) for the constrained COLS) proposed above. For purposes of illustration, we present constrained MLE results of k 0 = 0:05, 0:, 0:5, and 0: in columns (6)-(3) of Table and compare the values of BIC(k 0 ), showing that ~ k 0 = 0:5 achieves the minimum of BIC(k 0 ). Thus, the constrained MLE of and u are positive, 0:689 and 0:05 respectively. Furthermore, consistent with the negative population skewness of the composed error, the skewness of constrained MLE residuals ( 0:0599) has the desired sign. Since the constraint slightly adjusts the coe cients of constrained MLE, as expected, the rest of the coe cients are very close to the unconstrained MLE and OLS. For example, the constrained estimated coe cient of variable Log fuel is 0:3907 (column 0), while its unconstrained counterpart With the exception of perhaps Green and Mayes (99), Mester (997) and Parmeter and Racine (0), there appear to be very few empirical studies with wrong skewness in the literature. As in Greene (007, Table.), we use this panel data example as a cross-sectional one only for the purpose of illustration. Inconsistent with the statements of Waldman (98) and Greene (007) the MLE with positive achieves a slightly bigger value of log-likelihood than OLS for this dataset. Similarly, the inconsistency between OLS and MLE in the presence of positive OLS residual skewness by using FROTIER is discussed by Simar and Wilson (009). Greene (007, p.0) notes: "... for this data set, and more generally, when the OLS residuals are positively skewed, then there is a second maximizer of the log-likelihood, OLS, that may be superior to the stochastic frontier." 7

is 0:3836 (in column 4) and OLS coe cient is 0:388 (in column ). Consistent with the analysis in Section 4., the di erence between the constrained MLE slope coe cients and its OLS (and unconstrained MLE) counterparts is positively related to the magnitude of k 0. The bigger the value of k 0, the larger is the di erence. However, this di erence is relatively small. For example, the constrained estimated coe cients of variable Log fuel using k 0 = 0: is 0:3939 (in column of Table ), compared with the OLS 0:388 and the unconstrained MLE 0:3836 (in columns and 4 of Table ). This is also the case for v and. In stark contrast to this small di erence in slope coe cients, the residual skewness and estimated k change signi - cantly, since they are implicitly determined by the chosen value of k 0 in the constraint. Another important point observed in Table is that the value of the likelihood decreases with k 0. 3 The results of constrained COLS are reported in columns (6)-(3) of Table 3 and are very close to their constrained MLE counterparts for given values of k 0 = 0:05, 0:, 0:5, and 0:. 4 However, for the constrained COLS, the optimal value of k 0 is 0: by applying Mallows C p -type criterion C(k 0 ) proposed above. (Table 3 reports C(k 0 ) instead of C(k 0 ).) This is slightly di erent from ~k 0 = 0:5 by minimizing BIC(k 0 ) in the constrained MLE. Therefore, the constrained COLS of u is 0:0853 and skewness is 0:035 in column (8). It is worth mentioning that the value of criterion C(0:5) is nearly equal to C(0:) in this empirical example, implying that BIC(k 0 ) for the constrained MLE and C(k 0 ) for the constrained COLS result in similar optimal values of k 0. Since the proposed nite sample adjustment restricts the signal-to-noise ratio, it indirectly a ects the estimated u. In this example, it is 0:05, for the constrained MLE. Consequently, the mean technical e ciency estimate, exp(^ u=)[ (^ u )], depends on the chosen value of k 0. However, e ciency rankings appear to be preserved under di erent choices of k 0. For the unconstrained MLE, the least e cient rm is the 79 th with technical e ciency.8958. If we impose k 0 = 0:05, 0:, 0:5, 0: in the constraint, the technical e ciency becomes.8583,.8308,.805,.77 3 This property can be obtained by the equation (3) in Waldman (98, p.78): r l = 3 4 6s 3 P e 3 i where can be regarded as changing from 0 as in the analysis in Section 3.. Since 4 < 0, in the presence P wrong skewness ( e 3 i > 0), the log-likelihood decreases with the imposed value of (and k 0). 4 The constant term is calculated by OLS intercept plus p ^ u=. The standard errors formula of the COLS estimators of constant term, and (not ) can be found in Coelli (995). 8

respectively, and it remains lowest among the 56 rms. The most e cient rm is the 50 th with technical e ciency.9696, 0.9669,.9655,.9644,.9636 for the unconstrained MLE and constrained MLE with k 0 = 0:05, 0:, 0:5, 0:, respectively. This is also the case for the median rm. 7 Conclusions This paper studies the wrong skewness issue in parametric stochastic frontier models. Following Simar and Wilson s (00) opinion, we consider wrong skewness to be a consequence of estimation in nite samples when the signal-to-noise ratio is small. In nite samples the data may fail to be informative enough to detect the existence of ine ciency term in stochastic frontier models, even though the population signal-to-noise ratio could be fairly large. Thus, the resulting residuals could display skewness in either direction with probability of as high as 50%. As an alternative to the usual "solutions" to the wrong skew problem, we propose a feasible nite sample adjustment to existing estimates. When there is evidence of ine ciency, it is reasonable to impose a lower bound on the signal-to-noise ratio in the normal-half normal model, equivalent to a negative upper bound on the residual skewness. Thus, we propose to use this negative bound on residual skewness as a constraint in the MLE and COLS in the event of wrong skewness. The idea of the proposed constrained estimators is to slightly adjust the slope coe cients in nite samples. They provide a point estimate that yields a negative residual skewness, though a correct sign of residual skewness is not always guaranteed. Since the constraint is based on k 0, the choice of k 0 a ects estimation results. A model selection approach is proposed to select k 0. Monte Carlo experiments show that the bias of constrained estimates is less of a concern when sample size is large and signal-to-noise ratio increases. The empirical example in this paper also shows that the value k 0 has little e ect on the estimated slope coe cients and v,, while the residual skewness and estimated k are implicitly determined by the value of k 0. In this sense, the proposed method can be regarded as a nite sample adjustment to existing estimators, rather than a new estimator. When the sample size is large, since wrong skewness is less likely to occur, such adjustment becomes unnecessary. 9

References [] Aigner, D.J., C.A.K. Lovell, and P. Schmidt, 977, Formulation and Estimation of Stochastic Frontier Production Function Models, Journal of Econometrics 6, -37. [] Amemiya, T., 985. Advanced Econometrics, Cambridge, MA: Harvard University Press. [3] Almanidis, P. and R. C. Sickles, 0, The Skewness Issue in Stochastic Frontier Models: Fact of Fiction? In I. van Keilegom and P. W. Wilson (Eds.), Exploring Research Frontiers in Contemporary Statistics and Econometrics. Springer Verlag, Berlin Heidelberg. [4] Almanidis, P., J. Qian and R. Sickles, 04, Stochastic Frontier with Bounded E ciency, In Festschrift in Honor of Peter Schmidt: Econometric Methods and Applications. RC Sickles and WC Horrace (eds.) Springer Science & Business Media, ew York, Y, (04): 47-8.. [5] Badunenko, O., D. Henderson and S. Kumbhakar, 0, When, Where and How to Perform E ciency Estimation, Journal of the Royal Statistical Society, Series A, 75, 863-89. [6] Bai, J. and S. g, 00. Determining the umber of Factors in Approximate Factor Models, Econometrica, 70, 9-. [7] Carree, M., 00, Technological ine ciency and the skewness of the error component in stochastic frontier analysis, Economics Letters 77, 0-07. [8] Coelli, T., 995, Estimators and Hypothesis Tests for a Stochastic Frontier Function: A Monte Carlo Analysis, Journal of Productivity Analysis, 6, 47-68. [9] Coelli, T., 996, A guide to frontier version 4.: A computer program for stochastic frontier production and cost function estimation, CEPA working paper o. 96/07, Centre for E ciency and Productivity Analysis, University of ew England, Arimidale, SW 35, Australia. [0] Feng, Q. and W. C. Horrace, 0, Alternative Technical E ciency Measures: Skew, Bias and Scale, Journal of Applied Econometrics 7, 53-68. [] Green, A. and Mayes, D., 99, Technical Ine ciency in Manufacturing Industries, Economic Journal 0, 53-538. [] Greene, W., 980, On the Estimation of a Flexible Frontier Production Model, Journal of Econometrics, 3, 0-5. [3] Greene, W., 995, LIMDEP Version 7.0 User s Manual, ew York: Econometric Software, Inc. 0

[4] Greene, W., 007, The Econometric Approach to E ciency Analysis, in H.O. Fried, C. A. K. Lovell and S. S. Schmidt, eds., The Measurement of Productive E ciency: Techniques and Applications. ew York: Oxford University Press. [5] Greene, W., 0, Econometric Analysis, 7th edition, Pearson. [6] Hafner, C., H. Manner and L. Simar, 03, The Wrong Skewness Problem in Stochastic Frontier Models: A ew Approach, working paper. [7] Kim, M., Kim, Y. and P. Schmidt, 007, On the accuracy of bootstrap con dence intervals for e ciency Levels in Stochastic Frontier Models with Panel Data, Journal of Productivity Analysis, 8, 65-8. [8] Kumbhakar, S. and K. Lovell, 000, Stochastic Frontier Analysis, Cambridge University Press, Cambridge, UK. [9] Q. Li, 996, Estimating a Stochastic Production Frontier When the Adjusted Error is Symmetric, Economics Letters, 5, -8. [0] Kumbhakar, S., C. Parmeter and E. Tsionas, 03, A Zero Ine cient Stochastic Frontier Model, Journal of Econometrics, 7, 66-76. [] Meeusen,W., and J. van den Broeck, 977, E ciency Estimation from Cobb-Douglas Production Functions with Composed Error, International Economic Review, 8, 435 444. [] Mester, L.J., 997, Measuring E ciency at US Banks: Accounting for Heterogeneity is Important, European Journal of Operational Research, 98, 30 4. [3] Moon, H. R., and F. Schorfheide, 009, Estimation with Overidentifying Inequality Moment Conditions, Journal of Econometrics, 53, 36-54. [4] Olson, J., Schmidt, P. and D. M. Waldman, 980, A Monte Carlo Study of Estimators of Stochastic Frontier Production Functions, Journal of Econometrics,3, 67-8. [5] Parmeter, C. F. and J. S. Racine, 0, Smooth Constrained Frontier Analysis, Recent Advances and Future Directions in Causality, Prediction, and Speci cation Analysis: Essays in Honor of Halbert L. White Jr., edited by X. Chen and.e. Swanson, Chapter 8, 463-489, Springer-Verlag: ew York. [6] Simar, L. and P. W. Wilson (00), Inferences from Cross-Sectional Stochastic Frontier Models, Econometric Reviews, 9, 6-98.