Working Papers:

Nonparametric Estimation of Triangular Simultaneous Equations Models under Weak Identification” (Latest Version: October 6, 2015. Under Review)

        - Matlab Codes Download: ZIP file (Latest Version: March 2014.)

Abstract:  This paper analyzes the problem of weak instruments on identification, estimation, and inference in a simple nonparametric model of a triangular system. The paper derives a necessary and sufficient rank condition for identification, based on which weak identification is established. Then nonparametric weak instruments are defined as a sequence of reduced form functions where the associated rank shrinks to zero. The problem of weak instruments is characterized to be similar to the ill-posed inverse problem, which motivates the introduction of a regularization scheme. The paper proposes a penalized series estimation method to alleviate the effects of weak instruments. The rate of convergence of the resulting estimator is given, and it is shown that weak instruments slow down the rate and penalization derives a faster rate. Consistency and asymptotic normality results are also derived. Monte Carlo results are presented, and an empirical example is given, where the effect of class size on test scores is estimated nonparametrically.

Identification in a Generalization of Bivariate Probit Models with Dummy Endogenous Regressors with Edward Vytlacil (Latest Version: June 9, 2015. Resubmitted, The Journal of Econometrics)

Abstract:  This paper provides identification results for a class of models specified by a triangular system of two equations with binary endogenous variables. The joint distribution of the latent error terms is specified through a parametric copula structure that satisfies a particular dependence ordering, while the marginal distributions are allowed to be arbitrary but known. This class of models is broad and includes bivariate probit models as a special case. The paper demonstrates that having an exclusion restriction is necessary and sufficient for globally identification in a model without common exogenous covariates, where the excluded variable is allowed to be binary. Having an exclusion restriction is sufficient in models with common exogenous covariates that are present in both equations. The paper then extends the identification analyses to a model where the marginal distributions of the error terms are unknown. 

Estimation and Inference with a (Nearly) Singular Jacobian”* with Adam McCloskey (Latest Version: November 18, 2015.)

Abstract:  This paper develops extremum estimation and inference results for nonlinear models with very general forms of potential identification failure when the source of this identification failure is known. We examine models that may have a general deficient rank Jacobian in certain parts of the parameter space, leading to an identified set that is a sub-manifold of the parameter space. We examine standard extremum estimators and Wald statistics under a comprehensive class of parameter sequences characterizing the strength of identification of the model parameters, ranging from non-identification to strong identification. Allowing for a general singular Jacobian as the limiting point of weak identification allows us to study estimation and inference in many models to which previous results in the weak identification literature do not apply. Using the asymptotic results, we propose two hypothesis testing methods that make use of a standard Wald statistic and data-dependent critical values, leading to tests with correct asymptotic size regardless of identification strength and good power properties. Importantly, this allows one to directly conduct uniform inference on low-dimensional functions of the model parameters, including one-dimensional subvectors. The paper
focuses on three examples of models to illustrate the results: sample selection models, models of potential outcomes with endogenous treatment and threshold crossing models.

* This paper is motivated by my earlier working paper titled as 
“Identification and Inference in a Bivariate Probit Model With Weak Instruments” (2009) (Slides for the latter paper is available upon request)

CQIV: Stata Module to Perform Censored Quantile Instrumental Variable Regression with Victor Chernozhukov, Ivan Fernandez-Val, and Amanda Kowalski (Latest Version: June 2012.)
        - Stata Code Download: CQIV Stata ado fileCQIV Stata help file

Cybersecurity Policy Designs and Evaluations: A Field Experiment and Economic Theory” with Yun-Sik Choi, Jin Hyuk Choi, Shu He, Gene Moo Lee, and Andy Whinston (Latest Version: November 5, 2015. Under Review)

Abstract:  Cyber-insecurity has been a serious threat to the world. A suboptimal cybersecurity environment is partly due to organizations' underinvestment and the lack of suitable policies. The motivation of this paper stems from related policy questions: how to develop a socially desirable cybersecurity environment; and how to design policies for governments and other organizations that can ensure a sufficient level of cybersecurity. This paper addresses these questions by exploring two mutually related themes. The first theme considers information asymmetry and peer effects; the second theme studies attacker-defender interaction and considers cyber-insurance policies. In relation to these themes, the paper designs and evaluates several cybersecurity policies via both empirical and theoretical analyses.
    In the first part, as a policy devise to alleviate information asymmetry and to achieve transparency in cybersecurity information sharing practice, we introduce a cybersecurity evaluation agency along with regulations on information disclosure. To empirically evaluate the effectiveness of such institution, we conduct a large-scale randomized field experiment on 7,919 U.S. organizations. Specifically, we generate organizations' security reports based on outbound spam and industry peer rankings, then share the reports with the subjects in either private or public ways. We find evidence that the security information sharing combined with a publicity treatment has significant effects on spam reduction for large spammers. Moreover, significant peer effects are observed among industry peers after the experiment.
    As the second part of the paper, we turn to theoretical analyses and introduce economic models to conduct more comprehensive policy analyses on cybersecurity. The first model is a dynamic model that incorporates strategic interaction between defending organizations and attackers and that reveals a mechanism under which the players' actions affect security outcomes. The second model is a cyber-insurance-reinsurance framework that suggests the importance of cyber-insurance and the role of governments as ultimate risk takers, in promoting cyber-insurance businesses. By computing a simple version of this model, we find that the existence of a cyber-insurance market does encourage organizations to make cybersecurity investments, provided that the organizations underestimate losses incurred by cyber-insecurity. As applications of this model, we consider cyber-insurance for cloud computing and software validation and verification (V&V). Lastly, we propose creating a security reputation measure. 

Measuring Trends with Multiple Unknown Truncation Points: The Standard of Living During Industrialization in Britain” (Latest Version: November 18, 2015. Under Review)

Abstract:  Historical data are typically deficient and require suitable econometric methods to draw inference. This paper revisits two procedures commonly practiced in the literature of economic history to estimate trends of certain historical variables when the data are truncated with unknown and possibly multiple truncation points: the reduced-sample procedure by Komlos and Kim (1990) and the adjustment procedure. In this paper, we show that the reduced-sample procedure is justified under weaker assumptions, which implies that results obtained from the procedure is more robust than it is originally thought to be. Motivated from this finding, we provide a generalized version of the method. Next, we show that the adjustment procedure may over- or under-estimate the trend, especially when the variable of interest follows distributions that are commonly assumed in the literature. Using micro-data on the heights of British soldiers and poverty-stricken boys, we illustrate our analyses in the context of the long-term trend in the standard of living during the Industrial Revolution in Britain. 

Work in Progress:

“Multiple Treatments with Strategic Interaction: The Effects of Lobbying on Policy Changes” with Karam Kang

“Estimation and Inference in a Generalization of Bivariate Probit Models with Endogenous Regressors” with Sungwon Lee


Invalidity of the Bootstrap and the m out of n Bootstrap for Confidence Interval Endpoints Defined by Moment Inequalities,” with Donald Andrews, Econometrics Journal (2009), Volume 12, pp. S172–S199.

Abstract:  This paper analyses the finite-sample and asymptotic properties of several bootstrap and m out of n bootstrap methods for constructing confidence interval (CI) endpoints in models defined by moment inequalities. In particular, we consider using these methods directly to construct CI endpoints. By considering two very simple models, the paper shows that neither the bootstrap nor the m out of n bootstrap is valid in finite samples or in a uniform asymptotic sense in general when applied directly to construct CI endpoints.
    In contrast, other results in the literature show that other ways of applying the bootstrap, m out of n bootstrap, and subsampling do lead to uniformly asymptotically valid confidence sets in moment inequality models. Thus, the uniform asymptotic validity of resampling methods in moment inequality models depends on the way in which the resampling methods are employed.