guide:55e14f6e47: Difference between revisions

From Stochiki
No edit summary
 
mNo edit summary
 
(2 intermediate revisions by 2 users not shown)
Line 1: Line 1:
<div class="d-none"><math>
\newcommand{\edis}{\stackrel{d}{=}}
\newcommand{\fd}{\stackrel{f.d.}{\rightarrow}}
\newcommand{\dom}{\operatorname{dom}}
\newcommand{\eig}{\operatorname{eig}}
\newcommand{\epi}{\operatorname{epi}}
\newcommand{\lev}{\operatorname{lev}}
\newcommand{\card}{\operatorname{card}}
\newcommand{\comment}{\textcolor{Green}}   
\newcommand{\B}{\mathbb{B}}   
\newcommand{\C}{\mathbb{C}}
\newcommand{\G}{\mathbb{G}}
\newcommand{\M}{\mathbb{M}}
\newcommand{\N}{\mathbb{N}}
\newcommand{\Q}{\mathbb{Q}}
\newcommand{\T}{\mathbb{T}}
\newcommand{\R}{\mathbb{R}}
\newcommand{\E}{\mathbb{E}}
\newcommand{\W}{\mathbb{W}}   
\newcommand{\bU}{\mathfrak{U}}   
\newcommand{\bu}{\mathfrak{u}}
\newcommand{\bI}{\mathfrak{I}} 
\newcommand{\cA}{\mathcal{A}} 
\newcommand{\cB}{\mathcal{B}} 
\newcommand{\cC}{\mathcal{C}}
\newcommand{\cD}{\mathcal{D}}
\newcommand{\cE}{\mathcal{E}}
\newcommand{\cF}{\mathcal{F}}
\newcommand{\cG}{\mathcal{G}}
\newcommand{\cg}{\mathcal{g}}
\newcommand{\cH}{\mathcal{H}}
\newcommand{\cI}{\mathcal{I}}
\newcommand{\cJ}{\mathcal{J}}
\newcommand{\cK}{\mathcal{K}} 
\newcommand{\cL}{\mathcal{L}} 
\newcommand{\cM}{\mathcal{M}}
\newcommand{\cN}{\mathcal{N}}
\newcommand{\cO}{\mathcal{O}}
\newcommand{\cP}{\mathcal{P}}
\newcommand{\cQ}{\mathcal{Q}} 
\newcommand{\cR}{\mathcal{R}}
\newcommand{\cS}{\mathcal{S}}
\newcommand{\cT}{\mathcal{T}}
\newcommand{\cU}{\mathcal{U}}
\newcommand{\cu}{\mathcal{u}} 
\newcommand{\cV}{\mathcal{V}}
\newcommand{\cW}{\mathcal{W}} 
\newcommand{\cX}{\mathcal{X}} 
\newcommand{\cY}{\mathcal{Y}} 
\newcommand{\cZ}{\mathcal{Z}} 
\newcommand{\sF}{\mathsf{F}} 
\newcommand{\sM}{\mathsf{M}} 
\newcommand{\sG}{\mathsf{G}} 
\newcommand{\sT}{\mathsf{T}} 
\newcommand{\sB}{\mathsf{B}} 
\newcommand{\sC}{\mathsf{C}} 
\newcommand{\sP}{\mathsf{P}} 
\newcommand{\sQ}{\mathsf{Q}} 
\newcommand{\sq}{\mathsf{q}} 
\newcommand{\sR}{\mathsf{R}} 
\newcommand{\sS}{\mathsf{S}}
\newcommand{\sd}{\mathsf{d}} 
\newcommand{\cp}{\mathsf{p}}
\newcommand{\cc}{\mathsf{c}}
\newcommand{\cf}{\mathsf{f}}
\newcommand{\eU}{{\boldsymbol{U}}}
\newcommand{\eb}{{\boldsymbol{b}}}
\newcommand{\ed}{{\boldsymbol{d}}}
\newcommand{\eu}{{\boldsymbol{u}}}
\newcommand{\ew}{{\boldsymbol{w}}}
\newcommand{\ep}{{\boldsymbol{p}}}
\newcommand{\eX}{{\boldsymbol{X}}}
\newcommand{\ex}{{\boldsymbol{x}}}
\newcommand{\eY}{{\boldsymbol{Y}}}
\newcommand{\eB}{{\boldsymbol{B}}}
\newcommand{\eC}{{\boldsymbol{C}}}
\newcommand{\eD}{{\boldsymbol{D}}}
\newcommand{\eW}{{\boldsymbol{W}}}
\newcommand{\eR}{{\boldsymbol{R}}}
\newcommand{\eQ}{{\boldsymbol{Q}}}
\newcommand{\eS}{{\boldsymbol{S}}}
\newcommand{\eT}{{\boldsymbol{T}}}
\newcommand{\eA}{{\boldsymbol{A}}}
\newcommand{\eH}{{\boldsymbol{H}}}
\newcommand{\ea}{{\boldsymbol{a}}}
\newcommand{\ey}{{\boldsymbol{y}}}
\newcommand{\eZ}{{\boldsymbol{Z}}}
\newcommand{\eG}{{\boldsymbol{G}}}
\newcommand{\ez}{{\boldsymbol{z}}}
\newcommand{\es}{{\boldsymbol{s}}}
\newcommand{\et}{{\boldsymbol{t}}}
\newcommand{\ev}{{\boldsymbol{v}}}
\newcommand{\ee}{{\boldsymbol{e}}}
\newcommand{\eq}{{\boldsymbol{q}}}
\newcommand{\bnu}{{\boldsymbol{\nu}}}
\newcommand{\barX}{\overline{\eX}}   


\newcommand{\eps}{\varepsilon}
\newcommand{\Eps}{\mathcal{E}}
\newcommand{\carrier}{{\mathfrak{X}}}
\newcommand{\Ball}{{\mathbb{B}}^{d}}
\newcommand{\Sphere}{{\mathbb{S}}^{d-1}}
\newcommand{\salg}{\mathfrak{F}}
\newcommand{\ssalg}{\mathfrak{B}}
\newcommand{\one}{\mathbf{1}}
\newcommand{\Prob}[1]{\P\{#1\}}
\newcommand{\yL}{\ey_{\mathrm{L}}}
\newcommand{\yU}{\ey_{\mathrm{U}}}
\newcommand{\yLi}{\ey_{\mathrm{L}i}}
\newcommand{\yUi}{\ey_{\mathrm{U}i}}
\newcommand{\xL}{\ex_{\mathrm{L}}}
\newcommand{\xU}{\ex_{\mathrm{U}}}
\newcommand{\vL}{\ev_{\mathrm{L}}}
\newcommand{\vU}{\ev_{\mathrm{U}}}
\newcommand{\dist}{\mathbf{d}}
\newcommand{\rhoH}{\dist_{\mathrm{H}}}
\newcommand{\ti}{\to\infty}
\newcommand{\comp}[1]{#1^\mathrm{c}}
\newcommand{\ThetaI}{\Theta_{\mathrm{I}}}
\newcommand{\crit}{q}
\newcommand{\CS}{CS_n}
\newcommand{\CI}{CI_n}
\newcommand{\cv}[1]{\hat{c}_{n,1-\alpha}(#1)}
\newcommand{\idr}[1]{\mathcal{H}_\sP[#1]}
\newcommand{\outr}[1]{\mathcal{O}_\sP[#1]}
\newcommand{\idrn}[1]{\hat{\mathcal{H}}_{\sP_n}[#1]}
\newcommand{\outrn}[1]{\mathcal{O}_{\sP_n}[#1]}
\newcommand{\email}[1]{\texttt{#1}}
\newcommand{\possessivecite}[1]{<ref name="#1"></ref>'s \citeyear{#1}}
\newcommand\xqed[1]{%
  \leavevmode\unskip\penalty9999 \hbox{}\nobreak\hfill
  \quad\hbox{#1}}
\newcommand\qedex{\xqed{$\triangle$}}
\newcommand\independent{\perp\!\!\!\perp}
\DeclareMathOperator{\Int}{Int}
\DeclareMathOperator{\conv}{conv}
\DeclareMathOperator{\cov}{Cov}
\DeclareMathOperator{\var}{Var}
\DeclareMathOperator{\Sel}{Sel}
\DeclareMathOperator{\Bel}{Bel}
\DeclareMathOperator{\cl}{cl}
\DeclareMathOperator{\sgn}{sgn}
\DeclareMathOperator{\essinf}{essinf}
\DeclareMathOperator{\esssup}{esssup}
\newcommand{\mathds}{\mathbb}
\renewcommand{\P}{\mathbb{P}}
</math>
</div>
===Why Partial Identification?===
Knowing the population distribution that data are drawn from, what can one learn about a parameter of interest?
It has long been understood that assumptions about the data generating process (DGP) play a crucial role in answering this ''identification question'' at the core of all empirical research.
Inevitably, assumptions brought to bear enjoy a varying degree of credibility.
Some are rooted in economic theory (e.g., optimizing behavior) or in information available to the researcher on the DGP (e.g., randomization mechanisms).
These assumptions can be argued to be highly credible.
Others are driven by concerns for tractability and the desire to answer the identification question with a certain level of precision (e.g., functional form and distributional assumptions).
These are arguably less credible.
Early on, <ref name="koo:rei50"><span style="font-variant-caps:small-caps">Koopmans, T.C.,  <span style="font-variant-caps:normal">and</span> O.Reiersol</span>  (1950): “The  Identification of Structural Characteristics” ''The Annals of  Mathematical Statistics'', 21(2), 165--181.</ref> highlighted the importance of imposing restrictions based on prior knowledge of the phenomenon under analysis and some criteria of simplicity, but not for the purpose of identifiability of a parameter that the researcher happens to be interested in, stating (p.169):
<blockquote>One might regard problems of identifiability as a necessary part of the specification problem.
We would consider such a classification acceptable, provided the temptation to specify models in such a way as to produce identifiability of relevant characteristics is resisted.</blockquote>
Much work, spanning multiple fields, has been devoted to putting forward strategies to carry out empirical research while relaxing distributional, functional form, or behavioral assumptions.
One example, embodied in the research program on semiparameteric and nonparametric methods, is to characterize sufficient sets of assumptions, that exclude many suspect ones --sometimes as many as possible-- to guarantee that point identification of specific economically interesting parameters attains.
This literature is reviewed in, e.g., <ref name="mat07"><span style="font-variant-caps:small-caps">Matzkin, R.L.</span>  (2007): “Chapter 73 -- Nonparametric identification” in  ''Handbook of Econometrics'', ed. by J.J. Heckman,  <span style="font-variant-caps:normal">and</span> E.E.  Leamer, vol.6, chap.73, pp. 5307 -- 5368. Elsevier.</ref><ref name="mat13"><span style="font-variant-caps:small-caps">Matzkin, R.L.</span>  (2013): “Nonparametric Identification in Structural  Economic Models” ''Annual Review of Economics'', 5(1), 457--486.</ref>, and is not discussed here.
Another example, embodied in the research program on Bayesian model uncertainty, is to specify multiple models (i.e., multiple sets of assumptions), put a prior on the parameters of each model and on each model, embed the various separate models within one large hierarchical mixture model, and obtain model posterior probabilities which can be used for a variety of inferences and decisions.
This literature is reviewed in, e.g., <ref name="was00"><span style="font-variant-caps:small-caps">Wasserman, L.</span>  (2000): “Bayesian Model Selection and Model  Averaging” ''Journal of Mathematical Psychology'', 44(1), 92 -- 107.</ref> and <ref name="cly:geo04"><span style="font-variant-caps:small-caps">Clyde, M.,  <span style="font-variant-caps:normal">and</span> E.I. George</span>  (2004): “Model Uncertainty”  ''Statist. Sci.'', 19(1), 81--94.</ref>, and is not discussed here.
The approach considered here fixes a set of assumptions and a parameter of interest a priori, in the spirit of <ref name="koo:rei50"/>, and asks what can be learned about that parameter given the available data, recognizing that even partial information can be illuminating for empirical research, while enjoying wider credibility thanks to the weaker assumptions imposed.
The bounding methods at the core of this approach appeared in the literature nearly a century ago.
Arguably, the first exemplar that leverages economic reasoning is given by the work of <ref name="mar:and44"><span style="font-variant-caps:small-caps">Marschak, J.,  <span style="font-variant-caps:normal">and</span> W.H. Andrews</span>  (1944): “Random  Simultaneous Equations and the Theory of Production” ''Econometrica'',  12(3/4), 143--205.</ref>.
They provided bounds on Cobb-Douglas production functions in models of supply and demand, building on optimization principles and restrictions from microeconomic theory.
<ref name="lea81"><span style="font-variant-caps:small-caps">Leamer, E.E.</span>  (1981): “Is it a Demand Curve, Or Is It A Supply  Curve? Partial Identification through Inequality Constraints” ''The  Review of Economics and Statistics'', 63(3), 319--327.</ref> revisited their analysis to obtain bounds on the elasticities of demand and supply in a linear simultaneous equations system with uncorrelated errors.
The first exemplars that do not rely on specific economic models appear in <ref name="gin21"><span style="font-variant-caps:small-caps">Gini, C.</span>  (1921): “Sull'interpolazione di una retta quando i valori  della variabile indipendente sono affetti da errori accidentali”  ''Metroeconomica'', 1(3), 63–82.</ref>, <ref name="fri34"><span style="font-variant-caps:small-caps">Frisch, R.</span>  (1934): ''Statistical Confluence Analysis by Means of  Complete Regression Systems'', Okonomiske Institutt Oslo: Publikasjon.  Universitetets {\O}konomiske Instituut.</ref>, and <ref name="rei41"><span style="font-variant-caps:small-caps">Reiersol, O.</span>  (1941): “Confluence Analysis by Means of Lag Moments  and Other Methods of Confluence Analysis” ''Econometrica'', 9(1), 1--24.</ref>, who bounded the coefficient of a simple linear regression in the presence of measurement error.
These results were extended to the general linear regression model with errors in all variables by <ref name="kle:lea84"><span style="font-variant-caps:small-caps">Klepper, S.,  <span style="font-variant-caps:normal">and</span> E.E. Leamer</span>  (1984): “Consistent Sets of  Estimates for Regressions with Errors in All Variables”  ''Econometrica'', 52(1), 163--183.</ref> and <ref name="lea87"><span style="font-variant-caps:small-caps">Leamer, E.E.</span>  (1987): “Errors in Variables in Linear Systems”  ''Econometrica'', 55(4), 893--909.</ref>.
This chapter surveys some of the methods proposed over the last thirty years in the microeconometrics literature to further this approach.
These methods belong to the systematic program on ’'partial identification'' analysis started with <ref name="man89"><span style="font-variant-caps:small-caps">Manski, C.F.</span>  (1989): “Anatomy of the Selection Problem” ''The  Journal of Human Resources'', 24(3), 343--360.</ref><ref name="man90"><span style="font-variant-caps:small-caps">Manski, C.F.</span>  (1990): “Nonparametric Bounds on Treatment Effects”  ''The American Economic Review Papers and Proceedings'', 80(2), 319--323.</ref><ref name="man95"><span style="font-variant-caps:small-caps">Manski, C.F.</span>  (1995): ''Identification Problems in the Social  Sciences''. Harvard University Press.</ref><ref name="man03"><span style="font-variant-caps:small-caps">Manski, C.F.</span>  (2003): ''Partial Identification of Probability  Distributions'', Springer Series in Statistics. Springer.</ref><ref name="man07a"><span style="font-variant-caps:small-caps">Manski, C.F.</span>  (2007a): ''Identification for Prediction and Decision''.  Harvard University Press.</ref><ref name="man13book"><span style="font-variant-caps:small-caps">Manski, C.F.</span>  (2013b): ''Public Policy in an Uncertain World:  Analysis and Decisions''. Harvard University Press.</ref> and developed by several authors since the early 1990s.
Within this program, the focus shifts from points to sets: the researcher aims to learn what is the set of values for the parameters of interest that can generate the same distribution of observables as the one in the data, for some DGP consistent with the maintained assumptions.
In other words, the focus is on the ''set of observationally equivalent'' values, which henceforth I refer to as the parameters'  ''sharp identification region''.
In the partial identification paradigm, empirical analysis begins with characterizing this set using the data alone.
This is a nonparametric approach that dispenses with all assumptions, except basic restrictions on the sampling process such that the distribution of the observable variables can be learned as data accumulate.
In subsequent steps, one incorporates additional assumptions into the analysis, reporting how each assumption (or set of assumptions) affects what one can learn about the parameters of interest, i.e., how it modifies and possibly shrinks the sharp identification region.
Point identification may result from the process of increasingly strengthening the maintained assumptions, but it is not the goal in itself.
Rather, the objective is to make transparent the relative role played by the data and the assumptions in shaping the inference that one draws.
There are several strands of independent, but thematically related literatures that are not discussed in this chapter.
As a consequence, many relevant contributions are left out of the presentation and the references.
One example is the literature in finance.
<ref name="han:jag91"><span style="font-variant-caps:small-caps">Hansen, L.P.,  <span style="font-variant-caps:normal">and</span> R.Jagannathan</span>  (1991): “Implications of  Security Market Data for Models of Dynamic Economies” ''Journal of  Political Economy'', 99(2), 225--262.</ref> developed nonparametric bounds for the admissible set for means and standard deviations of intertemporal marginal rates of substitution (IMRS) of consumers. 
The bounds were developed exploiting the condition, satisfied in many finance models, that the equilibrium price of any traded security equals the expectation (conditioned on current information) of the product's future payoff and the IMRS of any consumer.<ref group="Notes">
{{Ref|name=han:jag91}} deduce a duality relation with the mean variance theory of {{Ref|name=mar52}} and {{Ref|name=fam96}}, but the relation does not apply to the sharp bounds they derive. 
In the Arbitrage Pricing Model {{Ref|name=ros76}}, bounds on extensions of existing pricing functions, consistent with the absence of arbitrage opportunities, were considered by {{Ref|name=har:kre79}} and {{Ref|name=kre81}}.
</ref>
<ref name="lut96"><span style="font-variant-caps:small-caps">Luttmer, E. G.J.</span>  (1996): “Asset Pricing in Economies with  Frictions” ''Econometrica'', 64(6), 1439--1467.</ref> extended the analysis to economies with frictions. 
<ref name="han:hea:lut95"><span style="font-variant-caps:small-caps">Hansen, L.P., J.Heaton,  <span style="font-variant-caps:normal">and</span> E.G.J. Luttmer</span>  (1995):  “Econometric Evaluation of Asset Pricing Models” ''The Review of  Financial Studies'', 8(2), 237--274.</ref> developed econometric tools to estimate the regions, to assess asset pricing models, and to provide nonparametric characterizations of asset pricing anomalies. 
Earlier on, the existence of volatility bounds on IMRSs were noted by <ref name="shi82"><span style="font-variant-caps:small-caps">Shiller, R.J.</span>  (1982): “Consumption, asset markets and macroeconomic  fluctuations” ''Carnegie-Rochester Conference Series on Public Policy'',  17, 203 -- 238.</ref> and <ref name="han82comment"><span style="font-variant-caps:small-caps">Hansen, L.P.</span>  (1982a): “Consumption, asset markets, and  macroeconomic fluctuations: A comment” ''Carnegie-Rochester Conference  Series on Public Policy'', 17, 239 -- 250.</ref>. 
The bounding arguments that build on the minimum-volatility frontier for stochastic discount factors proposed by <ref name="han:jag91"/> have become a litmus test to detect anomalies in asset pricing models (see, e.g. <ref name="shi03"><span style="font-variant-caps:small-caps">Shiller, R.J.</span>  (2003): “From Efficient Markets Theory to Behavioral  Finance” ''Journal of Economic Perspectives'', 17(1), 83--104.</ref>{{rp|at=p. 89}}).
I refer to the textbook presentations in <ref name="lju:sar04"><span style="font-variant-caps:small-caps">Ljungqvist, L.,  <span style="font-variant-caps:normal">and</span> T.Sargent</span>  (2004): ''Recursive  Macroeconomic Theory'', vol.1. The MIT Press, 2 edn.</ref>{{rp|at=Chapter 13}} and <ref name="coc05"><span style="font-variant-caps:small-caps">Cochrane, J.H.</span>  (2005): ''Asset Pricing''. Princeton University  Press, Princeton, New Jersey, second edn.</ref>{{rp|at=Chapters 5 and 21}}, and the review articles by <ref name="fer03"><span style="font-variant-caps:small-caps">Ferson, W.E.</span>  (2003): “Chapter 12 -- Tests of multifactor pricing  models, volatility bounds and portfolio performance” in ''Financial  Markets and Asset Pricing'', vol.1 of ''Handbook of the Economics of  Finance'', pp. 743 -- 802. Elsevier.</ref> and <ref name="cam14"><span style="font-variant-caps:small-caps">Campbell, J.Y.</span>  (2014): “Empirical Asset Pricing: Eugene Fama, Lars  Peter Hansen, and Robert Shiller” ''The Scandinavian Journal of  Economics'', 116(3), 593--634.</ref>, for a careful presentation of this literature.
In macroeconomics, <ref name="fau98"><span style="font-variant-caps:small-caps">Faust, J.</span>  (1998): “The robustness of identified VAR conclusions  about money” ''Carnegie-Rochester Conference Series on Public Policy'',  49, 207 -- 244.</ref>, <ref name="can:den02"><span style="font-variant-caps:small-caps">Canova, F.,  <span style="font-variant-caps:normal">and</span> G.{De Nicolo}</span>  (2002): “Monetary  disturbances matter for business fluctuations in the G-7” ''Journal of  Monetary Economics'', 49(6), 1131 -- 1159.</ref>, and <ref name="uhl05"><span style="font-variant-caps:small-caps">Uhlig, H.</span>  (2005): “What are the effects of monetary policy on  output? Results from an agnostic identification procedure” ''Journal of  Monetary Economics'', 52(2), 381 -- 419.</ref> proposed bounds for impulse response functions in sign-restricted structural vector autoregression models, and carried out Bayesian inference with a non-informative prior for the non-identified parameters.
I refer to <ref name="kil:lut17"><span style="font-variant-caps:small-caps">Kilian, L.,  <span style="font-variant-caps:normal">and</span> H.L\"{u}tkepohl</span>  (2017): ''Structural  Vector Autoregressive Analysis'', Themes in Modern Econometrics. Cambridge  University Press.</ref>{{rp|at=Chapter 13}} for a careful presentation of this literature.
In microeconomic theory, bounds were derived from inequalities resulting as necessary and sufficient conditions that data on an individual's choice need to satisfy in order to be consistent with optimizing behavior, as in the research pioneered by <ref name="sam38"><span style="font-variant-caps:small-caps">Samuelson, P.A.</span>  (1938): “A Note on the Pure Theory of Consumer's  Behaviour” ''Economica'', 5(17), 61--71.</ref> and advanced early on by <ref name="hou50"><span style="font-variant-caps:small-caps">Houthakker, H.S.</span>  (1950): “Revealed Preference and the Utility  Function” ''Economica'', 17(66), 159--174.</ref> and <ref name="ric66"><span style="font-variant-caps:small-caps">Richter, M.K.</span>  (1966): “Revealed Preference Theory”  ''Econometrica'', 34(3), 635--645.</ref>.
<ref name="afr67"><span style="font-variant-caps:small-caps">Afriat, S.N.</span>  (1967): “The Construction of Utility Functions from  Expenditure Data” ''International Economic Review'', 8(1), 67--77.</ref> and <ref name="var82"><span style="font-variant-caps:small-caps">Varian, H.R.</span>  (1982): “The Nonparametric Approach to Demand  Analysis” ''Econometrica'', 50(4), 945--973.</ref> extended this research program to revealed preference extrapolation.
Notably, in this work no stochastic terms enter the analysis.
<ref name="blo:mar60"><span style="font-variant-caps:small-caps">Block, H.D.,  <span style="font-variant-caps:normal">and</span> J.Marschak</span>  (1960): “Random Orderings  and Stochastic Theories of Responses” in ''Contributions to Probability  and Statistics: Essays in Honor of Harold Hotelling'', ed. by I.Olkin, pp.  97--132. Stanford University Press.</ref>, <ref name="mar60"><span style="font-variant-caps:small-caps">Marschak, J.</span>  (1960): “Binary Choice Constraints on Random Utility  Indicators” in ''Stanford Symposium on Mathematical Methods in the  Social Sciences'', ed. by K.Arrow. Stanford University Press.</ref>, <ref name="hal73"><span style="font-variant-caps:small-caps">Hall, R.E.</span>  (1973): “On the statistical theory of unobserved  components” MIT Working Paper 117, available at  [https://dspace.mit.edu/bitstream/handle/1721.1/63972/onstatisticalthe00hall.pdf?sequence=1 https://dspace.mit.edu/bitstream/handle/1721.1/63972/onstatisticalthe00hall.pdf?sequence=1].</ref>, <ref name="mcf75"><span style="font-variant-caps:small-caps">McFadden, D.L.</span>  (1975): “Tchebyscheff bounds for the space of agent  characteristics” ''Journal of Mathematical Economics'', 2(2), 225 --  242.</ref>, <ref name="fal78"><span style="font-variant-caps:small-caps">Falmagne, J.</span>  (1978): “A representation theorem for finite random  scale systems” ''Journal of Mathematical Psychology'', 18(1), 52 -- 72.</ref>, and <ref name="mcf:ric91"><span style="font-variant-caps:small-caps">McFadden, D.L.,  <span style="font-variant-caps:normal">and</span> M.K. Richter</span>  (1991): “Stochastic  rationality and revealed stochastic preference” in ''Preferences,  Uncertainty and Rationality'', ed. by J.S. Chipman, D.L. McFadden,  <span style="font-variant-caps:normal">  and</span> M.K. Richter, pp. 161--186. Westview Press.</ref>, extended revealed preference arguments to random utility models, and obtained bounds on the distributions of preferences.
I refer to the survey articles by <ref name="cra:der14"><span style="font-variant-caps:small-caps">Crawford, I.,  <span style="font-variant-caps:normal">and</span> B.DeRock</span>  (2014): “Empirical Revealed  Preference” ''Annual Review of Economics'', 6(1), 503--524.</ref> and <ref name="blu19"><span style="font-variant-caps:small-caps">Blundell, R.</span>  (2019): “Revealed preference” in ''Handbook of  Econometrics''. Elsevier.</ref>{{rp|at=Chapter XXX in this Volume}} for a careful presentation of this literature.
A complementary approach to partial identification is given by sensitivity analysis, advocated for in different ways by, e.g., <ref name="gil:lea83"><span style="font-variant-caps:small-caps">Gilstein, C.Z.,  <span style="font-variant-caps:normal">and</span> E.E. Leamer</span>  (1983): “Robust Sets of  Regression Estimates” ''Econometrica'', 51(2), 321--333.</ref>, <ref name="ros:rub83"><span style="font-variant-caps:small-caps">Rosenbaum, P.R.,  <span style="font-variant-caps:normal">and</span> D.B. Rubin</span>  (1983): “Assessing  Sensitivity to an Unobserved Binary Covariate in an Observational Study with  Binary Outcome” ''Journal of the Royal Statistical Society. Series B  (Methodological)'', 45(2), 212--218.</ref>, <ref name="lea85"><span style="font-variant-caps:small-caps">Leamer, E.E.</span>  (1985): “Sensitivity Analyses Would Help” ''The  American Economic Review'', 75(3), 308--313.</ref>, <ref name="ros95"><span style="font-variant-caps:small-caps">Rosenbaum, P.R.</span>  (1995): ''Observational Studies''. Springer.</ref>, <ref name="imb03"><span style="font-variant-caps:small-caps">Imbens, G.W.</span>  (2003): “Sensitivity to Exogeneity Assumptions in  Program Evaluation” ''American Economic Review'', 93(2), 126--132.</ref>, and others.
Within this approach, the analysis begins with a fully parametric model that point identifies the parameter of interest.
One then reports the set of values for this parameter that result when the more suspicious assumptions are relaxed.
Related literatures, not discussed in this chapter, abound also outside Economics.
For example, in probability theory, <ref name="hoe40"><span style="font-variant-caps:small-caps">Hoeffding, W.</span>  (1940): “Masstabinvariante Korrelationstheorie”  ''Schriften des Matematischen Instituts und des Instituts f\"{u}r  Angewandte MAthematik des Universit\"{a}t Berlin'', 5(3), 179--233.</ref> and <ref name="fre51"><span style="font-variant-caps:small-caps">Fréchet, M.R.</span>  (1951): “Sur les tableaux de correlation dont les  marges sont donnes” ''Annales de I’Université de Lyon A'', 3, 53--77.</ref> put forward bounds on the joint distributions of random variables, and <ref name="mak81"><span style="font-variant-caps:small-caps">Makarov, G.D.</span>  (1981): “Estimates for the distribution function of a  sum of two random variables when the marginal distributions are fixed”  ''Theory of Probability and its Applications'', 26(4), 803--806.</ref>, <ref name="rus82"><span style="font-variant-caps:small-caps">R\"{u}schendorf, L.</span>  (1982): “Random Variables with Maximum Sums”  ''Advances in Applied Probability'', 14(3), 623--632.</ref>, and <ref name="fra:nel:sch87"><span style="font-variant-caps:small-caps">Frank, M.J., R.B. Nelsen,  <span style="font-variant-caps:normal">and</span> B.Schweizer</span>  (1987):  “Best-possible bounds for the distribution of a sum --- a problem of  Kolmogorov” ''Probability Theory and Related Fields'', 74(2), 199--211.</ref> on the sum of random variables, when only marginal distributions are observed.
The literature on probability bounds is discussed in the textbook by <ref name="sho:wel09"><span style="font-variant-caps:small-caps">Shorack, G.R.,  <span style="font-variant-caps:normal">and</span> J.A. Wellner</span>  (2009): ''Empirical  Processes with Applications to Statistics''. Society for Industrial and  Applied Mathematics.</ref>{{rp|at=Appendix A}}.
Addressing problems faced in economics, sociology, epidemiology, geography, history, political science, and more, <ref name="dun:dav53"><span style="font-variant-caps:small-caps">Duncan, O.D.,  <span style="font-variant-caps:normal">and</span> B.Davis</span>  (1953): “An Alternative to  Ecological Correlation” ''American Sociological Review'', 18(6),  665--666.</ref> derived bounds on correlations among variables measured at the individual level based on observable correlations among variables measured at the aggregate level.
The so called ecological inference problem they studied, and the associated literature, is discussed in the survey article by <ref name="cho:man09"><span style="font-variant-caps:small-caps">Cho, W.T.,  <span style="font-variant-caps:normal">and</span> C.F. Manski</span>  (2009):  “Cross‐Level/Ecological Inference” in ''Oxford Handbook of Political  Methodology'', ed. by J.M. Box-Steffensmeier, H.E. Brady,  <span style="font-variant-caps:normal">and</span>  D.Collier, chap.24, pp. 530--569. Oxford University Press.</ref> and references therein.
===Goals and Structure of this Chapter===
To carry out econometric analysis with partial identification, one needs: (1) computationally feasible characterizations of the parameters' sharp identification region; (2) methods to estimate this region; and (3) methods to test hypotheses and construct confidence sets.
The goal of this chapter is to provide insights into the challenges posed by each of these desiderata, and into some of their solutions.
In order to discuss the partial identification literature in microeconometrics with some level of detail while keeping this chapter to a manageable length, I focus on a selection of papers and not on a complete survey of the literature.
As a consequence, many relevant contributions are left out of the presentation and the references.
I also do not discuss the important but
separate topic of statistical decisions in the presence of partial identification, for which I refer to the textbook treatments in <ref name="man05"><span style="font-variant-caps:small-caps">Manski, C.F.</span>  (2005): ''Social Choice with Partial Knowledge of  Treatment Response''. Princeton University Press.</ref><ref name="man07a"/> and to the review by <ref name="hir:por19"><span style="font-variant-caps:small-caps">Hirano, K.,  <span style="font-variant-caps:normal">and</span> J.R. Porter</span>  (2019): “Statistical Decision  Rules in Econometrics” in ''Handbook of Econometrics''. Elsevier.</ref>{{rp|at=Chapter XXX in this Volume}}.
The presumption in identification analysis that the distribution from which the data are drawn is known allows one to keep separate the identification question from the distinct question of statistical inference from a finite sample.
I use the same separation in this chapter.
I assume solid knowledge of the topics covered in first year Economics PhD courses in econometrics and microeconomic theory.
I begin in [[guide:Ec36399528#sec:prob:distr |Section]] with the analysis of what can be learned about features of probability distributions that are well defined in the absence of an economic model, such as moments, quantiles, cumulative distribution functions, etc., when one faces measurement problems.
Specifically, I focus on cases where the ''data is incomplete'', either due to selectively observed data or to interval measurements.
I refer to <ref name="man95"/><ref name="man03"/><ref name="man07a"/> for textbook treatments of many other cases.
I lay out formally the maintained assumptions for several examples, and then discuss in detail what is the source of the identification problem.
I conclude with providing tractable characterizations of what can be learned about the parameters of interest, with formal proofs.
I show that even in simple problems, great care may be needed to obtain the sharp identification region.
It is often easier to characterize an ''outer region'', i.e., a collection of values for the parameter of interest that contains the sharp one but may contain also additional values. Outer regions are useful because of their simplicity and because in certain applications they may suffice to answer questions of great interest, e.g., whether a policy intervention has a nonnegative effect.
However, compared to the sharp identification region they may afford the researcher less useful predictions, and a lower ability to test for misspecification, because they do not harness all the information in the observed data and maintained assumptions.
In [[guide:521939d27a#sec:structural |Section]] I use the same approach to study what can be learned about features of parameters of structural econometric models when the ''model is incomplete'' <ref name="tam03"><span style="font-variant-caps:small-caps">Tamer, E.</span>  (2003): “Incomplete Simultaneous Discrete Response Model  with Multiple Equilibria” ''The Review of Economic Studies'', 70(1),  147--165.</ref><ref name="hai:tam03"><span style="font-variant-caps:small-caps">Haile, P.A.,  <span style="font-variant-caps:normal">and</span> E.Tamer</span>  (2003): “Inference with an  Incomplete Model of English Auctions” ''Journal of Political Economy'',  111(1), 1--51.</ref><ref name="cil:tam09"><span style="font-variant-caps:small-caps">Ciliberto, F.,  <span style="font-variant-caps:normal">and</span> E.Tamer</span>  (2009): “Market Structure and  Multiple Equilibria in Airline Markets” ''Econometrica'', 77(6),  1791--1828.</ref>.
Specifically, I discuss single agent discrete choice models under a variety of challenging situations (interval measured as well as endogenous explanatory variables; unobserved as well as counterfactual choice sets); finite discrete games with multiple equilibria; auction models under weak assumptions on bidding behavior; and network formation models.
Again I formally derive sharp identification regions for several examples.
I conclude each of these sections with a brief discussion of further theoretical advances and empirical applications that is meant to give a sense of the breadth of the approach, but not to be exhaustive.
I refer to the recent survey by <ref name="ho:ros17"><span style="font-variant-caps:small-caps">Ho, K.,  <span style="font-variant-caps:normal">and</span> A.M. Rosen</span>  (2017): “Partial Identification in  Applied Research: Benefits and Challenges” in ''Advances in Economics  and Econometrics: Eleventh World Congress'', ed. by B.Honoré, A.Pakes,  M.Piazzesi,  <span style="font-variant-caps:normal">and</span> L.Samuelson, vol.1 of ''Econometric Society  Monographs'', pp. 307--359. Cambridge University Press.</ref> for a thorough discussion of empirical applications of partial identification methods.
In [[guide:6d1a428897#sec:inference |Section]] I discuss finite sample inference.
I limit myself to highlighting the challenges that one faces for consistent estimation when the identified object is a set, and several coverage notions and requirements that have been proposed over the last 20 years.
I refer to the recent survey by <ref name="can:sha17"><span style="font-variant-caps:small-caps">Canay, I.A.,  <span style="font-variant-caps:normal">and</span> A.M. Shaikh</span>  (2017): “Practical and  Theoretical Advances in Inference for Partially Identified Models” in  ''Advances in Economics and Econometrics: Eleventh World Congress'', ed.  by B.Honoré, A.Pakes, M.Piazzesi,  <span style="font-variant-caps:normal">and</span> L.Samuelson, vol.2 of  ''Econometric Society Monographs'', p. 271–306. Cambridge University  Press.</ref> for a thorough discussion of methods to tests hypotheses and build confidence sets in moment inequality models.
In [[guide:7b0105e1fc#sec:misspec |Section]] I discuss the distinction between refutable and non-refutable assumptions, and how model misspecification may be detectable in the presence of the former, even within the partial identification paradigm.
I then highlight certain challenges that model misspecification presents for the interpretation of sharp identification (as well as outer) regions, and for the construction of confidence sets.
In [[guide:A85a6b6ff1#sec:computations |Section]] I highlight that while most of the sharp identification regions characterized in [[guide:Ec36399528#sec:prob:distr |Section]] can be easily computed, many of the ones in [[guide:521939d27a#sec:structural |Section]] are more challenging.
This is because the latter are obtained as level sets of criterion functions in moderately dimensional spaces, and tracing out these level sets or their boundaries is a non-trivial computational problem.
In [[guide:A16691a720#sec:future |Section]] I conclude providing some considerations on what I view as open questions for future research.
I refer to <ref name="tam10"><span style="font-variant-caps:small-caps">Tamer, E.</span>  (2010): “Partial Identification in Econometrics”  ''Annual Review of Economics'', 2, 167--195.</ref> for an earlier review of this literature, and to <ref name="lew18"><span style="font-variant-caps:small-caps">Lewbel, A.</span>  (2018): “The Identification Zoo - Meanings of  Identification in Econometrics” ''Journal of Economic Literature'',  forthcoming.</ref> for a careful presentation of the many notions of identification that are used across the econometrics literature, including an important historical account of how these notions developed over time.
===Random Set Theory as a Tool for Partial Identification Analysis===
Throughout [[guide:Ec36399528#sec:prob:distr |Sections]] and [[guide:521939d27a#sec:structural |Section]], a simple organizing principle for much of partial identification analysis emerges.
The cause of the identification problems discussed can be traced back to a collection of random variables that are consistent with the available data and maintained assumptions.
For the problems studied in [[guide:Ec36399528#sec:prob:distr |Section]], this set is often a simple function of the observed variables.
The incompleteness of the data stems from the fact that instead of observing the singleton variables of interest, one observes set-valued variables to which these belong, but one has no information on their exact value within the sets.
For the problems studied in [[guide:521939d27a#sec:structural |Section]], the collection of random variables consistent with the maintained assumptions comprises what the model predicts for the endogenous variable(s).
The incompleteness of the model stems from the fact that instead of making a singleton prediction for the variable(s) of interest, the model makes multiple predictions but does not specify how one is chosen.
The central role of set-valued objects, both stochastic and nonstochastic, in partial identification renders ''random set theory'' a natural toolkit to aid the analysis.<ref group="Notes" >The first idea of a general random set in the form of a region that depends on chance appears in {{ref|name=kol50}}, originally published in 1933.
For another early example where confidence regions are explicitly described as random sets, see {{ref|name=haa44}}{{rp|at=p. 67}}.
The role of random sets in this chapter is different.</ref>
This theory originates in the seminal contributions of <ref name="cho53"><span style="font-variant-caps:small-caps">Choquet, G.</span>  (1953/54): “Theory of capacities” ''Annales de  l'Institut Fourier (Grenoble)'', 5, 131--295.</ref>, <ref name="aum65"><span style="font-variant-caps:small-caps">Aumann, R.J.</span>  (1965): “Integrals of set-valued functions”  ''Journal of Mathematical Analysis and Applications'', 12(1), 1--12.</ref>, and <ref name="deb67"><span style="font-variant-caps:small-caps">Debreu, G.</span>  (1967): “Integration of correspondences” in  ''Proceedings of the Fifth Berkeley Symposium in Mathematical Statistics  and Probability'', vol.2, pp. 351--372. University of California Press.</ref>, with the first self contained treatment of the theory given by <ref name="mat75"><span style="font-variant-caps:small-caps">Matheron, G.</span>  (1975): ''Random Sets and Integral Geometry''. Wiley,  New York.</ref>. 
I refer to <ref name="mo1"><span style="font-variant-caps:small-caps">Molchanov, I.</span>  (2017): ''Theory of Random Sets''. Springer, London,  2 edn.</ref> for a textbook presentation, and to <ref name="mol:mol14"><span style="font-variant-caps:small-caps">Molchanov, I.,  <span style="font-variant-caps:normal">and</span> F.Molinari</span>  (2014): “Applications of  Random Set Theory in Econometrics” ''Annual Review of Economics'', 6(1),  229--251.</ref><ref name="mol:mol18"><span style="font-variant-caps:small-caps">Molchanov, I.,  <span style="font-variant-caps:normal">and</span> F.Molinari</span>  (2018): ''Random Sets in Econometrics''. Econometric  Society Monograph Series, Cambridge University Press, Cambridge UK.</ref> for a treatment focusing on its applications in econometrics.
<ref name="ber:mol08"><span style="font-variant-caps:small-caps">Beresteanu, A.,  <span style="font-variant-caps:normal">and</span> F.Molinari</span>  (2008): “Asymptotic  Properties for a Class of Partially Identified Models” ''Econometrica'',  76(4), 763--814.</ref> introduce the use of random set theory in econometrics to carry out identification analysis and statistical inference with incomplete data.
<ref name="ber:mol:mol11"><span style="font-variant-caps:small-caps">Beresteanu, A., I.Molchanov,  <span style="font-variant-caps:normal">and</span> F.Molinari</span>  (2011): “Sharp identification regions in models with  convex moment predictions” ''Econometrica'', 79(6), 1785--1821.</ref><ref name="ber:mol:mol12"><span style="font-variant-caps:small-caps">Beresteanu, A., I.Molchanov,  <span style="font-variant-caps:normal">and</span> F.Molinari</span>  (2012): “Partial identification using random set theory”  ''Journal of Econometrics'', 166(1), 17 -- 32, with errata at  [https://molinari.economics.cornell.edu/docs/NOTE_BMM2012_v3.pdf https://molinari.economics.cornell.edu/docs/NOTE_BMM2012_v3.pdf].</ref> propose it to characterize sharp identification regions both with incomplete data and with incomplete models.
<ref name="gal:hen11"><span style="font-variant-caps:small-caps">Galichon, A.,  <span style="font-variant-caps:normal">and</span> M.Henry</span>  (2011): “Set Identification in Models with Multiple  Equilibria” ''The Review of Economic Studies'', 78(4), 1264--1298.</ref> propose the use of optimal transportation methods that in some applications deliver the same characterizations as the random set methods.
I do not discuss optimal transportation methods in this chapter, but refer to <ref name="gal16"><span style="font-variant-caps:small-caps">Galichon, A.</span>  (2016): ''Optimal Transport Methods in Economics''.  Princeton University Press.</ref> for a thorough treatment.
Over the last ten years, random set methods have been used to unify a number of specific results in partial identification, and to produce a general methodology for identification analysis that dispenses completely with case-by-case distinctions.
In particular, as I show throughout the chapter, the methods allow for simple and tractable characterizations of sharp identification regions.
The collection of these results establishes that indeed this is a useful tool to carry out econometrics with partial identification, as exemplified by its prominent role both in this chapter and in Chapter XXX in this Volume by <ref name="che:ros19"><span style="font-variant-caps:small-caps">Chesher, A.,  <span style="font-variant-caps:normal">and</span> A.M. Rosen</span>  (2019): “Generalized instrumental variable models,  methods, and applications” in ''Handbook of Econometrics''. Elsevier.</ref>, which focuses on general classes of instrumental variable models.
The random sets approach complements the more traditional one, based on mathematical tools for (single valued) random vectors, that proved extremely productive since the beginning of the research program in partial identification.
This chapter shows that to fruitfully apply random set theory for identification and inference, the econometrician needs to carry out three fundamental steps.
First, she needs to define the random closed set that is relevant for the problem under consideration using all information given by the available data and maintained assumptions.
This is a delicate task, but one that is typically carried out in identification analysis regardless of whether random set theory is applied.
Indeed, throughout the chapter I highlight how relevant random closed sets were characterized in partial identification analysis since the early 1990s, albeit the connection to the theory of random sets was not made. As a second step, the econometrician needs to determine how the observable random variables relate to the random closed set.
Often, one of two cases occurs: either the observable variables determine a random set to which the unobservable variable of interest belongs with probability one, as in incomplete data scenarios; or the (expectation of the) (un)observable variable belongs to (the expectation of) a random set determined by the model, as in incomplete model scenarios.
Finally, the econometrician needs to determine which tool from random set theory should be utilized.
To date, new applications of random set theory to econometrics have fruitfully exploited (Aumann) expectations and their support functions, (Choquet) capacity functionals, and laws of large numbers and central limit theorems for random sets.
[[guide:379e0dcd67#app:RCS |Appendix]] reports basic definitions from random set theory of these concepts, as well as some useful theorems.
The chapter explains in detail through applications to important identification problems how these steps can be carried out.
<span id="tab:notation"/>
{|class="table"
|+ Notation Used
|-
|<math>(\Omega,\salg,\P)</math> || Nonatomic probability space
|-
|<math>\R^d,\|\cdot\|</math> || Euclidean space equipped with the Euclidean norm
|-
|<math>\cF,\cG,\cK</math> || Collection of closed, open, and compact subsets of <math>\R^d</math> (respectively)
|-
|<math>\Sphere = \{x \in \R^d: \|x\| = 1\}</math> || Unit sphere in <math>\R^d</math>
|-
|<math>\Ball = \{x \in \R^d: \|x\| \leq 1\}</math> || Unit ball in <math>\R^d</math>
|-
|<math>\conv(A),\cl(A),|B|</math> || Convex hull and closure of a set <math>A\subset\R^d</math> (respectively), and cardinality of a finite set <math>B\subset\R^d</math>
|-
|<math>\ex,\ey,\ez,\dots</math> || Random vectors
|-
|<math>x,y,z,\dots</math> || Realizations of random vectors or deterministic vectors
|-
|<math>\eX,\eY,\eZ,\dots</math> || Random sets
|-
|<math>X,Y,Z,\dots</math> || Realizations of random sets or deterministic sets
|-
|<math>\epsilon,\eps,\nu,\zeta</math> || Unobserved random variables (heterogeneity)
|-
|<math>\Theta,\theta,\vartheta</math> || Parameter space, data generating value for the  parameter vector, and a generic element of <math>\Theta</math> 
|-
|<math>\sR</math> || Joint distribution of all variables (observable and unobservable)
|-
|<math>\sP</math> || Joint distribution of the observable variables
|-
|<math>\sQ</math> || Joint distribution whose features one wants to learn
|-
|<math>\sM</math> || A joint distribution of observed variables implied by the model
|-
|<math>\sq_{\tau}(\alpha)</math> || Quantile function at level <math>\alpha \in (0,1)</math> for a random variable distributed <math>\tau\in\{\sR,\sP,\sQ\}</math>
|-
|<math>\E_\tau</math> || Expectation operator associated with distribution <math>\tau\in\{\sR,\sP,\sQ\}</math>
|-
|<math>\sT_\eX(K)=\Prob{\eX\cap K\neq\emptyset},\, K\in\cK</math> || Capacity functional of random set <math>\eX</math>
|-
|<math>\sC_\eX(F)=\Prob{\eX\subset F},\, F\in\cF</math> || Containment functional of random set <math>\eX</math>
|-
|<math>\stackrel{p}{\rightarrow},\stackrel{\text{a.s.}}{\rightarrow},\Rightarrow</math> || Convergence in probability, convergence almost surely, and weak convergence (respectively)
|-
|<math>\ex\edis\ey</math> || <math>\ex</math> and <math>\ey</math> have the same distribution
|-
|<math>\ex\independent\ey</math> || Statistical independence between random variables <math>\ex</math> and <math>\ey</math>
|-
|<math>x^\top y</math> || Inner product between vectors <math>x</math> and <math>y</math>, <math>x,y\in\R^d</math>
|-
|<math>\bU,\bu</math> || Family of utility functions and one of its elements
|-
|<math>\crit_\sP</math> || Criterion function that aggregates violations of the population moment inequalities
|-
|<math>\crit_n</math> || Criterion function that aggregates violations of the sample moment inequalities
|-
|<math>\idr{\cdot}</math> || Sharp identification region of the functional in square brackets (a function of <math>\sP</math>)
|-
|<math>\outr{\cdot}</math> || An outer region of the functional in square brackets (a function of <math>\sP</math>)
|}
===<span id="subsec:notation"></span>Notation===
This chapter employs consistent notation that is summarized in [[#tab:notation |Table]].
Some important conventions are as follows:
<math>\ey</math> denotes outcome variables, <math>(\ex,\ew)</math> denote explanatory variables, and <math>\ez</math> denotes instrumental variables (i.e., variables that satisfy some form of independence with the outcome or with the unobservable variables, possibly conditional on <math>\ex,\ew</math>).
I denote by <math>\sP</math> the joint distribution of all observable variables.
Identification analysis is carried out using the information contained in this distribution, and finite sample inference is carried out under the presumption that one draws a random sample of size <math>n</math> from <math>\sP</math>.
I denote by <math>\sQ</math> the joint distribution whose features the researcher wants to learn.
If <math>\sQ</math> were identified given the observed data (e.g., if it were a marginal of <math>\sP</math>), point identification of the parameter or functional of interest would attain.
I denote by <math>\sR</math> the joint distribution of all variables, observable and unobservable ones; both <math>\sP</math> and <math>\sQ</math> can be obtained from it.
In the context of structural models, I denote by <math>\sM</math> a distribution for the observable variables that is consistent with the model.
I note that model incompleteness typically implies that <math>\sM</math> is not unique.
I let <math>\idr{\cdot}</math> denote the sharp identification region of the functional in square brackets, and <math>\outr{\cdot}</math> an outer region.
In both cases, the regions are indexed by <math>\sP</math>, because they depend on the distribution of the observed data.
==General references==
{{cite arXiv|last1=Molinari|first1=Francesca|year=2020|title=Microeconometrics with Partial Identification|eprint=2004.11751|class=econ.EM}}
==Notes==
{{Reflist|group=Notes}}
==References==
{{Reflist}}

Latest revision as of 04:16, 31 May 2024

[math] \newcommand{\edis}{\stackrel{d}{=}} \newcommand{\fd}{\stackrel{f.d.}{\rightarrow}} \newcommand{\dom}{\operatorname{dom}} \newcommand{\eig}{\operatorname{eig}} \newcommand{\epi}{\operatorname{epi}} \newcommand{\lev}{\operatorname{lev}} \newcommand{\card}{\operatorname{card}} \newcommand{\comment}{\textcolor{Green}} \newcommand{\B}{\mathbb{B}} \newcommand{\C}{\mathbb{C}} \newcommand{\G}{\mathbb{G}} \newcommand{\M}{\mathbb{M}} \newcommand{\N}{\mathbb{N}} \newcommand{\Q}{\mathbb{Q}} \newcommand{\T}{\mathbb{T}} \newcommand{\R}{\mathbb{R}} \newcommand{\E}{\mathbb{E}} \newcommand{\W}{\mathbb{W}} \newcommand{\bU}{\mathfrak{U}} \newcommand{\bu}{\mathfrak{u}} \newcommand{\bI}{\mathfrak{I}} \newcommand{\cA}{\mathcal{A}} \newcommand{\cB}{\mathcal{B}} \newcommand{\cC}{\mathcal{C}} \newcommand{\cD}{\mathcal{D}} \newcommand{\cE}{\mathcal{E}} \newcommand{\cF}{\mathcal{F}} \newcommand{\cG}{\mathcal{G}} \newcommand{\cg}{\mathcal{g}} \newcommand{\cH}{\mathcal{H}} \newcommand{\cI}{\mathcal{I}} \newcommand{\cJ}{\mathcal{J}} \newcommand{\cK}{\mathcal{K}} \newcommand{\cL}{\mathcal{L}} \newcommand{\cM}{\mathcal{M}} \newcommand{\cN}{\mathcal{N}} \newcommand{\cO}{\mathcal{O}} \newcommand{\cP}{\mathcal{P}} \newcommand{\cQ}{\mathcal{Q}} \newcommand{\cR}{\mathcal{R}} \newcommand{\cS}{\mathcal{S}} \newcommand{\cT}{\mathcal{T}} \newcommand{\cU}{\mathcal{U}} \newcommand{\cu}{\mathcal{u}} \newcommand{\cV}{\mathcal{V}} \newcommand{\cW}{\mathcal{W}} \newcommand{\cX}{\mathcal{X}} \newcommand{\cY}{\mathcal{Y}} \newcommand{\cZ}{\mathcal{Z}} \newcommand{\sF}{\mathsf{F}} \newcommand{\sM}{\mathsf{M}} \newcommand{\sG}{\mathsf{G}} \newcommand{\sT}{\mathsf{T}} \newcommand{\sB}{\mathsf{B}} \newcommand{\sC}{\mathsf{C}} \newcommand{\sP}{\mathsf{P}} \newcommand{\sQ}{\mathsf{Q}} \newcommand{\sq}{\mathsf{q}} \newcommand{\sR}{\mathsf{R}} \newcommand{\sS}{\mathsf{S}} \newcommand{\sd}{\mathsf{d}} \newcommand{\cp}{\mathsf{p}} \newcommand{\cc}{\mathsf{c}} \newcommand{\cf}{\mathsf{f}} \newcommand{\eU}{{\boldsymbol{U}}} \newcommand{\eb}{{\boldsymbol{b}}} \newcommand{\ed}{{\boldsymbol{d}}} \newcommand{\eu}{{\boldsymbol{u}}} \newcommand{\ew}{{\boldsymbol{w}}} \newcommand{\ep}{{\boldsymbol{p}}} \newcommand{\eX}{{\boldsymbol{X}}} \newcommand{\ex}{{\boldsymbol{x}}} \newcommand{\eY}{{\boldsymbol{Y}}} \newcommand{\eB}{{\boldsymbol{B}}} \newcommand{\eC}{{\boldsymbol{C}}} \newcommand{\eD}{{\boldsymbol{D}}} \newcommand{\eW}{{\boldsymbol{W}}} \newcommand{\eR}{{\boldsymbol{R}}} \newcommand{\eQ}{{\boldsymbol{Q}}} \newcommand{\eS}{{\boldsymbol{S}}} \newcommand{\eT}{{\boldsymbol{T}}} \newcommand{\eA}{{\boldsymbol{A}}} \newcommand{\eH}{{\boldsymbol{H}}} \newcommand{\ea}{{\boldsymbol{a}}} \newcommand{\ey}{{\boldsymbol{y}}} \newcommand{\eZ}{{\boldsymbol{Z}}} \newcommand{\eG}{{\boldsymbol{G}}} \newcommand{\ez}{{\boldsymbol{z}}} \newcommand{\es}{{\boldsymbol{s}}} \newcommand{\et}{{\boldsymbol{t}}} \newcommand{\ev}{{\boldsymbol{v}}} \newcommand{\ee}{{\boldsymbol{e}}} \newcommand{\eq}{{\boldsymbol{q}}} \newcommand{\bnu}{{\boldsymbol{\nu}}} \newcommand{\barX}{\overline{\eX}} \newcommand{\eps}{\varepsilon} \newcommand{\Eps}{\mathcal{E}} \newcommand{\carrier}{{\mathfrak{X}}} \newcommand{\Ball}{{\mathbb{B}}^{d}} \newcommand{\Sphere}{{\mathbb{S}}^{d-1}} \newcommand{\salg}{\mathfrak{F}} \newcommand{\ssalg}{\mathfrak{B}} \newcommand{\one}{\mathbf{1}} \newcommand{\Prob}[1]{\P\{#1\}} \newcommand{\yL}{\ey_{\mathrm{L}}} \newcommand{\yU}{\ey_{\mathrm{U}}} \newcommand{\yLi}{\ey_{\mathrm{L}i}} \newcommand{\yUi}{\ey_{\mathrm{U}i}} \newcommand{\xL}{\ex_{\mathrm{L}}} \newcommand{\xU}{\ex_{\mathrm{U}}} \newcommand{\vL}{\ev_{\mathrm{L}}} \newcommand{\vU}{\ev_{\mathrm{U}}} \newcommand{\dist}{\mathbf{d}} \newcommand{\rhoH}{\dist_{\mathrm{H}}} \newcommand{\ti}{\to\infty} \newcommand{\comp}[1]{#1^\mathrm{c}} \newcommand{\ThetaI}{\Theta_{\mathrm{I}}} \newcommand{\crit}{q} \newcommand{\CS}{CS_n} \newcommand{\CI}{CI_n} \newcommand{\cv}[1]{\hat{c}_{n,1-\alpha}(#1)} \newcommand{\idr}[1]{\mathcal{H}_\sP[#1]} \newcommand{\outr}[1]{\mathcal{O}_\sP[#1]} \newcommand{\idrn}[1]{\hat{\mathcal{H}}_{\sP_n}[#1]} \newcommand{\outrn}[1]{\mathcal{O}_{\sP_n}[#1]} \newcommand{\email}[1]{\texttt{#1}} \newcommand{\possessivecite}[1]{\ltref name="#1"\gt\lt/ref\gt's \citeyear{#1}} \newcommand\xqed[1]{% \leavevmode\unskip\penalty9999 \hbox{}\nobreak\hfill \quad\hbox{#1}} \newcommand\qedex{\xqed{$\triangle$}} \newcommand\independent{\perp\!\!\!\perp} \DeclareMathOperator{\Int}{Int} \DeclareMathOperator{\conv}{conv} \DeclareMathOperator{\cov}{Cov} \DeclareMathOperator{\var}{Var} \DeclareMathOperator{\Sel}{Sel} \DeclareMathOperator{\Bel}{Bel} \DeclareMathOperator{\cl}{cl} \DeclareMathOperator{\sgn}{sgn} \DeclareMathOperator{\essinf}{essinf} \DeclareMathOperator{\esssup}{esssup} \newcommand{\mathds}{\mathbb} \renewcommand{\P}{\mathbb{P}} [/math]

Why Partial Identification?

Knowing the population distribution that data are drawn from, what can one learn about a parameter of interest? It has long been understood that assumptions about the data generating process (DGP) play a crucial role in answering this identification question at the core of all empirical research. Inevitably, assumptions brought to bear enjoy a varying degree of credibility. Some are rooted in economic theory (e.g., optimizing behavior) or in information available to the researcher on the DGP (e.g., randomization mechanisms). These assumptions can be argued to be highly credible. Others are driven by concerns for tractability and the desire to answer the identification question with a certain level of precision (e.g., functional form and distributional assumptions). These are arguably less credible. Early on, [1] highlighted the importance of imposing restrictions based on prior knowledge of the phenomenon under analysis and some criteria of simplicity, but not for the purpose of identifiability of a parameter that the researcher happens to be interested in, stating (p.169):

One might regard problems of identifiability as a necessary part of the specification problem. We would consider such a classification acceptable, provided the temptation to specify models in such a way as to produce identifiability of relevant characteristics is resisted.

Much work, spanning multiple fields, has been devoted to putting forward strategies to carry out empirical research while relaxing distributional, functional form, or behavioral assumptions. One example, embodied in the research program on semiparameteric and nonparametric methods, is to characterize sufficient sets of assumptions, that exclude many suspect ones --sometimes as many as possible-- to guarantee that point identification of specific economically interesting parameters attains. This literature is reviewed in, e.g., [2][3], and is not discussed here.

Another example, embodied in the research program on Bayesian model uncertainty, is to specify multiple models (i.e., multiple sets of assumptions), put a prior on the parameters of each model and on each model, embed the various separate models within one large hierarchical mixture model, and obtain model posterior probabilities which can be used for a variety of inferences and decisions. This literature is reviewed in, e.g., [4] and [5], and is not discussed here.

The approach considered here fixes a set of assumptions and a parameter of interest a priori, in the spirit of [1], and asks what can be learned about that parameter given the available data, recognizing that even partial information can be illuminating for empirical research, while enjoying wider credibility thanks to the weaker assumptions imposed. The bounding methods at the core of this approach appeared in the literature nearly a century ago. Arguably, the first exemplar that leverages economic reasoning is given by the work of [6]. They provided bounds on Cobb-Douglas production functions in models of supply and demand, building on optimization principles and restrictions from microeconomic theory. [7] revisited their analysis to obtain bounds on the elasticities of demand and supply in a linear simultaneous equations system with uncorrelated errors. The first exemplars that do not rely on specific economic models appear in [8], [9], and [10], who bounded the coefficient of a simple linear regression in the presence of measurement error. These results were extended to the general linear regression model with errors in all variables by [11] and [12].

This chapter surveys some of the methods proposed over the last thirty years in the microeconometrics literature to further this approach. These methods belong to the systematic program on ’'partial identification analysis started with [13][14][15][16][17][18] and developed by several authors since the early 1990s. Within this program, the focus shifts from points to sets: the researcher aims to learn what is the set of values for the parameters of interest that can generate the same distribution of observables as the one in the data, for some DGP consistent with the maintained assumptions. In other words, the focus is on the set of observationally equivalent values, which henceforth I refer to as the parameters' sharp identification region. In the partial identification paradigm, empirical analysis begins with characterizing this set using the data alone. This is a nonparametric approach that dispenses with all assumptions, except basic restrictions on the sampling process such that the distribution of the observable variables can be learned as data accumulate. In subsequent steps, one incorporates additional assumptions into the analysis, reporting how each assumption (or set of assumptions) affects what one can learn about the parameters of interest, i.e., how it modifies and possibly shrinks the sharp identification region. Point identification may result from the process of increasingly strengthening the maintained assumptions, but it is not the goal in itself. Rather, the objective is to make transparent the relative role played by the data and the assumptions in shaping the inference that one draws.

There are several strands of independent, but thematically related literatures that are not discussed in this chapter. As a consequence, many relevant contributions are left out of the presentation and the references. One example is the literature in finance. [19] developed nonparametric bounds for the admissible set for means and standard deviations of intertemporal marginal rates of substitution (IMRS) of consumers. The bounds were developed exploiting the condition, satisfied in many finance models, that the equilibrium price of any traded security equals the expectation (conditioned on current information) of the product's future payoff and the IMRS of any consumer.[Notes 1] [20] extended the analysis to economies with frictions. [21] developed econometric tools to estimate the regions, to assess asset pricing models, and to provide nonparametric characterizations of asset pricing anomalies. Earlier on, the existence of volatility bounds on IMRSs were noted by [22] and [23]. The bounding arguments that build on the minimum-volatility frontier for stochastic discount factors proposed by [19] have become a litmus test to detect anomalies in asset pricing models (see, e.g. [24](p. 89)). I refer to the textbook presentations in [25](Chapter 13) and [26](Chapters 5 and 21), and the review articles by [27] and [28], for a careful presentation of this literature.

In macroeconomics, [29], [30], and [31] proposed bounds for impulse response functions in sign-restricted structural vector autoregression models, and carried out Bayesian inference with a non-informative prior for the non-identified parameters. I refer to [32](Chapter 13) for a careful presentation of this literature.

In microeconomic theory, bounds were derived from inequalities resulting as necessary and sufficient conditions that data on an individual's choice need to satisfy in order to be consistent with optimizing behavior, as in the research pioneered by [33] and advanced early on by [34] and [35]. [36] and [37] extended this research program to revealed preference extrapolation. Notably, in this work no stochastic terms enter the analysis. [38], [39], [40], [41], [42], and [43], extended revealed preference arguments to random utility models, and obtained bounds on the distributions of preferences. I refer to the survey articles by [44] and [45](Chapter XXX in this Volume) for a careful presentation of this literature.

A complementary approach to partial identification is given by sensitivity analysis, advocated for in different ways by, e.g., [46], [47], [48], [49], [50], and others. Within this approach, the analysis begins with a fully parametric model that point identifies the parameter of interest. One then reports the set of values for this parameter that result when the more suspicious assumptions are relaxed.

Related literatures, not discussed in this chapter, abound also outside Economics. For example, in probability theory, [51] and [52] put forward bounds on the joint distributions of random variables, and [53], [54], and [55] on the sum of random variables, when only marginal distributions are observed. The literature on probability bounds is discussed in the textbook by [56](Appendix A). Addressing problems faced in economics, sociology, epidemiology, geography, history, political science, and more, [57] derived bounds on correlations among variables measured at the individual level based on observable correlations among variables measured at the aggregate level. The so called ecological inference problem they studied, and the associated literature, is discussed in the survey article by [58] and references therein.

Goals and Structure of this Chapter

To carry out econometric analysis with partial identification, one needs: (1) computationally feasible characterizations of the parameters' sharp identification region; (2) methods to estimate this region; and (3) methods to test hypotheses and construct confidence sets. The goal of this chapter is to provide insights into the challenges posed by each of these desiderata, and into some of their solutions. In order to discuss the partial identification literature in microeconometrics with some level of detail while keeping this chapter to a manageable length, I focus on a selection of papers and not on a complete survey of the literature. As a consequence, many relevant contributions are left out of the presentation and the references. I also do not discuss the important but separate topic of statistical decisions in the presence of partial identification, for which I refer to the textbook treatments in [59][17] and to the review by [60](Chapter XXX in this Volume).

The presumption in identification analysis that the distribution from which the data are drawn is known allows one to keep separate the identification question from the distinct question of statistical inference from a finite sample. I use the same separation in this chapter. I assume solid knowledge of the topics covered in first year Economics PhD courses in econometrics and microeconomic theory.

I begin in Section with the analysis of what can be learned about features of probability distributions that are well defined in the absence of an economic model, such as moments, quantiles, cumulative distribution functions, etc., when one faces measurement problems. Specifically, I focus on cases where the data is incomplete, either due to selectively observed data or to interval measurements. I refer to [15][16][17] for textbook treatments of many other cases. I lay out formally the maintained assumptions for several examples, and then discuss in detail what is the source of the identification problem. I conclude with providing tractable characterizations of what can be learned about the parameters of interest, with formal proofs. I show that even in simple problems, great care may be needed to obtain the sharp identification region. It is often easier to characterize an outer region, i.e., a collection of values for the parameter of interest that contains the sharp one but may contain also additional values. Outer regions are useful because of their simplicity and because in certain applications they may suffice to answer questions of great interest, e.g., whether a policy intervention has a nonnegative effect. However, compared to the sharp identification region they may afford the researcher less useful predictions, and a lower ability to test for misspecification, because they do not harness all the information in the observed data and maintained assumptions.

In Section I use the same approach to study what can be learned about features of parameters of structural econometric models when the model is incomplete [61][62][63]. Specifically, I discuss single agent discrete choice models under a variety of challenging situations (interval measured as well as endogenous explanatory variables; unobserved as well as counterfactual choice sets); finite discrete games with multiple equilibria; auction models under weak assumptions on bidding behavior; and network formation models. Again I formally derive sharp identification regions for several examples.

I conclude each of these sections with a brief discussion of further theoretical advances and empirical applications that is meant to give a sense of the breadth of the approach, but not to be exhaustive. I refer to the recent survey by [64] for a thorough discussion of empirical applications of partial identification methods.

In Section I discuss finite sample inference. I limit myself to highlighting the challenges that one faces for consistent estimation when the identified object is a set, and several coverage notions and requirements that have been proposed over the last 20 years. I refer to the recent survey by [65] for a thorough discussion of methods to tests hypotheses and build confidence sets in moment inequality models.

In Section I discuss the distinction between refutable and non-refutable assumptions, and how model misspecification may be detectable in the presence of the former, even within the partial identification paradigm. I then highlight certain challenges that model misspecification presents for the interpretation of sharp identification (as well as outer) regions, and for the construction of confidence sets.

In Section I highlight that while most of the sharp identification regions characterized in Section can be easily computed, many of the ones in Section are more challenging. This is because the latter are obtained as level sets of criterion functions in moderately dimensional spaces, and tracing out these level sets or their boundaries is a non-trivial computational problem. In Section I conclude providing some considerations on what I view as open questions for future research. I refer to [66] for an earlier review of this literature, and to [67] for a careful presentation of the many notions of identification that are used across the econometrics literature, including an important historical account of how these notions developed over time.

Random Set Theory as a Tool for Partial Identification Analysis

Throughout Sections and Section, a simple organizing principle for much of partial identification analysis emerges. The cause of the identification problems discussed can be traced back to a collection of random variables that are consistent with the available data and maintained assumptions. For the problems studied in Section, this set is often a simple function of the observed variables. The incompleteness of the data stems from the fact that instead of observing the singleton variables of interest, one observes set-valued variables to which these belong, but one has no information on their exact value within the sets. For the problems studied in Section, the collection of random variables consistent with the maintained assumptions comprises what the model predicts for the endogenous variable(s). The incompleteness of the model stems from the fact that instead of making a singleton prediction for the variable(s) of interest, the model makes multiple predictions but does not specify how one is chosen.

The central role of set-valued objects, both stochastic and nonstochastic, in partial identification renders random set theory a natural toolkit to aid the analysis.[Notes 2] This theory originates in the seminal contributions of [68], [69], and [70], with the first self contained treatment of the theory given by [71]. I refer to [72] for a textbook presentation, and to [73][74] for a treatment focusing on its applications in econometrics.

[75] introduce the use of random set theory in econometrics to carry out identification analysis and statistical inference with incomplete data. [76][77] propose it to characterize sharp identification regions both with incomplete data and with incomplete models. [78] propose the use of optimal transportation methods that in some applications deliver the same characterizations as the random set methods. I do not discuss optimal transportation methods in this chapter, but refer to [79] for a thorough treatment.

Over the last ten years, random set methods have been used to unify a number of specific results in partial identification, and to produce a general methodology for identification analysis that dispenses completely with case-by-case distinctions. In particular, as I show throughout the chapter, the methods allow for simple and tractable characterizations of sharp identification regions. The collection of these results establishes that indeed this is a useful tool to carry out econometrics with partial identification, as exemplified by its prominent role both in this chapter and in Chapter XXX in this Volume by [80], which focuses on general classes of instrumental variable models. The random sets approach complements the more traditional one, based on mathematical tools for (single valued) random vectors, that proved extremely productive since the beginning of the research program in partial identification.

This chapter shows that to fruitfully apply random set theory for identification and inference, the econometrician needs to carry out three fundamental steps. First, she needs to define the random closed set that is relevant for the problem under consideration using all information given by the available data and maintained assumptions. This is a delicate task, but one that is typically carried out in identification analysis regardless of whether random set theory is applied. Indeed, throughout the chapter I highlight how relevant random closed sets were characterized in partial identification analysis since the early 1990s, albeit the connection to the theory of random sets was not made. As a second step, the econometrician needs to determine how the observable random variables relate to the random closed set. Often, one of two cases occurs: either the observable variables determine a random set to which the unobservable variable of interest belongs with probability one, as in incomplete data scenarios; or the (expectation of the) (un)observable variable belongs to (the expectation of) a random set determined by the model, as in incomplete model scenarios. Finally, the econometrician needs to determine which tool from random set theory should be utilized. To date, new applications of random set theory to econometrics have fruitfully exploited (Aumann) expectations and their support functions, (Choquet) capacity functionals, and laws of large numbers and central limit theorems for random sets. Appendix reports basic definitions from random set theory of these concepts, as well as some useful theorems. The chapter explains in detail through applications to important identification problems how these steps can be carried out.

Notation Used
[math](\Omega,\salg,\P)[/math] Nonatomic probability space
[math]\R^d,\|\cdot\|[/math] Euclidean space equipped with the Euclidean norm
[math]\cF,\cG,\cK[/math] Collection of closed, open, and compact subsets of [math]\R^d[/math] (respectively)
[math]\Sphere = \{x \in \R^d: \|x\| = 1\}[/math] Unit sphere in [math]\R^d[/math]
[math]\Ball = \{x \in \R^d: \|x\| \leq 1\}[/math] Unit ball in [math]\R^d[/math]
[math]\conv(A),\cl(A),|B|[/math] Convex hull and closure of a set [math]A\subset\R^d[/math] (respectively), and cardinality of a finite set [math]B\subset\R^d[/math]
[math]\ex,\ey,\ez,\dots[/math] Random vectors
[math]x,y,z,\dots[/math] Realizations of random vectors or deterministic vectors
[math]\eX,\eY,\eZ,\dots[/math] Random sets
[math]X,Y,Z,\dots[/math] Realizations of random sets or deterministic sets
[math]\epsilon,\eps,\nu,\zeta[/math] Unobserved random variables (heterogeneity)
[math]\Theta,\theta,\vartheta[/math] Parameter space, data generating value for the parameter vector, and a generic element of [math]\Theta[/math]
[math]\sR[/math] Joint distribution of all variables (observable and unobservable)
[math]\sP[/math] Joint distribution of the observable variables
[math]\sQ[/math] Joint distribution whose features one wants to learn
[math]\sM[/math] A joint distribution of observed variables implied by the model
[math]\sq_{\tau}(\alpha)[/math] Quantile function at level [math]\alpha \in (0,1)[/math] for a random variable distributed [math]\tau\in\{\sR,\sP,\sQ\}[/math]
[math]\E_\tau[/math] Expectation operator associated with distribution [math]\tau\in\{\sR,\sP,\sQ\}[/math]
[math]\sT_\eX(K)=\Prob{\eX\cap K\neq\emptyset},\, K\in\cK[/math] Capacity functional of random set [math]\eX[/math]
[math]\sC_\eX(F)=\Prob{\eX\subset F},\, F\in\cF[/math] Containment functional of random set [math]\eX[/math]
[math]\stackrel{p}{\rightarrow},\stackrel{\text{a.s.}}{\rightarrow},\Rightarrow[/math] Convergence in probability, convergence almost surely, and weak convergence (respectively)
[math]\ex\edis\ey[/math] [math]\ex[/math] and [math]\ey[/math] have the same distribution
[math]\ex\independent\ey[/math] Statistical independence between random variables [math]\ex[/math] and [math]\ey[/math]
[math]x^\top y[/math] Inner product between vectors [math]x[/math] and [math]y[/math], [math]x,y\in\R^d[/math]
[math]\bU,\bu[/math] Family of utility functions and one of its elements
[math]\crit_\sP[/math] Criterion function that aggregates violations of the population moment inequalities
[math]\crit_n[/math] Criterion function that aggregates violations of the sample moment inequalities
[math]\idr{\cdot}[/math] Sharp identification region of the functional in square brackets (a function of [math]\sP[/math])
[math]\outr{\cdot}[/math] An outer region of the functional in square brackets (a function of [math]\sP[/math])

Notation

This chapter employs consistent notation that is summarized in Table. Some important conventions are as follows: [math]\ey[/math] denotes outcome variables, [math](\ex,\ew)[/math] denote explanatory variables, and [math]\ez[/math] denotes instrumental variables (i.e., variables that satisfy some form of independence with the outcome or with the unobservable variables, possibly conditional on [math]\ex,\ew[/math]). I denote by [math]\sP[/math] the joint distribution of all observable variables. Identification analysis is carried out using the information contained in this distribution, and finite sample inference is carried out under the presumption that one draws a random sample of size [math]n[/math] from [math]\sP[/math]. I denote by [math]\sQ[/math] the joint distribution whose features the researcher wants to learn. If [math]\sQ[/math] were identified given the observed data (e.g., if it were a marginal of [math]\sP[/math]), point identification of the parameter or functional of interest would attain. I denote by [math]\sR[/math] the joint distribution of all variables, observable and unobservable ones; both [math]\sP[/math] and [math]\sQ[/math] can be obtained from it. In the context of structural models, I denote by [math]\sM[/math] a distribution for the observable variables that is consistent with the model. I note that model incompleteness typically implies that [math]\sM[/math] is not unique. I let [math]\idr{\cdot}[/math] denote the sharp identification region of the functional in square brackets, and [math]\outr{\cdot}[/math] an outer region. In both cases, the regions are indexed by [math]\sP[/math], because they depend on the distribution of the observed data.

General references

Molinari, Francesca (2020). "Microeconometrics with Partial Identification". arXiv:2004.11751 [econ.EM].

Notes

  1. [1] deduce a duality relation with the mean variance theory of [2] and [3], but the relation does not apply to the sharp bounds they derive. In the Arbitrage Pricing Model [4], bounds on extensions of existing pricing functions, consistent with the absence of arbitrage opportunities, were considered by [5] and [6].
  2. The first idea of a general random set in the form of a region that depends on chance appears in [7], originally published in 1933. For another early example where confidence regions are explicitly described as random sets, see [8](p. 67). The role of random sets in this chapter is different.

References

  1. 1.0 1.1 Koopmans, T.C., and O.Reiersol (1950): “The Identification of Structural Characteristics” The Annals of Mathematical Statistics, 21(2), 165--181.
  2. Matzkin, R.L. (2007): “Chapter 73 -- Nonparametric identification” in Handbook of Econometrics, ed. by J.J. Heckman, and E.E. Leamer, vol.6, chap.73, pp. 5307 -- 5368. Elsevier.
  3. Matzkin, R.L. (2013): “Nonparametric Identification in Structural Economic Models” Annual Review of Economics, 5(1), 457--486.
  4. Wasserman, L. (2000): “Bayesian Model Selection and Model Averaging” Journal of Mathematical Psychology, 44(1), 92 -- 107.
  5. Clyde, M., and E.I. George (2004): “Model Uncertainty” Statist. Sci., 19(1), 81--94.
  6. Marschak, J., and W.H. Andrews (1944): “Random Simultaneous Equations and the Theory of Production” Econometrica, 12(3/4), 143--205.
  7. Leamer, E.E. (1981): “Is it a Demand Curve, Or Is It A Supply Curve? Partial Identification through Inequality Constraints” The Review of Economics and Statistics, 63(3), 319--327.
  8. Gini, C. (1921): “Sull'interpolazione di una retta quando i valori della variabile indipendente sono affetti da errori accidentali” Metroeconomica, 1(3), 63–82.
  9. Frisch, R. (1934): Statistical Confluence Analysis by Means of Complete Regression Systems, Okonomiske Institutt Oslo: Publikasjon. Universitetets {\O}konomiske Instituut.
  10. Reiersol, O. (1941): “Confluence Analysis by Means of Lag Moments and Other Methods of Confluence Analysis” Econometrica, 9(1), 1--24.
  11. Klepper, S., and E.E. Leamer (1984): “Consistent Sets of Estimates for Regressions with Errors in All Variables” Econometrica, 52(1), 163--183.
  12. Leamer, E.E. (1987): “Errors in Variables in Linear Systems” Econometrica, 55(4), 893--909.
  13. Manski, C.F. (1989): “Anatomy of the Selection Problem” The Journal of Human Resources, 24(3), 343--360.
  14. Manski, C.F. (1990): “Nonparametric Bounds on Treatment Effects” The American Economic Review Papers and Proceedings, 80(2), 319--323.
  15. 15.0 15.1 Manski, C.F. (1995): Identification Problems in the Social Sciences. Harvard University Press.
  16. 16.0 16.1 Manski, C.F. (2003): Partial Identification of Probability Distributions, Springer Series in Statistics. Springer.
  17. 17.0 17.1 17.2 Manski, C.F. (2007a): Identification for Prediction and Decision. Harvard University Press.
  18. Manski, C.F. (2013b): Public Policy in an Uncertain World: Analysis and Decisions. Harvard University Press.
  19. 19.0 19.1 Hansen, L.P., and R.Jagannathan (1991): “Implications of Security Market Data for Models of Dynamic Economies” Journal of Political Economy, 99(2), 225--262.
  20. Luttmer, E. G.J. (1996): “Asset Pricing in Economies with Frictions” Econometrica, 64(6), 1439--1467.
  21. Hansen, L.P., J.Heaton, and E.G.J. Luttmer (1995): “Econometric Evaluation of Asset Pricing Models” The Review of Financial Studies, 8(2), 237--274.
  22. Shiller, R.J. (1982): “Consumption, asset markets and macroeconomic fluctuations” Carnegie-Rochester Conference Series on Public Policy, 17, 203 -- 238.
  23. Hansen, L.P. (1982a): “Consumption, asset markets, and macroeconomic fluctuations: A comment” Carnegie-Rochester Conference Series on Public Policy, 17, 239 -- 250.
  24. Shiller, R.J. (2003): “From Efficient Markets Theory to Behavioral Finance” Journal of Economic Perspectives, 17(1), 83--104.
  25. Ljungqvist, L., and T.Sargent (2004): Recursive Macroeconomic Theory, vol.1. The MIT Press, 2 edn.
  26. Cochrane, J.H. (2005): Asset Pricing. Princeton University Press, Princeton, New Jersey, second edn.
  27. Ferson, W.E. (2003): “Chapter 12 -- Tests of multifactor pricing models, volatility bounds and portfolio performance” in Financial Markets and Asset Pricing, vol.1 of Handbook of the Economics of Finance, pp. 743 -- 802. Elsevier.
  28. Campbell, J.Y. (2014): “Empirical Asset Pricing: Eugene Fama, Lars Peter Hansen, and Robert Shiller” The Scandinavian Journal of Economics, 116(3), 593--634.
  29. Faust, J. (1998): “The robustness of identified VAR conclusions about money” Carnegie-Rochester Conference Series on Public Policy, 49, 207 -- 244.
  30. Canova, F., and G.{De Nicolo} (2002): “Monetary disturbances matter for business fluctuations in the G-7” Journal of Monetary Economics, 49(6), 1131 -- 1159.
  31. Uhlig, H. (2005): “What are the effects of monetary policy on output? Results from an agnostic identification procedure” Journal of Monetary Economics, 52(2), 381 -- 419.
  32. Kilian, L., and H.L\"{u}tkepohl (2017): Structural Vector Autoregressive Analysis, Themes in Modern Econometrics. Cambridge University Press.
  33. Samuelson, P.A. (1938): “A Note on the Pure Theory of Consumer's Behaviour” Economica, 5(17), 61--71.
  34. Houthakker, H.S. (1950): “Revealed Preference and the Utility Function” Economica, 17(66), 159--174.
  35. Richter, M.K. (1966): “Revealed Preference Theory” Econometrica, 34(3), 635--645.
  36. Afriat, S.N. (1967): “The Construction of Utility Functions from Expenditure Data” International Economic Review, 8(1), 67--77.
  37. Varian, H.R. (1982): “The Nonparametric Approach to Demand Analysis” Econometrica, 50(4), 945--973.
  38. Block, H.D., and J.Marschak (1960): “Random Orderings and Stochastic Theories of Responses” in Contributions to Probability and Statistics: Essays in Honor of Harold Hotelling, ed. by I.Olkin, pp. 97--132. Stanford University Press.
  39. Marschak, J. (1960): “Binary Choice Constraints on Random Utility Indicators” in Stanford Symposium on Mathematical Methods in the Social Sciences, ed. by K.Arrow. Stanford University Press.
  40. Hall, R.E. (1973): “On the statistical theory of unobserved components” MIT Working Paper 117, available at https://dspace.mit.edu/bitstream/handle/1721.1/63972/onstatisticalthe00hall.pdf?sequence=1.
  41. McFadden, D.L. (1975): “Tchebyscheff bounds for the space of agent characteristics” Journal of Mathematical Economics, 2(2), 225 -- 242.
  42. Falmagne, J. (1978): “A representation theorem for finite random scale systems” Journal of Mathematical Psychology, 18(1), 52 -- 72.
  43. McFadden, D.L., and M.K. Richter (1991): “Stochastic rationality and revealed stochastic preference” in Preferences, Uncertainty and Rationality, ed. by J.S. Chipman, D.L. McFadden, and M.K. Richter, pp. 161--186. Westview Press.
  44. Crawford, I., and B.DeRock (2014): “Empirical Revealed Preference” Annual Review of Economics, 6(1), 503--524.
  45. Blundell, R. (2019): “Revealed preference” in Handbook of Econometrics. Elsevier.
  46. Gilstein, C.Z., and E.E. Leamer (1983): “Robust Sets of Regression Estimates” Econometrica, 51(2), 321--333.
  47. Rosenbaum, P.R., and D.B. Rubin (1983): “Assessing Sensitivity to an Unobserved Binary Covariate in an Observational Study with Binary Outcome” Journal of the Royal Statistical Society. Series B (Methodological), 45(2), 212--218.
  48. Leamer, E.E. (1985): “Sensitivity Analyses Would Help” The American Economic Review, 75(3), 308--313.
  49. Rosenbaum, P.R. (1995): Observational Studies. Springer.
  50. Imbens, G.W. (2003): “Sensitivity to Exogeneity Assumptions in Program Evaluation” American Economic Review, 93(2), 126--132.
  51. Hoeffding, W. (1940): “Masstabinvariante Korrelationstheorie” Schriften des Matematischen Instituts und des Instituts f\"{u}r Angewandte MAthematik des Universit\"{a}t Berlin, 5(3), 179--233.
  52. Fréchet, M.R. (1951): “Sur les tableaux de correlation dont les marges sont donnes” Annales de I’Université de Lyon A, 3, 53--77.
  53. Makarov, G.D. (1981): “Estimates for the distribution function of a sum of two random variables when the marginal distributions are fixed” Theory of Probability and its Applications, 26(4), 803--806.
  54. R\"{u}schendorf, L. (1982): “Random Variables with Maximum Sums” Advances in Applied Probability, 14(3), 623--632.
  55. Frank, M.J., R.B. Nelsen, and B.Schweizer (1987): “Best-possible bounds for the distribution of a sum --- a problem of Kolmogorov” Probability Theory and Related Fields, 74(2), 199--211.
  56. Shorack, G.R., and J.A. Wellner (2009): Empirical Processes with Applications to Statistics. Society for Industrial and Applied Mathematics.
  57. Duncan, O.D., and B.Davis (1953): “An Alternative to Ecological Correlation” American Sociological Review, 18(6), 665--666.
  58. Cho, W.T., and C.F. Manski (2009): “Cross‐Level/Ecological Inference” in Oxford Handbook of Political Methodology, ed. by J.M. Box-Steffensmeier, H.E. Brady, and D.Collier, chap.24, pp. 530--569. Oxford University Press.
  59. Manski, C.F. (2005): Social Choice with Partial Knowledge of Treatment Response. Princeton University Press.
  60. Hirano, K., and J.R. Porter (2019): “Statistical Decision Rules in Econometrics” in Handbook of Econometrics. Elsevier.
  61. Tamer, E. (2003): “Incomplete Simultaneous Discrete Response Model with Multiple Equilibria” The Review of Economic Studies, 70(1), 147--165.
  62. Haile, P.A., and E.Tamer (2003): “Inference with an Incomplete Model of English Auctions” Journal of Political Economy, 111(1), 1--51.
  63. Ciliberto, F., and E.Tamer (2009): “Market Structure and Multiple Equilibria in Airline Markets” Econometrica, 77(6), 1791--1828.
  64. Ho, K., and A.M. Rosen (2017): “Partial Identification in Applied Research: Benefits and Challenges” in Advances in Economics and Econometrics: Eleventh World Congress, ed. by B.Honoré, A.Pakes, M.Piazzesi, and L.Samuelson, vol.1 of Econometric Society Monographs, pp. 307--359. Cambridge University Press.
  65. Canay, I.A., and A.M. Shaikh (2017): “Practical and Theoretical Advances in Inference for Partially Identified Models” in Advances in Economics and Econometrics: Eleventh World Congress, ed. by B.Honoré, A.Pakes, M.Piazzesi, and L.Samuelson, vol.2 of Econometric Society Monographs, p. 271–306. Cambridge University Press.
  66. Tamer, E. (2010): “Partial Identification in Econometrics” Annual Review of Economics, 2, 167--195.
  67. Lewbel, A. (2018): “The Identification Zoo - Meanings of Identification in Econometrics” Journal of Economic Literature, forthcoming.
  68. Choquet, G. (1953/54): “Theory of capacities” Annales de l'Institut Fourier (Grenoble), 5, 131--295.
  69. Aumann, R.J. (1965): “Integrals of set-valued functions” Journal of Mathematical Analysis and Applications, 12(1), 1--12.
  70. Debreu, G. (1967): “Integration of correspondences” in Proceedings of the Fifth Berkeley Symposium in Mathematical Statistics and Probability, vol.2, pp. 351--372. University of California Press.
  71. Matheron, G. (1975): Random Sets and Integral Geometry. Wiley, New York.
  72. Molchanov, I. (2017): Theory of Random Sets. Springer, London, 2 edn.
  73. Molchanov, I., and F.Molinari (2014): “Applications of Random Set Theory in Econometrics” Annual Review of Economics, 6(1), 229--251.
  74. Molchanov, I., and F.Molinari (2018): Random Sets in Econometrics. Econometric Society Monograph Series, Cambridge University Press, Cambridge UK.
  75. Beresteanu, A., and F.Molinari (2008): “Asymptotic Properties for a Class of Partially Identified Models” Econometrica, 76(4), 763--814.
  76. Beresteanu, A., I.Molchanov, and F.Molinari (2011): “Sharp identification regions in models with convex moment predictions” Econometrica, 79(6), 1785--1821.
  77. Beresteanu, A., I.Molchanov, and F.Molinari (2012): “Partial identification using random set theory” Journal of Econometrics, 166(1), 17 -- 32, with errata at https://molinari.economics.cornell.edu/docs/NOTE_BMM2012_v3.pdf.
  78. Galichon, A., and M.Henry (2011): “Set Identification in Models with Multiple Equilibria” The Review of Economic Studies, 78(4), 1264--1298.
  79. Galichon, A. (2016): Optimal Transport Methods in Economics. Princeton University Press.
  80. Chesher, A., and A.M. Rosen (2019): “Generalized instrumental variable models, methods, and applications” in Handbook of Econometrics. Elsevier.