id
int64 28.8k
36k
| category
stringclasses 3
values | text
stringlengths 44
3.03k
| title
stringlengths 10
236
| published
stringlengths 19
19
| author
stringlengths 6
943
| link
stringlengths 66
127
| primary_category
stringclasses 62
values |
---|---|---|---|---|---|---|---|
29,425 | em | This paper provides an introduction to structural estimation methods for
matching markets with transferable utility. | Structural Estimation of Matching Markets with Transferable Utility | 2021-09-16 15:29:14 | Alfred Galichon, Bernard Salanié | http://arxiv.org/abs/2109.07932v1, http://arxiv.org/pdf/2109.07932v1 | econ.EM |
29,133 | em | In this paper, we propose a model which simulates odds distributions of
pari-mutuel betting system under two hypotheses on the behavior of bettors: 1.
The amount of bets increases very rapidly as the deadline for betting comes
near. 2. Each bettor bets on a horse which gives the largest expectation value
of the benefit. The results can be interpreted as such efficient behaviors do
not serve to extinguish the FL bias but even produce stronger FL bias. | Efficiency in Micro-Behaviors and FL Bias | 2018-05-11 05:17:42 | Kurihara Kazutaka, Yohei Tutiya | http://arxiv.org/abs/1805.04225v1, http://arxiv.org/pdf/1805.04225v1 | econ.EM |
29,931 | em | This paper derives the efficiency bound for estimating the parameters of
dynamic panel data models in the presence of an increasing number of incidental
parameters. We study the efficiency problem by formulating the dynamic panel as
a simultaneous equations system, and show that the quasi-maximum likelihood
estimator (QMLE) applied to the system achieves the efficiency bound.
Comparison of QMLE with fixed effects estimators is made. | Efficiency of QMLE for dynamic panel data models with interactive effects | 2023-12-13 06:56:34 | Jushan Bai | http://arxiv.org/abs/2312.07881v1, http://arxiv.org/pdf/2312.07881v1 | econ.EM |
28,940 | em | Endogeneity and missing data are common issues in empirical research. We
investigate how both jointly affect inference on causal parameters.
Conventional methods to estimate the variance, which treat the imputed data as
if it was observed in the first place, are not reliable. We derive the
asymptotic variance and propose a heteroskedasticity robust variance estimator
for two-stage least squares which accounts for the imputation. Monte Carlo
simulations support our theoretical findings. | On the Effect of Imputation on the 2SLS Variance | 2019-03-26 19:42:59 | Helmut Farbmacher, Alexander Kann | http://arxiv.org/abs/1903.11004v1, http://arxiv.org/pdf/1903.11004v1 | econ.EM |
28,947 | em | We propose a model selection criterion to detect purely causal from purely
noncausal models in the framework of quantile autoregressions (QAR). We also
present asymptotics for the i.i.d. case with regularly varying distributed
innovations in QAR. This new modelling perspective is appealing for
investigating the presence of bubbles in economic and financial time series,
and is an alternative to approximate maximum likelihood methods. We illustrate
our analysis using hyperinflation episodes in Latin American countries. | Identification of Noncausal Models by Quantile Autoregressions | 2019-04-11 23:49:57 | Alain Hecq, Li Sun | http://arxiv.org/abs/1904.05952v1, http://arxiv.org/pdf/1904.05952v1 | econ.EM |
28,965 | em | Complex functions have multiple uses in various fields of study, so analyze
their characteristics it is of extensive interest to other sciences. This work
begins with a particular class of rational functions of a complex variable;
over this is deduced two elementals properties concerning the residues and is
proposed one results which establishes one lower bound for the p-norm of the
residues vector. Applications to the autoregressive processes are presented and
the exemplifications are indicated in historical data of electric generation
and econometric series. | On the residues vectors of a rational class of complex functions. Application to autoregressive processes | 2019-07-12 23:46:56 | Guillermo Daniel Scheidereiter, Omar Roberto Faure | http://arxiv.org/abs/1907.05949v1, http://arxiv.org/pdf/1907.05949v1 | econ.EM |
29,080 | em | Many econometric models can be analyzed as finite mixtures. We focus on
two-component mixtures and we show that they are nonparametrically point
identified by a combination of an exclusion restriction and tail restrictions.
Our identification analysis suggests simple closed-form estimators of the
component distributions and mixing proportions, as well as a specification
test. We derive their asymptotic properties using results on tail empirical
processes and we present a simulation study that documents their finite-sample
performance. | Inference on two component mixtures under tail restrictions | 2021-02-11 22:27:47 | Marc Henry, Koen Jochmans, Bernard Salanié | http://arxiv.org/abs/2102.06232v1, http://arxiv.org/pdf/2102.06232v1 | econ.EM |
29,093 | em | This paper establishes an extended representation theorem for unit-root VARs.
A specific algebraic technique is devised to recover stationarity from the
solution of the model in the form of a cointegrating transformation. Closed
forms of the results of interest are derived for integrated processes up to the
4-th order. An extension to higher-order processes turns out to be within the
reach on an induction argument. | Cointegrated Solutions of Unit-Root VARs: An Extended Representation Theorem | 2021-02-21 18:28:20 | Mario Faliva, Maria Grazia Zoia | http://arxiv.org/abs/2102.10626v1, http://arxiv.org/pdf/2102.10626v1 | econ.EM |
29,015 | em | The randomization inference literature studying randomized controlled trials
(RCTs) assumes that units' potential outcomes are deterministic. This
assumption is unlikely to hold, as stochastic shocks may take place during the
experiment. In this paper, we consider the case of an RCT with individual-level
treatment assignment, and we allow for individual-level and cluster-level (e.g.
village-level) shocks. We show that one can draw inference on the ATE
conditional on the realizations of the cluster-level shocks, using
heteroskedasticity-robust standard errors, or on the ATE netted out of those
shocks, using cluster-robust standard errors. | Clustering and External Validity in Randomized Controlled Trials | 2019-12-02 22:30:25 | Antoine Deeb, Clément de Chaisemartin | http://arxiv.org/abs/1912.01052v7, http://arxiv.org/pdf/1912.01052v7 | econ.EM |
28,777 | em | This article reviews recent advances in fixed effect estimation of panel data
models for long panels, where the number of time periods is relatively large.
We focus on semiparametric models with unobserved individual and time effects,
where the distribution of the outcome variable conditional on covariates and
unobserved effects is specified parametrically, while the distribution of the
unobserved effects is left unrestricted. Compared to existing reviews on long
panels (Arellano and Hahn 2007; a section in Arellano and Bonhomme 2011) we
discuss models with both individual and time effects, split-panel Jackknife
bias corrections, unbalanced panels, distribution and quantile effects, and
other extensions. Understanding and correcting the incidental parameter bias
caused by the estimation of many fixed effects is our main focus, and the
unifying theme is that the order of this bias is given by the simple formula
p/n for all models discussed, with p the number of estimated parameters and n
the total sample size. | Fixed Effect Estimation of Large T Panel Data Models | 2017-09-26 15:46:13 | Iván Fernández-Val, Martin Weidner | http://arxiv.org/abs/1709.08980v2, http://arxiv.org/pdf/1709.08980v2 | econ.EM |
28,778 | em | This paper considers the identification of treatment effects on conditional
transition probabilities. We show that even under random assignment only the
instantaneous average treatment effect is point identified. Since treated and
control units drop out at different rates, randomization only ensures the
comparability of treatment and controls at the time of randomization, so that
long-run average treatment effects are not point identified. Instead we derive
informative bounds on these average treatment effects. Our bounds do not impose
(semi)parametric restrictions, for example, proportional hazards. We also
explore various assumptions such as monotone treatment response, common shocks
and positively correlated outcomes that tighten the bounds. | Bounds On Treatment Effects On Transitions | 2017-09-26 15:46:40 | Johan Vikström, Geert Ridder, Martin Weidner | http://arxiv.org/abs/1709.08981v1, http://arxiv.org/pdf/1709.08981v1 | econ.EM |
28,779 | em | We propose an inference procedure for estimators defined by mathematical
programming problems, focusing on the important special cases of linear
programming (LP) and quadratic programming (QP). In these settings, the
coefficients in both the objective function and the constraints of the
mathematical programming problem may be estimated from data and hence involve
sampling error. Our inference approach exploits the characterization of the
solutions to these programming problems by complementarity conditions; by doing
so, we can transform the problem of doing inference on the solution of a
constrained optimization problem (a non-standard inference problem) into one
involving inference based on a set of inequalities with pre-estimated
coefficients, which is much better understood. We evaluate the performance of
our procedure in several Monte Carlo simulations and an empirical application
to the classic portfolio selection problem in finance. | Inference on Estimators defined by Mathematical Programming | 2017-09-26 19:24:52 | Yu-Wei Hsieh, Xiaoxia Shi, Matthew Shum | http://arxiv.org/abs/1709.09115v1, http://arxiv.org/pdf/1709.09115v1 | econ.EM |
28,780 | em | We analyze the empirical content of the Roy model, stripped down to its
essential features, namely sector specific unobserved heterogeneity and
self-selection on the basis of potential outcomes. We characterize sharp bounds
on the joint distribution of potential outcomes and testable implications of
the Roy self-selection model under an instrumental constraint on the joint
distribution of potential outcomes we call stochastically monotone instrumental
variable (SMIV). We show that testing the Roy model selection is equivalent to
testing stochastic monotonicity of observed outcomes relative to the
instrument. We apply our sharp bounds to the derivation of a measure of
departure from Roy self-selection to identify values of observable
characteristics that induce the most costly misallocation of talent and sector
and are therefore prime targets for intervention. Special emphasis is put on
the case of binary outcomes, which has received little attention in the
literature to date. For richer sets of outcomes, we emphasize the distinction
between pointwise sharp bounds and functional sharp bounds, and its importance,
when constructing sharp bounds on functional features, such as inequality
measures. We analyze a Roy model of college major choice in Canada and Germany
within this framework, and we take a new look at the under-representation of
women in~STEM. | Sharp bounds and testability of a Roy model of STEM major choices | 2017-09-27 02:25:35 | Ismael Mourifie, Marc Henry, Romuald Meango | http://arxiv.org/abs/1709.09284v2, http://arxiv.org/pdf/1709.09284v2 | econ.EM |
28,781 | em | The ongoing net neutrality debate has generated a lot of heated discussions
on whether or not monetary interactions should be regulated between content and
access providers. Among the several topics discussed, `differential pricing'
has recently received attention due to `zero-rating' platforms proposed by some
service providers. In the differential pricing scheme, Internet Service
Providers (ISPs) can exempt data access charges for on content from certain CPs
(zero-rated) while no exemption is on content from other CPs. This allows the
possibility for Content Providers (CPs) to make `sponsorship' agreements to
zero-rate their content and attract more user traffic. In this paper, we study
the effect of differential pricing on various players in the Internet. We first
consider a model with a monopolistic ISP and multiple CPs where users select
CPs based on the quality of service (QoS) and data access charges. We show that
in a differential pricing regime 1) a CP offering low QoS can make have higher
surplus than a CP offering better QoS through sponsorships. 2) Overall QoS
(mean delay) for end users can degrade under differential pricing schemes. In
the oligopolistic market with multiple ISPs, users tend to select the ISP with
lowest ISP resulting in same type of conclusions as in the monopolistic market.
We then study how differential pricing effects the revenue of ISPs. | Zero-rating of Content and its Effect on the Quality of Service in the Internet | 2017-09-27 07:51:32 | Manjesh K. Hanawal, Fehmina Malik, Yezekael Hayel | http://arxiv.org/abs/1709.09334v2, http://arxiv.org/pdf/1709.09334v2 | econ.EM |
28,782 | em | The uncertainty and robustness of Computable General Equilibrium models can
be assessed by conducting a Systematic Sensitivity Analysis. Different methods
have been used in the literature for SSA of CGE models such as Gaussian
Quadrature and Monte Carlo methods. This paper explores the use of Quasi-random
Monte Carlo methods based on the Halton and Sobol' sequences as means to
improve the efficiency over regular Monte Carlo SSA, thus reducing the
computational requirements of the SSA. The findings suggest that by using
low-discrepancy sequences, the number of simulations required by the regular MC
SSA methods can be notably reduced, hence lowering the computational time
required for SSA of CGE models. | Quasi-random Monte Carlo application in CGE systematic sensitivity analysis | 2017-09-28 01:54:30 | Theodoros Chatzivasileiadis | http://arxiv.org/abs/1709.09755v1, http://arxiv.org/pdf/1709.09755v1 | econ.EM |
28,783 | em | We propose a method of estimating the linear-in-means model of peer effects
in which the peer group, defined by a social network, is endogenous in the
outcome equation for peer effects. Endogeneity is due to unobservable
individual characteristics that influence both link formation in the network
and the outcome of interest. We propose two estimators of the peer effect
equation that control for the endogeneity of the social connections using a
control function approach. We leave the functional form of the control function
unspecified and treat it as unknown. To estimate the model, we use a sieve
semiparametric approach, and we establish asymptotics of the semiparametric
estimator. | Estimation of Peer Effects in Endogenous Social Networks: Control Function Approach | 2017-09-28 18:41:48 | Ida Johnsson, Hyungsik Roger Moon | http://arxiv.org/abs/1709.10024v3, http://arxiv.org/pdf/1709.10024v3 | econ.EM |
28,784 | em | This paper considers the problem of forecasting a collection of short time
series using cross sectional information in panel data. We construct point
predictors using Tweedie's formula for the posterior mean of heterogeneous
coefficients under a correlated random effects distribution. This formula
utilizes cross-sectional information to transform the unit-specific (quasi)
maximum likelihood estimator into an approximation of the posterior mean under
a prior distribution that equals the population distribution of the random
coefficients. We show that the risk of a predictor based on a non-parametric
estimate of the Tweedie correction is asymptotically equivalent to the risk of
a predictor that treats the correlated-random-effects distribution as known
(ratio-optimality). Our empirical Bayes predictor performs well compared to
various competitors in a Monte Carlo study. In an empirical application we use
the predictor to forecast revenues for a large panel of bank holding companies
and compare forecasts that condition on actual and severely adverse
macroeconomic conditions. | Forecasting with Dynamic Panel Data Models | 2017-09-29 01:46:48 | Laura Liu, Hyungsik Roger Moon, Frank Schorfheide | http://arxiv.org/abs/1709.10193v1, http://arxiv.org/pdf/1709.10193v1 | econ.EM |
28,785 | em | There is a fast growing literature that set-identifies structural vector
autoregressions (SVARs) by imposing sign restrictions on the responses of a
subset of the endogenous variables to a particular structural shock
(sign-restricted SVARs). Most methods that have been used to construct
pointwise coverage bands for impulse responses of sign-restricted SVARs are
justified only from a Bayesian perspective. This paper demonstrates how to
formulate the inference problem for sign-restricted SVARs within a
moment-inequality framework. In particular, it develops methods of constructing
confidence bands for impulse response functions of sign-restricted SVARs that
are valid from a frequentist perspective. The paper also provides a comparison
of frequentist and Bayesian coverage bands in the context of an empirical
application - the former can be substantially wider than the latter. | Inference for VARs Identified with Sign Restrictions | 2017-09-29 02:25:13 | Eleonora Granziera, Hyungsik Roger Moon, Frank Schorfheide | http://arxiv.org/abs/1709.10196v2, http://arxiv.org/pdf/1709.10196v2 | econ.EM |
28,786 | em | We systematically investigate the effect heterogeneity of job search
programmes for unemployed workers. To investigate possibly heterogeneous
employment effects, we combine non-experimental causal empirical models with
Lasso-type estimators. The empirical analyses are based on rich administrative
data from Swiss social security records. We find considerable heterogeneities
only during the first six months after the start of training. Consistent with
previous results of the literature, unemployed persons with fewer employment
opportunities profit more from participating in these programmes. Furthermore,
we also document heterogeneous employment effects by residence status. Finally,
we show the potential of easy-to-implement programme participation rules for
improving average employment effects of these active labour market programmes. | Heterogeneous Employment Effects of Job Search Programmes: A Machine Learning Approach | 2017-09-29 11:21:08 | Michael Knaus, Michael Lechner, Anthony Strittmatter | http://dx.doi.org/10.3368/jhr.57.2.0718-9615R1, http://arxiv.org/abs/1709.10279v2, http://arxiv.org/pdf/1709.10279v2 | econ.EM |
28,787 | em | Dynamic contracts with multiple agents is a classical decentralized
decision-making problem with asymmetric information. In this paper, we extend
the single-agent dynamic incentive contract model in continuous-time to a
multi-agent scheme in finite horizon and allow the terminal reward to be
dependent on the history of actions and incentives. We first derive a set of
sufficient conditions for the existence of optimal contracts in the most
general setting and conditions under which they form a Nash equilibrium. Then
we show that the principal's problem can be converted to solving
Hamilton-Jacobi-Bellman (HJB) equation requiring a static Nash equilibrium.
Finally, we provide a framework to solve this problem by solving partial
differential equations (PDE) derived from backward stochastic differential
equations (BSDE). | A Note on the Multi-Agent Contracts in Continuous Time | 2017-10-01 20:07:08 | Qi Luo, Romesh Saigal | http://arxiv.org/abs/1710.00377v2, http://arxiv.org/pdf/1710.00377v2 | econ.EM |
28,788 | em | This paper presents a new estimator of the intercept of a linear regression
model in cases where the outcome varaible is observed subject to a selection
rule. The intercept is often in this context of inherent interest; for example,
in a program evaluation context, the difference between the intercepts in
outcome equations for participants and non-participants can be interpreted as
the difference in average outcomes of participants and their counterfactual
average outcomes if they had chosen not to participate. The new estimator can
under mild conditions exhibit a rate of convergence in probability equal to
$n^{-p/(2p+1)}$, where $p\ge 2$ is an integer that indexes the strength of
certain smoothness assumptions. This rate of convergence is shown in this
context to be the optimal rate of convergence for estimation of the intercept
parameter in terms of a minimax criterion. The new estimator, unlike other
proposals in the literature, is under mild conditions consistent and
asymptotically normal with a rate of convergence that is the same regardless of
the degree to which selection depends on unobservables in the outcome equation.
Simulation evidence and an empirical example are included. | Rate-Optimal Estimation of the Intercept in a Semiparametric Sample-Selection Model | 2017-10-04 03:02:22 | Chuan Goh | http://arxiv.org/abs/1710.01423v3, http://arxiv.org/pdf/1710.01423v3 | econ.EM |
28,789 | em | Gale, Kuhn and Tucker (1950) introduced two ways to reduce a zero-sum game by
packaging some strategies with respect to a probability distribution on them.
In terms of value, they gave conditions for a desirable reduction. We show that
a probability distribution for a desirable reduction relies on optimal
strategies in the original game. Also, we correct an improper example given by
them to show that the reverse of a theorem does not hold. | A Note on Gale, Kuhn, and Tucker's Reductions of Zero-Sum Games | 2017-10-06 12:45:42 | Shuige Liu | http://arxiv.org/abs/1710.02326v1, http://arxiv.org/pdf/1710.02326v1 | econ.EM |
28,790 | em | This study proposes a simple technique for propensity score matching for
multiple treatment levels under the strong unconfoundedness assumption with the
help of the Aitchison distance proposed in the field of compositional data
analysis (CODA). | Propensity score matching for multiple treatment levels: A CODA-based contribution | 2017-10-24 03:27:47 | Hajime Seya, Takahiro Yoshida | http://arxiv.org/abs/1710.08558v1, http://arxiv.org/pdf/1710.08558v1 | econ.EM |
28,791 | em | We consider an index model of dyadic link formation with a homophily effect
index and a degree heterogeneity index. We provide nonparametric identification
results in a single large network setting for the potentially nonparametric
homophily effect function, the realizations of unobserved individual fixed
effects and the unknown distribution of idiosyncratic pairwise shocks, up to
normalization, for each possible true value of the unknown parameters. We
propose a novel form of scale normalization on an arbitrary interquantile
range, which is not only theoretically robust but also proves particularly
convenient for the identification analysis, as quantiles provide direct
linkages between the observable conditional probabilities and the unknown index
values. We then use an inductive "in-fill and out-expansion" algorithm to
establish our main results, and consider extensions to more general settings
that allow nonseparable dependence between homophily and degree heterogeneity,
as well as certain extents of network sparsity and weaker assumptions on the
support of unobserved heterogeneity. As a byproduct, we also propose a concept
called "modeling equivalence" as a refinement of "observational equivalence",
and use it to provide a formal discussion about normalization, identification
and their interplay with counterfactuals. | Nonparametric Identification in Index Models of Link Formation | 2017-10-30 23:32:12 | Wayne Yuan Gao | http://arxiv.org/abs/1710.11230v5, http://arxiv.org/pdf/1710.11230v5 | econ.EM |
28,792 | em | Web search data are a valuable source of business and economic information.
Previous studies have utilized Google Trends web search data for economic
forecasting. We expand this work by providing algorithms to combine and
aggregate search volume data, so that the resulting data is both consistent
over time and consistent between data series. We give a brand equity example,
where Google Trends is used to analyze shopping data for 100 top ranked brands
and these data are used to nowcast economic variables. We describe the
importance of out of sample prediction and show how principal component
analysis (PCA) can be used to improve the signal to noise ratio and prevent
overfitting in nowcasting models. We give a finance example, where exploratory
data analysis and classification is used to analyze the relationship between
Google Trends searches and stock prices. | Aggregating Google Trends: Multivariate Testing and Analysis | 2017-12-08 19:18:10 | Stephen L. France, Yuying Shi | http://arxiv.org/abs/1712.03152v2, http://arxiv.org/pdf/1712.03152v2 | econ.EM |
28,793 | em | We propose a new inferential methodology for dynamic economies that is robust
to misspecification of the mechanism generating frictions. Economies with
frictions are treated as perturbations of a frictionless economy that are
consistent with a variety of mechanisms. We derive a representation for the law
of motion for such economies and we characterize parameter set identification.
We derive a link from model aggregate predictions to distributional information
contained in qualitative survey data and specify conditions under which the
identified set is refined. The latter is used to semi-parametrically estimate
distortions due to frictions in macroeconomic variables. Based on these
estimates, we propose a novel test for complete models. Using consumer and
business survey data collected by the European Commission, we apply our method
to estimate distortions due to financial frictions in the Spanish economy. We
investigate the implications of these estimates for the adequacy of the
standard model of financial frictions SW-BGG (Smets and Wouters (2007),
Bernanke, Gertler, and Gilchrist (1999)). | Set Identified Dynamic Economies and Robustness to Misspecification | 2017-12-11 11:41:11 | Andreas Tryphonides | http://arxiv.org/abs/1712.03675v2, http://arxiv.org/pdf/1712.03675v2 | econ.EM |
28,794 | em | This paper defines the class of $\mathcal{H}$-valued autoregressive (AR)
processes with a unit root of finite type, where $\mathcal{H}$ is an infinite
dimensional separable Hilbert space, and derives a generalization of the
Granger-Johansen Representation Theorem valid for any integration order
$d=1,2,\dots$. An existence theorem shows that the solution of an AR with a
unit root of finite type is necessarily integrated of some finite integer $d$
and displays a common trends representation with a finite number of common
stochastic trends of the type of (cumulated) bilateral random walks and an
infinite dimensional cointegrating space. A characterization theorem clarifies
the connections between the structure of the AR operators and $(i)$ the order
of integration, $(ii)$ the structure of the attractor space and the
cointegrating space, $(iii)$ the expression of the cointegrating relations, and
$(iv)$ the Triangular representation of the process. Except for the fact that
the number of cointegrating relations that are integrated of order 0 is
infinite, the representation of $\mathcal{H}$-valued ARs with a unit root of
finite type coincides with that of usual finite dimensional VARs, which
corresponds to the special case $\mathcal{H}=\mathbb{R}^p$. | Cointegration in functional autoregressive processes | 2017-12-20 18:23:20 | Massimo Franchi, Paolo Paruolo | http://dx.doi.org/10.1017/S0266466619000306, http://arxiv.org/abs/1712.07522v2, http://arxiv.org/pdf/1712.07522v2 | econ.EM |
28,795 | em | High-dimensional linear models with endogenous variables play an increasingly
important role in recent econometric literature. In this work we allow for
models with many endogenous variables and many instrument variables to achieve
identification. Because of the high-dimensionality in the second stage,
constructing honest confidence regions with asymptotically correct coverage is
non-trivial. Our main contribution is to propose estimators and confidence
regions that would achieve that. The approach relies on moment conditions that
have an additional orthogonal property with respect to nuisance parameters.
Moreover, estimation of high-dimension nuisance parameters is carried out via
new pivotal procedures. In order to achieve simultaneously valid confidence
regions we use a multiplier bootstrap procedure to compute critical values and
establish its validity. | Simultaneous Confidence Intervals for High-dimensional Linear Models with Many Endogenous Variables | 2017-12-21 20:33:40 | Alexandre Belloni, Christian Hansen, Whitney Newey | http://arxiv.org/abs/1712.08102v4, http://arxiv.org/pdf/1712.08102v4 | econ.EM |
28,796 | em | This paper investigates the impacts of major natural resource discoveries
since 1960 on life expectancy in the nations that they were resource poor prior
to the discoveries. Previous literature explains the relation between nations
wealth and life expectancy, but it has been silent about the impacts of
resource discoveries on life expectancy. We attempt to fill this gap in this
study. An important advantage of this study is that as the previous researchers
argued resource discovery could be an exogenous variable. We use longitudinal
data from 1960 to 2014 and we apply three modern empirical methods including
Difference-in-Differences, Event studies, and Synthetic Control approach, to
investigate the main question of the research which is 'how resource
discoveries affect life expectancy?'. The findings show that resource
discoveries in Ecuador, Yemen, Oman, and Equatorial Guinea have positive and
significant impacts on life expectancy, but the effects for the European
countries are mostly negative. | Resource Abundance and Life Expectancy | 2018-01-01 01:43:39 | Bahram Sanginabadi | http://arxiv.org/abs/1801.00369v1, http://arxiv.org/pdf/1801.00369v1 | econ.EM |
28,797 | em | In this paper we estimate a Bayesian vector autoregressive model with factor
stochastic volatility in the error term to assess the effects of an uncertainty
shock in the Euro area. This allows us to treat macroeconomic uncertainty as a
latent quantity during estimation. Only a limited number of contributions to
the literature estimate uncertainty and its macroeconomic consequences jointly,
and most are based on single country models. We analyze the special case of a
shock restricted to the Euro area, where member states are highly related by
construction. We find significant results of a decrease in real activity for
all countries over a period of roughly a year following an uncertainty shock.
Moreover, equity prices, short-term interest rates and exports tend to decline,
while unemployment levels increase. Dynamic responses across countries differ
slightly in magnitude and duration, with Ireland, Slovakia and Greece
exhibiting different reactions for some macroeconomic fundamentals. | Implications of macroeconomic volatility in the Euro area | 2018-01-09 16:20:42 | Niko Hauzenberger, Maximilian Böck, Michael Pfarrhofer, Anna Stelzer, Gregor Zens | http://arxiv.org/abs/1801.02925v2, http://arxiv.org/pdf/1801.02925v2 | econ.EM |
28,798 | em | We report a new result on lotteries --- that a well-funded syndicate has a
purely mechanical strategy to achieve expected returns of 10\% to 25\% in an
equiprobable lottery with no take and no carryover pool. We prove that an
optimal strategy (Nash equilibrium) in a game between the syndicate and other
players consists of betting one of each ticket (the "trump ticket"), and extend
that result to proportional ticket selection in non-equiprobable lotteries. The
strategy can be adjusted to accommodate lottery taxes and carryover pools. No
"irrationality" need be involved for the strategy to succeed --- it requires
only that a large group of non-syndicate bettors each choose a few tickets
independently. | A Method for Winning at Lotteries | 2018-01-05 22:35:17 | Steven D. Moffitt, William T. Ziemba | http://arxiv.org/abs/1801.02958v1, http://arxiv.org/pdf/1801.02958v1 | econ.EM |
28,799 | em | Despite its unusual payout structure, the Canadian 6/49 Lotto is one of the
few government sponsored lotteries that has the potential for a favorable
strategy we call "buying the pot." By buying the pot we mean that a syndicate
buys each ticket in the lottery, ensuring that it holds a jackpot winner. We
assume that the other bettors independently buy small numbers of tickets. This
paper presents (1) a formula for the syndicate's expected return, (2)
conditions under which buying the pot produces a significant positive expected
return, and (3) the implications of these findings for lottery design. | Does it Pay to Buy the Pot in the Canadian 6/49 Lotto? Implications for Lottery Design | 2018-01-06 00:58:18 | Steven D. Moffitt, William T. Ziemba | http://arxiv.org/abs/1801.02959v1, http://arxiv.org/pdf/1801.02959v1 | econ.EM |
28,800 | em | Dynamic Discrete Choice Models (DDCMs) are important in the structural
estimation literature. Since the structural errors are practically always
continuous and unbounded in nature, researchers often use the expected value
function. The idea to solve for the expected value function made solution more
practical and estimation feasible. However, as we show in this paper, the
expected value function is impractical compared to an alternative: the
integrated (ex ante) value function. We provide brief descriptions of the
inefficacy of the former, and benchmarks on actual problems with varying
cardinality of the state space and number of decisions. Though the two
approaches solve the same problem in theory, the benchmarks support the claim
that the integrated value function is preferred in practice. | Solving Dynamic Discrete Choice Models: Integrated or Expected Value Function? | 2018-01-11 23:26:00 | Patrick Kofod Mogensen | http://arxiv.org/abs/1801.03978v1, http://arxiv.org/pdf/1801.03978v1 | econ.EM |
28,801 | em | This paper develops a new model and estimation procedure for panel data that
allows us to identify heterogeneous structural breaks. We model individual
heterogeneity using a grouped pattern. For each group, we allow common
structural breaks in the coefficients. However, the number, timing, and size of
these breaks can differ across groups. We develop a hybrid estimation procedure
of the grouped fixed effects approach and adaptive group fused Lasso. We show
that our method can consistently identify the latent group structure, detect
structural breaks, and estimate the regression parameters. Monte Carlo results
demonstrate the good performance of the proposed method in finite samples. An
empirical application to the relationship between income and democracy
illustrates the importance of considering heterogeneous structural breaks. | Heterogeneous structural breaks in panel data models | 2018-01-15 09:19:28 | Ryo Okui, Wendun Wang | http://arxiv.org/abs/1801.04672v2, http://arxiv.org/pdf/1801.04672v2 | econ.EM |
28,802 | em | We characterize common assumption of rationality of 2-person games within an
incomplete information framework. We use the lexicographic model with
incomplete information and show that a belief hierarchy expresses common
assumption of rationality within a complete information framework if and only
if there is a belief hierarchy within the corresponding incomplete information
framework that expresses common full belief in caution, rationality, every good
choice is supported, and prior belief in the original utility functions. | Characterizing Assumption of Rationality by Incomplete Information | 2018-01-15 12:48:20 | Shuige Liu | http://arxiv.org/abs/1801.04714v1, http://arxiv.org/pdf/1801.04714v1 | econ.EM |
28,803 | em | We first show (1) the importance of investigating health expenditure process
using the order two Markov chain model, rather than the standard order one
model, which is widely used in the literature. Markov chain of order two is the
minimal framework that is capable of distinguishing those who experience a
certain health expenditure level for the first time from those who have been
experiencing that or other levels for some time. In addition, using the model
we show (2) that the probability of encountering a health shock first de-
creases until around age 10, and then increases with age, particularly, after
age 40, (3) that health shock distributions among different age groups do not
differ until their percentiles reach the median range, but that above the
median the health shock distributions of older age groups gradually start to
first-order dominate those of younger groups, and (4) that the persistency of
health shocks also shows a U-shape in relation to age. | Quantifying Health Shocks Over the Life Cycle | 2018-01-26 13:35:38 | Taiyo Fukai, Hidehiko Ichimura, Kyogo Kanazawa | http://arxiv.org/abs/1801.08746v1, http://arxiv.org/pdf/1801.08746v1 | econ.EM |
28,804 | em | We define a modification of the standard Kripke model, called the ordered
Kripke model, by introducing a linear order on the set of accessible states of
each state. We first show this model can be used to describe the lexicographic
belief hierarchy in epistemic game theory, and perfect rationalizability can be
characterized within this model. Then we show that each ordered Kripke model is
the limit of a sequence of standard probabilistic Kripke models with a modified
(common) belief operator, in the senses of structure and the
(epsilon-)permissibilities characterized within them. | Ordered Kripke Model, Permissibility, and Convergence of Probabilistic Kripke Model | 2018-01-26 14:46:28 | Shuige Liu | http://arxiv.org/abs/1801.08767v1, http://arxiv.org/pdf/1801.08767v1 | econ.EM |
28,805 | em | Why women avoid participating in a competition and how can we encourage them
to participate in it? In this paper, we investigate how social image concerns
affect women's decision to compete. We first construct a theoretical model and
show that participating in a competition, even under affirmative action
policies favoring women, is costly for women under public observability since
it deviates from traditional female gender norms, resulting in women's low
appearance in competitive environments. We propose and theoretically show that
introducing prosocial incentives in the competitive environment is effective
and robust to public observability since (i) it induces women who are
intrinsically motivated by prosocial incentives to the competitive environment
and (ii) it makes participating in a competition not costly for women from
social image point of view. We conduct a laboratory experiment where we
randomly manipulate the public observability of decisions to compete and test
our theoretical predictions. The results of the experiment are fairly
consistent with our theoretical predictions. We suggest that when designing
policies to promote gender equality in competitive environments, using
prosocial incentives through company philanthropy or other social
responsibility policies, either as substitutes or as complements to traditional
affirmative action policies, could be promising. | How Can We Induce More Women to Competitions? | 2018-01-27 11:51:44 | Masayuki Yagasaki, Mitsunosuke Morishita | http://arxiv.org/abs/1801.10518v1, http://arxiv.org/pdf/1801.10518v1 | econ.EM |
28,806 | em | The rational choice theory is based on this idea that people rationally
pursue goals for increasing their personal interests. In most conditions, the
behavior of an actor is not independent of the person and others' behavior.
Here, we present a new concept of rational choice as a hyper-rational choice
which in this concept, the actor thinks about profit or loss of other actors in
addition to his personal profit or loss and then will choose an action which is
desirable to him. We implement the hyper-rational choice to generalize and
expand the game theory. Results of this study will help to model the behavior
of people considering environmental conditions, the kind of behavior
interactive, valuation system of itself and others and system of beliefs and
internal values of societies. Hyper-rationality helps us understand how human
decision makers behave in interactive decisions. | Hyper-rational choice theory | 2018-01-12 02:16:09 | Madjid Eshaghi Gordji, Gholamreza Askari | http://arxiv.org/abs/1801.10520v2, http://arxiv.org/pdf/1801.10520v2 | econ.EM |
28,807 | em | We develop a new VAR model for structural analysis with mixed-frequency data.
The MIDAS-SVAR model allows to identify structural dynamic links exploiting the
information contained in variables sampled at different frequencies. It also
provides a general framework to test homogeneous frequency-based
representations versus mixed-frequency data models. A set of Monte Carlo
experiments suggests that the test performs well both in terms of size and
power. The MIDAS-SVAR is then used to study how monetary policy and financial
market volatility impact on the dynamics of gross capital inflows to the US.
While no relation is found when using standard quarterly data, exploiting the
variability present in the series within the quarter shows that the effect of
an interest rate shock is greater the longer the time lag between the month of
the shock and the end of the quarter | Structural analysis with mixed-frequency data: A MIDAS-SVAR model of US capital flows | 2018-02-02 21:12:12 | Emanuele Bacchiocchi, Andrea Bastianin, Alessandro Missale, Eduardo Rossi | http://arxiv.org/abs/1802.00793v1, http://arxiv.org/pdf/1802.00793v1 | econ.EM |
28,808 | em | The development and deployment of matching procedures that incentivize
truthful preference reporting is considered one of the major successes of
market design research. In this study, we test the degree to which these
procedures succeed in eliminating preference misrepresentation. We administered
an online experiment to 1,714 medical students immediately after their
participation in the medical residency match--a leading field application of
strategy-proof market design. When placed in an analogous, incentivized
matching task, we find that 23% of participants misrepresent their preferences.
We explore the factors that predict preference misrepresentation, including
cognitive ability, strategic positioning, overconfidence, expectations, advice,
and trust. We discuss the implications of this behavior for the design of
allocation mechanisms and the social welfare in markets that use them. | An Experimental Investigation of Preference Misrepresentation in the Residency Match | 2018-02-05 20:51:55 | Alex Rees-Jones, Samuel Skowronek | http://dx.doi.org/10.1073/pnas.1803212115, http://arxiv.org/abs/1802.01990v2, http://arxiv.org/pdf/1802.01990v2 | econ.EM |
28,809 | em | Consumers are creatures of habit, often periodic, tied to work, shopping and
other schedules. We analyzed one month of data from the world's largest
bike-sharing company to elicit demand behavioral cycles, initially using models
from animal tracking that showed large customers fit an Ornstein-Uhlenbeck
model with demand peaks at periodicities of 7, 12, 24 hour and 7-days. Lorenz
curves of bicycle demand showed that the majority of customer usage was
infrequent, and demand cycles from time-series models would strongly overfit
the data yielding unreliable models. Analysis of thresholded wavelets for the
space-time tensor of bike-sharing contracts was able to compress the data into
a 56-coefficient model with little loss of information, suggesting that
bike-sharing demand behavior is exceptionally strong and regular. Improvements
to predicted demand could be made by adjusting for 'noise' filtered by our
model from air quality and weather information and demand from infrequent
riders. | Prediction of Shared Bicycle Demand with Wavelet Thresholding | 2018-02-08 04:17:27 | J. Christopher Westland, Jian Mou, Dafei Yin | http://arxiv.org/abs/1802.02683v1, http://arxiv.org/pdf/1802.02683v1 | econ.EM |
28,810 | em | This paper describes a numerical method to solve for mean product qualities
which equates the real market share to the market share predicted by a discrete
choice model. The method covers a general class of discrete choice model,
including the pure characteristics model in Berry and Pakes(2007) and the
random coefficient logit model in Berry et al.(1995) (hereafter BLP). The
method transforms the original market share inversion problem to an
unconstrained convex minimization problem, so that any convex programming
algorithm can be used to solve the inversion. Moreover, such results also imply
that the computational complexity of inverting a demand model should be no more
than that of a convex programming problem. In simulation examples, I show the
method outperforms the contraction mapping algorithm in BLP. I also find the
method remains robust in pure characteristics models with near-zero market
shares. | A General Method for Demand Inversion | 2018-02-13 05:50:46 | Lixiong Li | http://arxiv.org/abs/1802.04444v3, http://arxiv.org/pdf/1802.04444v3 | econ.EM |
29,016 | em | This paper develops a set of test statistics based on bilinear forms in the
context of the extremum estimation framework with particular interest in
nonlinear hypothesis. We show that the proposed statistic converges to a
conventional chi-square limit. A Monte Carlo experiment suggests that the test
statistic works well in finite samples. | Bilinear form test statistics for extremum estimation | 2019-12-03 17:32:49 | Federico Crudu, Felipe Osorio | http://dx.doi.org/10.1016/j.econlet.2019.108885, http://arxiv.org/abs/1912.01410v1, http://arxiv.org/pdf/1912.01410v1 | econ.EM |
28,811 | em | We provide an epistemic foundation for cooperative games by proof theory via
studying the knowledge for players unanimously accepting only core payoffs. We
first transform each cooperative game into a decision problem where a player
can accept or reject any payoff vector offered to her based on her knowledge
about available cooperation. Then we use a modified KD-system in epistemic
logic, which can be regarded as a counterpart of the model for non-cooperative
games in Bonanno (2008), (2015), to describe a player's knowledge,
decision-making criterion, and reasoning process; especially, a formula called
C-acceptability is defined to capture the criterion for accepting a core payoff
vector. Within this syntactical framework, we characterize the core of a
cooperative game in terms of players' knowledge. Based on that result, we
discuss an epistemic inconsistency behind Debreu-Scarf Theorem, that is, the
increase of the number of replicas has invariant requirement on each
participant's knowledge from the aspect of competitive market, while requires
unbounded epistemic ability players from the aspect of cooperative game. | Knowledge and Unanimous Acceptance of Core Payoffs: An Epistemic Foundation for Cooperative Game Theory | 2018-02-13 15:49:12 | Shuige Liu | http://arxiv.org/abs/1802.04595v4, http://arxiv.org/pdf/1802.04595v4 | econ.EM |
28,812 | em | In this study interest centers on regional differences in the response of
housing prices to monetary policy shocks in the US. We address this issue by
analyzing monthly home price data for metropolitan regions using a
factor-augmented vector autoregression (FAVAR) model. Bayesian model estimation
is based on Gibbs sampling with Normal-Gamma shrinkage priors for the
autoregressive coefficients and factor loadings, while monetary policy shocks
are identified using high-frequency surprises around policy announcements as
external instruments. The empirical results indicate that monetary policy
actions typically have sizeable and significant positive effects on regional
housing prices, revealing differences in magnitude and duration. The largest
effects are observed in regions located in states on both the East and West
Coasts, notably California, Arizona and Florida. | The dynamic impact of monetary policy on regional housing prices in the US: Evidence based on factor-augmented vector autoregressions | 2018-02-16 12:08:34 | Manfred M. Fischer, Florian Huber, Michael Pfarrhofer, Petra Staufer-Steinnocher | http://arxiv.org/abs/1802.05870v1, http://arxiv.org/pdf/1802.05870v1 | econ.EM |
28,813 | em | We study the asymptotic properties of a class of estimators of the structural
parameters in dynamic discrete choice games. We consider K-stage policy
iteration (PI) estimators, where K denotes the number of policy iterations
employed in the estimation. This class nests several estimators proposed in the
literature such as those in Aguirregabiria and Mira (2002, 2007), Pesendorfer
and Schmidt-Dengler (2008), and Pakes et al. (2007). First, we establish that
the K-PML estimator is consistent and asymptotically normal for all K. This
complements findings in Aguirregabiria and Mira (2007), who focus on K=1 and K
large enough to induce convergence of the estimator. Furthermore, we show under
certain conditions that the asymptotic variance of the K-PML estimator can
exhibit arbitrary patterns as a function of K. Second, we establish that the
K-MD estimator is consistent and asymptotically normal for all K. For a
specific weight matrix, the K-MD estimator has the same asymptotic distribution
as the K-PML estimator. Our main result provides an optimal sequence of weight
matrices for the K-MD estimator and shows that the optimally weighted K-MD
estimator has an asymptotic distribution that is invariant to K. The invariance
result is especially unexpected given the findings in Aguirregabiria and Mira
(2007) for K-PML estimators. Our main result implies two new corollaries about
the optimal 1-MD estimator (derived by Pesendorfer and Schmidt-Dengler (2008)).
First, the optimal 1-MD estimator is optimal in the class of K-MD estimators.
In other words, additional policy iterations do not provide asymptotic
efficiency gains relative to the optimal 1-MD estimator. Second, the optimal
1-MD estimator is more or equally asymptotically efficient than any K-PML
estimator for all K. Finally, the appendix provides appropriate conditions
under which the optimal 1-MD estimator is asymptotically efficient. | On the iterated estimation of dynamic discrete choice games | 2018-02-19 18:19:35 | Federico A. Bugni, Jackson Bunting | http://arxiv.org/abs/1802.06665v4, http://arxiv.org/pdf/1802.06665v4 | econ.EM |
28,814 | em | This paper proposes nonparametric kernel-smoothing estimation for panel data
to examine the degree of heterogeneity across cross-sectional units. We first
estimate the sample mean, autocovariances, and autocorrelations for each unit
and then apply kernel smoothing to compute their density functions. The
dependence of the kernel estimator on bandwidth makes asymptotic bias of very
high order affect the required condition on the relative magnitudes of the
cross-sectional sample size (N) and the time-series length (T). In particular,
it makes the condition on N and T stronger and more complicated than those
typically observed in the long-panel literature without kernel smoothing. We
also consider a split-panel jackknife method to correct bias and construction
of confidence intervals. An empirical application and Monte Carlo simulations
illustrate our procedure in finite samples. | Kernel Estimation for Panel Data with Heterogeneous Dynamics | 2018-02-24 12:45:50 | Ryo Okui, Takahide Yanagi | http://arxiv.org/abs/1802.08825v4, http://arxiv.org/pdf/1802.08825v4 | econ.EM |
28,815 | em | People reason heuristically in situations resembling inferential puzzles such
as Bertrand's box paradox and the Monty Hall problem. The practical
significance of that fact for economic decision making is uncertain because a
departure from sound reasoning may, but does not necessarily, result in a
"cognitively biased" outcome different from what sound reasoning would have
produced. Criteria are derived here, applicable to both experimental and
non-experimental situations, for heuristic reasoning in an inferential-puzzle
situations to result, or not to result, in cognitively bias. In some
situations, neither of these criteria is satisfied, and whether or not agents'
posterior probability assessments or choices are cognitively biased cannot be
determined. | Identifying the occurrence or non occurrence of cognitive bias in situations resembling the Monty Hall problem | 2018-02-25 03:28:11 | Fatemeh Borhani, Edward J. Green | http://arxiv.org/abs/1802.08935v1, http://arxiv.org/pdf/1802.08935v1 | econ.EM |
28,816 | em | I analyse the solution method for the variational optimisation problem in the
rational inattention framework proposed by Christopher A. Sims. The solution,
in general, does not exist, although it may exist in exceptional cases. I show
that the solution does not exist for the quadratic and the logarithmic
objective functions analysed by Sims (2003, 2006). For a linear-quadratic
objective function a solution can be constructed under restrictions on all but
one of its parameters. This approach is, therefore, unlikely to be applicable
to a wider set of economic models. | On the solution of the variational optimisation in the rational inattention framework | 2018-02-27 16:21:46 | Nigar Hashimzade | http://arxiv.org/abs/1802.09869v2, http://arxiv.org/pdf/1802.09869v2 | econ.EM |
28,830 | em | This paper proposes a model-free approach to analyze panel data with
heterogeneous dynamic structures across observational units. We first compute
the sample mean, autocovariances, and autocorrelations for each unit, and then
estimate the parameters of interest based on their empirical distributions. We
then investigate the asymptotic properties of our estimators using double
asymptotics and propose split-panel jackknife bias correction and inference
based on the cross-sectional bootstrap. We illustrate the usefulness of our
procedures by studying the deviation dynamics of the law of one price. Monte
Carlo simulations confirm that the proposed bias correction is effective and
yields valid inference in small samples. | Panel Data Analysis with Heterogeneous Dynamics | 2018-03-26 10:53:47 | Ryo Okui, Takahide Yanagi | http://arxiv.org/abs/1803.09452v2, http://arxiv.org/pdf/1803.09452v2 | econ.EM |
28,817 | em | Many macroeconomic policy questions may be assessed in a case study
framework, where the time series of a treated unit is compared to a
counterfactual constructed from a large pool of control units. I provide a
general framework for this setting, tailored to predict the counterfactual by
minimizing a tradeoff between underfitting (bias) and overfitting (variance).
The framework nests recently proposed structural and reduced form machine
learning approaches as special cases. Furthermore, difference-in-differences
with matching and the original synthetic control are restrictive cases of the
framework, in general not minimizing the bias-variance objective. Using
simulation studies I find that machine learning methods outperform traditional
methods when the number of potential controls is large or the treated unit is
substantially different from the controls. Equipped with a toolbox of
approaches, I revisit a study on the effect of economic liberalisation on
economic growth. I find effects for several countries where no effect was found
in the original study. Furthermore, I inspect how a systematically important
bank respond to increasing capital requirements by using a large pool of banks
to estimate the counterfactual. Finally, I assess the effect of a changing
product price on product sales using a novel scanner dataset. | Synthetic Control Methods and Big Data | 2018-03-01 00:32:09 | Daniel Kinn | http://arxiv.org/abs/1803.00096v1, http://arxiv.org/pdf/1803.00096v1 | econ.EM |
28,818 | em | It is widely known that geographically weighted regression(GWR) is
essentially same as varying-coefficient model. In the former research about
varying-coefficient model, scholars tend to use multidimensional-kernel-based
locally weighted estimation(MLWE) so that information of both distance and
direction is considered. However, when we construct the local weight matrix of
geographically weighted estimation, distance among the locations in the
neighbor is the only factor controlling the value of entries of weight matrix.
In other word, estimation of GWR is distance-kernel-based. Thus, in this paper,
under stationary and limited dependent data with multidimensional subscripts,
we analyze the local mean squared properties of without any assumption of the
form of coefficient functions and compare it with MLWE. According to the
theoretical and simulation results, geographically-weighted locally linear
estimation(GWLE) is asymptotically more efficient than MLWE. Furthermore, a
relationship between optimal bandwith selection and design of scale parameters
is also obtained. | An Note on Why Geographically Weighted Regression Overcomes Multidimensional-Kernel-Based Varying-Coefficient Model | 2018-03-04 21:50:17 | Zihao Yuan | http://arxiv.org/abs/1803.01402v2, http://arxiv.org/pdf/1803.01402v2 | econ.EM |
28,819 | em | We study three pricing mechanisms' performance and their effects on the
participants in the data industry from the data supply chain perspective. A
win-win pricing strategy for the players in the data supply chain is proposed.
We obtain analytical solutions in each pricing mechanism, including the
decentralized and centralized pricing, Nash Bargaining pricing, and revenue
sharing mechanism. | Pricing Mechanism in Information Goods | 2018-03-05 10:37:06 | Xinming Li, Huaqing Wang | http://arxiv.org/abs/1803.01530v1, http://arxiv.org/pdf/1803.01530v1 | econ.EM |
28,820 | em | Spatial association and heterogeneity are two critical areas in the research
about spatial analysis, geography, statistics and so on. Though large amounts
of outstanding methods has been proposed and studied, there are few of them
tend to study spatial association under heterogeneous environment.
Additionally, most of the traditional methods are based on distance statistic
and spatial weighted matrix. However, in some abstract spatial situations,
distance statistic can not be applied since we can not even observe the
geographical locations directly. Meanwhile, under these circumstances, due to
invisibility of spatial positions, designing of weight matrix can not
absolutely avoid subjectivity. In this paper, a new entropy-based method, which
is data-driven and distribution-free, has been proposed to help us investigate
spatial association while fully taking the fact that heterogeneity widely
exist. Specifically, this method is not bounded with distance statistic or
weight matrix. Asymmetrical dependence is adopted to reflect the heterogeneity
in spatial association for each individual and the whole discussion in this
paper is performed on spatio-temporal data with only assuming stationary
m-dependent over time. | A Nonparametric Approach to Measure the Heterogeneous Spatial Association: Under Spatial Temporal Data | 2018-03-06 21:46:49 | Zihao Yuan | http://arxiv.org/abs/1803.02334v2, http://arxiv.org/pdf/1803.02334v2 | econ.EM |
28,821 | em | The last technical barriers to trade(TBT) between countries are Non-Tariff
Barriers(NTBs), meaning all trade barriers are possible other than Tariff
Barriers. And the most typical examples are (TBT), which refer to measure
Technical Regulation, Standards, Procedure for Conformity Assessment, Test &
Certification etc. Therefore, in order to eliminate TBT, WTO has made all
membership countries automatically enter into an agreement on TBT | A study of strategy to the remove and ease TBT for increasing export in GCC6 countries | 2018-03-09 09:39:31 | YongJae Kim | http://arxiv.org/abs/1803.03394v3, http://arxiv.org/pdf/1803.03394v3 | econ.EM |
28,822 | em | Understanding the effectiveness of alternative approaches to water
conservation is crucially important for ensuring the security and reliability
of water services for urban residents. We analyze data from one of the
longest-running "cash for grass" policies - the Southern Nevada Water
Authority's Water Smart Landscapes program, where homeowners are paid to
replace grass with xeric landscaping. We use a twelve year long panel dataset
of monthly water consumption records for 300,000 households in Las Vegas,
Nevada. Utilizing a panel difference-in-differences approach, we estimate the
average water savings per square meter of turf removed. We find that
participation in this program reduced the average treated household's
consumption by 18 percent. We find no evidence that water savings degrade as
the landscape ages, or that water savings per unit area are influenced by the
value of the rebate. Depending on the assumed time horizon of benefits from
turf removal, we find that the WSL program cost the water authority about $1.62
per thousand gallons of water saved, which compares favorably to alternative
means of water conservation or supply augmentation. | How Smart Are `Water Smart Landscapes'? | 2018-03-13 05:00:07 | Christa Brelsford, Joshua K. Abbott | http://arxiv.org/abs/1803.04593v1, http://arxiv.org/pdf/1803.04593v1 | econ.EM |
28,823 | em | The business cycles are generated by the oscillating macro-/micro-/nano-
economic output variables in the economy of the scale and the scope in the
amplitude/frequency/phase/time domains in the economics. The accurate forward
looking assumptions on the business cycles oscillation dynamics can optimize
the financial capital investing and/or borrowing by the economic agents in the
capital markets. The book's main objective is to study the business cycles in
the economy of the scale and the scope, formulating the Ledenyov unified
business cycles theory in the Ledenyov classic and quantum econodynamics. | Business Cycles in Economics | 2018-03-16 11:24:05 | Viktor O. Ledenyov, Dimitri O. Ledenyov | http://dx.doi.org/10.2139/ssrn.3134655, http://arxiv.org/abs/1803.06108v1, http://arxiv.org/pdf/1803.06108v1 | econ.EM |
28,824 | em | Unobserved heterogeneous treatment effects have been emphasized in the recent
policy evaluation literature (see e.g., Heckman and Vytlacil, 2005). This paper
proposes a nonparametric test for unobserved heterogeneous treatment effects in
a treatment effect model with a binary treatment assignment, allowing for
individuals' self-selection to the treatment. Under the standard local average
treatment effects assumptions, i.e., the no defiers condition, we derive
testable model restrictions for the hypothesis of unobserved heterogeneous
treatment effects. Also, we show that if the treatment outcomes satisfy a
monotonicity assumption, these model restrictions are also sufficient. Then, we
propose a modified Kolmogorov-Smirnov-type test which is consistent and simple
to implement. Monte Carlo simulations show that our test performs well in
finite samples. For illustration, we apply our test to study heterogeneous
treatment effects of the Job Training Partnership Act on earnings and the
impacts of fertility on family income, where the null hypothesis of homogeneous
treatment effects gets rejected in the second case but fails to be rejected in
the first application. | Testing for Unobserved Heterogeneous Treatment Effects with Observational Data | 2018-03-20 19:30:07 | Yu-Chin Hsu, Ta-Cheng Huang, Haiqing Xu | http://arxiv.org/abs/1803.07514v2, http://arxiv.org/pdf/1803.07514v2 | econ.EM |
28,825 | em | In the regression discontinuity design (RDD), it is common practice to assess
the credibility of the design by testing the continuity of the density of the
running variable at the cut-off, e.g., McCrary (2008). In this paper we propose
an approximate sign test for continuity of a density at a point based on the
so-called g-order statistics, and study its properties under two complementary
asymptotic frameworks. In the first asymptotic framework, the number q of
observations local to the cut-off is fixed as the sample size n diverges to
infinity, while in the second framework q diverges to infinity slowly as n
diverges to infinity. Under both of these frameworks, we show that the test we
propose is asymptotically valid in the sense that it has limiting rejection
probability under the null hypothesis not exceeding the nominal level. More
importantly, the test is easy to implement, asymptotically valid under weaker
conditions than those used by competing methods, and exhibits finite sample
validity under stronger conditions than those needed for its asymptotic
validity. In a simulation study, we find that the approximate sign test
provides good control of the rejection probability under the null hypothesis
while remaining competitive under the alternative hypothesis. We finally apply
our test to the design in Lee (2008), a well-known application of the RDD to
study incumbency advantage. | Testing Continuity of a Density via g-order statistics in the Regression Discontinuity Design | 2018-03-21 17:52:59 | Federico A. Bugni, Ivan A. Canay | http://arxiv.org/abs/1803.07951v6, http://arxiv.org/pdf/1803.07951v6 | econ.EM |
28,826 | em | In this paper, we propose the use of causal inference techniques for survival
function estimation and prediction for subgroups of the data, upto individual
units. Tree ensemble methods, specifically random forests were modified for
this purpose. A real world healthcare dataset was used with about 1800 patients
with breast cancer, which has multiple patient covariates as well as disease
free survival days (DFS) and a death event binary indicator (y). We use the
type of cancer curative intervention as the treatment variable (T=0 or 1,
binary treatment case in our example). The algorithm is a 2 step approach. In
step 1, we estimate heterogeneous treatment effects using a causalTree with the
DFS as the dependent variable. Next, in step 2, for each selected leaf of the
causalTree with distinctly different average treatment effect (with respect to
survival), we fit a survival forest to all the patients in that leaf, one
forest each for treatment T=0 as well as T=1 to get estimated patient level
survival curves for each treatment (more generally, any model can be used at
this step). Then, we subtract the patient level survival curves to get the
differential survival curve for a given patient, to compare the survival
function as a result of the 2 treatments. The path to a selected leaf also
gives us the combination of patient features and their values which are
causally important for the treatment effect difference at the leaf. | Causal Inference for Survival Analysis | 2018-03-22 06:22:19 | Vikas Ramachandra | http://arxiv.org/abs/1803.08218v1, http://arxiv.org/pdf/1803.08218v1 | econ.EM |
28,827 | em | Linear regressions with period and group fixed effects are widely used to
estimate treatment effects. We show that they estimate weighted sums of the
average treatment effects (ATE) in each group and period, with weights that may
be negative. Due to the negative weights, the linear regression coefficient may
for instance be negative while all the ATEs are positive. We propose another
estimator that solves this issue. In the two applications we revisit, it is
significantly different from the linear regression estimator. | Two-way fixed effects estimators with heterogeneous treatment effects | 2018-03-22 01:56:07 | Clément de Chaisemartin, Xavier D'Haultfœuille | http://dx.doi.org/10.1257/aer.20181169, http://arxiv.org/abs/1803.08807v7, http://arxiv.org/pdf/1803.08807v7 | econ.EM |
28,828 | em | We examine the effects of monetary policy on income inequality in Japan using
a novel econometric approach that jointly estimates the Gini coefficient based
on micro-level grouped data of households and the dynamics of macroeconomic
quantities. Our results indicate different effects on income inequality for
different types of households: A monetary tightening increases inequality when
income data is based on households whose head is employed (workers'
households), while the effect reverses over the medium term when considering a
broader definition of households. Differences in the relative strength of the
transmission channels can account for this finding. Finally we demonstrate that
the proposed joint estimation strategy leads to more informative inference
while results based on the frequently used two-step estimation approach yields
inconclusive results. | How does monetary policy affect income inequality in Japan? Evidence from grouped data | 2018-03-23 19:28:23 | Martin Feldkircher, Kazuhiko Kakamu | http://dx.doi.org/10.1007/s00181-021-02102-7, http://arxiv.org/abs/1803.08868v2, http://arxiv.org/pdf/1803.08868v2 | econ.EM |
28,829 | em | We develop inference for a two-sided matching model where the characteristics
of agents on one side of the market are endogenous due to pre-matching
investments. The model can be used to measure the impact of frictions in labour
markets using a single cross-section of matched employer-employee data. The
observed matching of workers to firms is the outcome of a discrete, two-sided
matching process where firms with heterogeneous preferences over education
sequentially choose workers according to an index correlated with worker
preferences over firms. The distribution of education arises in equilibrium
from a Bayesian game: workers, knowing the distribution of worker and firm
types, invest in education prior to the matching process. Although the observed
matching exhibits strong cross-sectional dependence due to the matching
process, we propose an asymptotically valid inference procedure that combines
discrete choice methods with simulation. | Schooling Choice, Labour Market Matching, and Wages | 2018-03-24 03:41:09 | Jacob Schwartz | http://arxiv.org/abs/1803.09020v6, http://arxiv.org/pdf/1803.09020v6 | econ.EM |
28,831 | em | In this paper, we assess the impact of climate shocks on futures markets for
agricultural commodities and a set of macroeconomic quantities for multiple
high-income economies. To capture relations among countries, markets, and
climate shocks, this paper proposes parsimonious methods to estimate
high-dimensional panel VARs. We assume that coefficients associated with
domestic lagged endogenous variables arise from a Gaussian mixture model while
further parsimony is achieved using suitable global-local shrinkage priors on
several regions of the parameter space. Our results point towards pronounced
global reactions of key macroeconomic quantities to climate shocks. Moreover,
the empirical findings highlight substantial linkages between regionally
located climate shifts and global commodity markets. | A Bayesian panel VAR model to analyze the impact of climate change on high-income economies | 2018-04-04 21:23:10 | Florian Huber, Tamás Krisztin, Michael Pfarrhofer | http://arxiv.org/abs/1804.01554v3, http://arxiv.org/pdf/1804.01554v3 | econ.EM |
28,832 | em | This paper provides a new methodology to analyze unobserved heterogeneity
when observed characteristics are modeled nonlinearly. The proposed model
builds on varying random coefficients (VRC) that are determined by nonlinear
functions of observed regressors and additively separable unobservables. This
paper proposes a novel estimator of the VRC density based on weighted sieve
minimum distance. The main example of sieve bases are Hermite functions which
yield a numerically stable estimation procedure. This paper shows inference
results that go beyond what has been shown in ordinary RC models. We provide in
each case rates of convergence and also establish pointwise limit theory of
linear functionals, where a prominent example is the density of potential
outcomes. In addition, a multiplier bootstrap procedure is proposed to
construct uniform confidence bands. A Monte Carlo study examines finite sample
properties of the estimator and shows that it performs well even when the
regressors associated to RC are far from being heavy tailed. Finally, the
methodology is applied to analyze heterogeneity in income elasticity of demand
for housing. | Varying Random Coefficient Models | 2018-04-09 20:16:52 | Christoph Breunig | http://arxiv.org/abs/1804.03110v4, http://arxiv.org/pdf/1804.03110v4 | econ.EM |
28,833 | em | We develop point-identification for the local average treatment effect when
the binary treatment contains a measurement error. The standard instrumental
variable estimator is inconsistent for the parameter since the measurement
error is non-classical by construction. We correct the problem by identifying
the distribution of the measurement error based on the use of an exogenous
variable that can even be a binary covariate. The moment conditions derived
from the identification lead to generalized method of moments estimation with
asymptotically valid inferences. Monte Carlo simulations and an empirical
illustration demonstrate the usefulness of the proposed procedure. | Inference on Local Average Treatment Effects for Misclassified Treatment | 2018-04-10 08:57:30 | Takahide Yanagi | http://arxiv.org/abs/1804.03349v1, http://arxiv.org/pdf/1804.03349v1 | econ.EM |
28,834 | em | This paper re-examines the Shapley value methods for attribution analysis in
the area of online advertising. As a credit allocation solution in cooperative
game theory, Shapley value method directly quantifies the contribution of
online advertising inputs to the advertising key performance indicator (KPI)
across multiple channels. We simplify its calculation by developing an
alternative mathematical formulation. The new formula significantly improves
the computational efficiency and therefore extends the scope of applicability.
Based on the simplified formula, we further develop the ordered Shapley value
method. The proposed method is able to take into account the order of channels
visited by users. We claim that it provides a more comprehensive insight by
evaluating the attribution of channels at different stages of user conversion
journeys. The proposed approaches are illustrated using a real-world online
advertising campaign dataset. | Shapley Value Methods for Attribution Modeling in Online Advertising | 2018-04-15 12:19:25 | Kaifeng Zhao, Seyed Hanif Mahboobi, Saeed R. Bagheri | http://arxiv.org/abs/1804.05327v1, http://arxiv.org/pdf/1804.05327v1 | econ.EM |
28,835 | em | To estimate the dynamic effects of an absorbing treatment, researchers often
use two-way fixed effects regressions that include leads and lags of the
treatment. We show that in settings with variation in treatment timing across
units, the coefficient on a given lead or lag can be contaminated by effects
from other periods, and apparent pretrends can arise solely from treatment
effects heterogeneity. We propose an alternative estimator that is free of
contamination, and illustrate the relative shortcomings of two-way fixed
effects regressions with leads and lags through an empirical application. | Estimating Dynamic Treatment Effects in Event Studies with Heterogeneous Treatment Effects | 2018-04-16 19:54:46 | Liyang Sun, Sarah Abraham | http://arxiv.org/abs/1804.05785v2, http://arxiv.org/pdf/1804.05785v2 | econ.EM |
28,836 | em | This paper offers a two-pronged critique of the empirical investigation of
the income distribution performed by physicists over the past decade. Their
finding rely on the graphical analysis of the observed distribution of
normalized incomes. Two central observations lead to the conclusion that the
majority of incomes are exponentially distributed, but neither each individual
piece of evidence nor their concurrent observation robustly proves that the
thermal and superthermal mixture fits the observed distribution of incomes
better than reasonable alternatives. A formal analysis using popular measures
of fit shows that while an exponential distribution with a power-law tail
provides a better fit of the IRS income data than the log-normal distribution
(often assumed by economists), the thermal and superthermal mixture's fit can
be improved upon further by adding a log-normal component. The economic
implications of the thermal and superthermal distribution of incomes, and the
expanded mixture are explored in the paper. | Revisiting the thermal and superthermal two-class distribution of incomes: A critical perspective | 2018-04-17 19:09:59 | Markus P. A. Schneider | http://dx.doi.org/10.1140/epjb/e2014-50501-x, http://arxiv.org/abs/1804.06341v1, http://arxiv.org/pdf/1804.06341v1 | econ.EM |
28,837 | em | Researchers increasingly leverage movement across multiple treatments to
estimate causal effects. While these "mover regressions" are often motivated by
a linear constant-effects model, it is not clear what they capture under weaker
quasi-experimental assumptions. I show that binary treatment mover regressions
recover a convex average of four difference-in-difference comparisons and are
thus causally interpretable under a standard parallel trends assumption.
Estimates from multiple-treatment models, however, need not be causal without
stronger restrictions on the heterogeneity of treatment effects and
time-varying shocks. I propose a class of two-step estimators to isolate and
combine the large set of difference-in-difference quasi-experiments generated
by a mover design, identifying mover average treatment effects under
conditional-on-covariate parallel trends and effect homogeneity restrictions. I
characterize the efficient estimators in this class and derive specification
tests based on the model's overidentifying restrictions. Future drafts will
apply the theory to the Finkelstein et al. (2016) movers design, analyzing the
causal effects of geography on healthcare utilization. | Estimating Treatment Effects in Mover Designs | 2018-04-18 16:42:55 | Peter Hull | http://arxiv.org/abs/1804.06721v1, http://arxiv.org/pdf/1804.06721v1 | econ.EM |
28,838 | em | The study aims to identify the institutional flaws of the current EU waste
management model by analysing the economic model of extended producer
responsibility and collective waste management systems and to create a model
for measuring the transaction costs borne by waste recovery organizations. The
model was approbated by analysing the Bulgarian collective waste management
systems that have been complying with the EU legislation for the last 10 years.
The analysis focuses on waste oils because of their economic importance and the
limited number of studies and analyses in this field as the predominant body of
research to date has mainly addressed packaging waste, mixed household waste or
discarded electrical and electronic equipment. The study aims to support the
process of establishing a circular economy in the EU, which was initiated in
2015. | Transaction Costs in Collective Waste Recovery Systems in the EU | 2018-04-18 18:40:15 | Shteryo Nozharov | http://arxiv.org/abs/1804.06792v1, http://arxiv.org/pdf/1804.06792v1 | econ.EM |
28,839 | em | We study the foundations of empirical equilibrium, a refinement of Nash
equilibrium that is based on a non-parametric characterization of empirical
distributions of behavior in games (Velez and Brown,2020b arXiv:1907.12408).
The refinement can be alternatively defined as those Nash equilibria that do
not refute the regular QRE theory of Goeree, Holt, and Palfrey (2005). By
contrast, some empirical equilibria may refute monotone additive randomly
disturbed payoff models. As a by product, we show that empirical equilibrium
does not coincide with refinements based on approximation by monotone additive
randomly disturbed payoff models, and further our understanding of the
empirical content of these models. | Empirical Equilibrium | 2018-04-21 18:38:24 | Rodrigo A. Velez, Alexander L. Brown | http://arxiv.org/abs/1804.07986v3, http://arxiv.org/pdf/1804.07986v3 | econ.EM |
28,840 | em | We analyze an operational policy for a multinational manufacturer to hedge
against exchange rate uncertainties and competition. We consider a single
product and single period. Because of long-lead times, the capacity investment
must done before the selling season begins when the exchange rate between the
two countries is uncertain. we consider a duopoly competition in the foreign
country. We model the exchange rate as a random variable. We investigate the
impact of competition and exchange rate on optimal capacities and optimal
prices. We show how competition can impact the decision of the home
manufacturer to enter the foreign market. | Price Competition with Geometric Brownian motion in Exchange Rate Uncertainty | 2018-04-22 21:33:53 | Murat Erkoc, Huaqing Wang, Anas Ahmed | http://arxiv.org/abs/1804.08153v1, http://arxiv.org/pdf/1804.08153v1 | econ.EM |
28,841 | em | Call centers' managers are interested in obtaining accurate point and
distributional forecasts of call arrivals in order to achieve an optimal
balance between service quality and operating costs. We present a strategy for
selecting forecast models of call arrivals which is based on three pillars: (i)
flexibility of the loss function; (ii) statistical evaluation of forecast
accuracy; (iii) economic evaluation of forecast performance using money
metrics. We implement fourteen time series models and seven forecast
combination schemes on three series of daily call arrivals. Although we focus
mainly on point forecasts, we also analyze density forecast evaluation. We show
that second moments modeling is important both for point and density
forecasting and that the simple Seasonal Random Walk model is always
outperformed by more general specifications. Our results suggest that call
center managers should invest in the use of forecast models which describe both
first and second moments of call arrivals. | Statistical and Economic Evaluation of Time Series Models for Forecasting Arrivals at Call Centers | 2018-04-23 12:57:42 | Andrea Bastianin, Marzio Galeotti, Matteo Manera | http://dx.doi.org/10.1007/s00181-018-1475-y, http://arxiv.org/abs/1804.08315v1, http://arxiv.org/pdf/1804.08315v1 | econ.EM |
28,842 | em | Economic inequality is one of the pivotal issues for most of economic and
social policy makers across the world to insure the sustainable economic growth
and justice. In the mainstream school of economics, namely neoclassical
theories, economic issues are dealt with in a mechanistic manner. Such a
mainstream framework is majorly focused on investigating a socio-economic
system based on an axiomatic scheme where reductionism approach plays a vital
role. The major limitations of such theories include unbounded rationality of
economic agents, reducing the economic aggregates to a set of predictable
factors and lack of attention to adaptability and the evolutionary nature of
economic agents. In tackling deficiencies of conventional economic models, in
the past two decades, some new approaches have been recruited. One of those
novel approaches is the Complex adaptive systems (CAS) framework which has
shown a very promising performance in action. In contrast to mainstream school,
under this framework, the economic phenomena are studied in an organic manner
where the economic agents are supposed to be both boundedly rational and
adaptive. According to it, the economic aggregates emerge out of the ways
agents of a system decide and interact. As a powerful way of modeling CASs,
Agent-based models (ABMs) has found a growing application among academicians
and practitioners. ABMs show that how simple behavioral rules of agents and
local interactions among them at micro-scale can generate surprisingly complex
patterns at macro-scale. In this paper, ABMs have been used to show (1) how an
economic inequality emerges in a system and to explain (2) how sadaqah as an
Islamic charity rule can majorly help alleviating the inequality and how
resource allocation strategies taken by charity entities can accelerate this
alleviation. | Economic inequality and Islamic Charity: An exploratory agent-based modeling approach | 2018-04-25 01:43:11 | Hossein Sabzian, Alireza Aliahmadi, Adel Azar, Madjid Mirzaee | http://arxiv.org/abs/1804.09284v1, http://arxiv.org/pdf/1804.09284v1 | econ.EM |
28,843 | em | This paper is concerned with inference about low-dimensional components of a
high-dimensional parameter vector $\beta^0$ which is identified through
instrumental variables. We allow for eigenvalues of the expected outer product
of included and excluded covariates, denoted by $M$, to shrink to zero as the
sample size increases. We propose a novel estimator based on desparsification
of an instrumental variable Lasso estimator, which is a regularized version of
2SLS with an additional correction term. This estimator converges to $\beta^0$
at a rate depending on the mapping properties of $M$ captured by a sparse link
condition. Linear combinations of our estimator of $\beta^0$ are shown to be
asymptotically normally distributed. Based on consistent covariance estimation,
our method allows for constructing confidence intervals and statistical tests
for single or low-dimensional components of $\beta^0$. In Monte-Carlo
simulations we analyze the finite sample behavior of our estimator. | Ill-posed Estimation in High-Dimensional Models with Instrumental Variables | 2018-06-02 19:41:24 | Christoph Breunig, Enno Mammen, Anna Simoni | http://arxiv.org/abs/1806.00666v2, http://arxiv.org/pdf/1806.00666v2 | econ.EM |
28,863 | em | By recasting indirect inference estimation as a prediction rather than a
minimization and by using regularized regressions, we can bypass the three
major problems of estimation: selecting the summary statistics, defining the
distance function and minimizing it numerically. By substituting regression
with classification we can extend this approach to model selection as well. We
present three examples: a statistical fit, the parametrization of a simple real
business cycle model and heuristics selection in a fishery agent-based model.
The outcome is a method that automatically chooses summary statistics, weighs
them and use them to parametrize models without running any direct
minimization. | Indirect inference through prediction | 2018-07-04 16:52:24 | Ernesto Carrella, Richard M. Bailey, Jens Koed Madsen | http://dx.doi.org/10.18564/jasss.4150, http://arxiv.org/abs/1807.01579v1, http://arxiv.org/pdf/1807.01579v1 | econ.EM |
28,844 | em | I propose a nonparametric iid bootstrap procedure for the empirical
likelihood, the exponential tilting, and the exponentially tilted empirical
likelihood estimators that achieves asymptotic refinements for t tests and
confidence intervals, and Wald tests and confidence regions based on such
estimators. Furthermore, the proposed bootstrap is robust to model
misspecification, i.e., it achieves asymptotic refinements regardless of
whether the assumed moment condition model is correctly specified or not. This
result is new, because asymptotic refinements of the bootstrap based on these
estimators have not been established in the literature even under correct model
specification. Monte Carlo experiments are conducted in dynamic panel data
setting to support the theoretical finding. As an application, bootstrap
confidence intervals for the returns to schooling of Hellerstein and Imbens
(1999) are calculated. The result suggests that the returns to schooling may be
higher. | Asymptotic Refinements of a Misspecification-Robust Bootstrap for Generalized Empirical Likelihood Estimators | 2018-06-04 07:54:48 | Seojeong Lee | http://dx.doi.org/10.1016/j.jeconom.2015.11.003, http://arxiv.org/abs/1806.00953v2, http://arxiv.org/pdf/1806.00953v2 | econ.EM |
28,845 | em | Many studies use shift-share (or ``Bartik'') instruments, which average a set
of shocks with exposure share weights. We provide a new econometric framework
for shift-share instrumental variable (SSIV) regressions in which
identification follows from the quasi-random assignment of shocks, while
exposure shares are allowed to be endogenous. The framework is motivated by an
equivalence result: the orthogonality between a shift-share instrument and an
unobserved residual can be represented as the orthogonality between the
underlying shocks and a shock-level unobservable. SSIV regression coefficients
can similarly be obtained from an equivalent shock-level regression, motivating
shock-level conditions for their consistency. We discuss and illustrate several
practical insights of this framework in the setting of Autor et al. (2013),
estimating the effect of Chinese import competition on manufacturing employment
across U.S. commuting zones. | Quasi-Experimental Shift-Share Research Designs | 2018-06-04 20:03:07 | Kirill Borusyak, Peter Hull, Xavier Jaravel | http://arxiv.org/abs/1806.01221v9, http://arxiv.org/pdf/1806.01221v9 | econ.EM |
28,846 | em | The implementation of a supervision and incentive process for identical
workers may lead to wage variance that stems from employer and employee
optimization. The harder it is to assess the nature of the labor output, the
more important such a process becomes, and the influence of such a process on
wage development growth. The dynamic model presented in this paper shows that
an employer will choose to pay a worker a starting wage that is less than what
he deserves, resulting in a wage profile that fits the classic profile in the
human-capital literature. The wage profile and wage variance rise at times of
technological advancements, which leads to increased turnover as older workers
are replaced by younger workers due to a rise in the relative marginal cost of
the former. | The Impact of Supervision and Incentive Process in Explaining Wage Profile and Variance | 2018-06-04 22:05:37 | Nitsa Kasir, Idit Sohlberg | http://arxiv.org/abs/1806.01332v1, http://arxiv.org/pdf/1806.01332v1 | econ.EM |
28,847 | em | I propose a nonparametric iid bootstrap that achieves asymptotic refinements
for t tests and confidence intervals based on GMM estimators even when the
model is misspecified. In addition, my bootstrap does not require recentering
the moment function, which has been considered as critical for GMM. Regardless
of model misspecification, the proposed bootstrap achieves the same sharp
magnitude of refinements as the conventional bootstrap methods which establish
asymptotic refinements by recentering in the absence of misspecification. The
key idea is to link the misspecified bootstrap moment condition to the large
sample theory of GMM under misspecification of Hall and Inoue (2003). Two
examples are provided: Combining data sets and invalid instrumental variables. | Asymptotic Refinements of a Misspecification-Robust Bootstrap for Generalized Method of Moments Estimators | 2018-06-05 04:13:06 | Seojeong Lee | http://dx.doi.org/10.1016/j.jeconom.2013.05.008, http://arxiv.org/abs/1806.01450v1, http://arxiv.org/pdf/1806.01450v1 | econ.EM |
28,848 | em | Under treatment effect heterogeneity, an instrument identifies the
instrument-specific local average treatment effect (LATE). With multiple
instruments, two-stage least squares (2SLS) estimand is a weighted average of
different LATEs. What is often overlooked in the literature is that the
postulated moment condition evaluated at the 2SLS estimand does not hold unless
those LATEs are the same. If so, the conventional heteroskedasticity-robust
variance estimator would be inconsistent, and 2SLS standard errors based on
such estimators would be incorrect. I derive the correct asymptotic
distribution, and propose a consistent asymptotic variance estimator by using
the result of Hall and Inoue (2003, Journal of Econometrics) on misspecified
moment condition models. This can be used to correctly calculate the standard
errors regardless of whether there is more than one LATE or not. | A Consistent Variance Estimator for 2SLS When Instruments Identify Different LATEs | 2018-06-05 04:36:49 | Seojeong Lee | http://dx.doi.org/10.1080/07350015.2016.1186555, http://arxiv.org/abs/1806.01457v1, http://arxiv.org/pdf/1806.01457v1 | econ.EM |
28,849 | em | We propose leave-out estimators of quadratic forms designed for the study of
linear models with unrestricted heteroscedasticity. Applications include
analysis of variance and tests of linear restrictions in models with many
regressors. An approximation algorithm is provided that enables accurate
computation of the estimator in very large datasets. We study the large sample
properties of our estimator allowing the number of regressors to grow in
proportion to the number of observations. Consistency is established in a
variety of settings where plug-in methods and estimators predicated on
homoscedasticity exhibit first-order biases. For quadratic forms of increasing
rank, the limiting distribution can be represented by a linear combination of
normal and non-central $\chi^2$ random variables, with normality ensuing under
strong identification. Standard error estimators are proposed that enable tests
of linear restrictions and the construction of uniformly valid confidence
intervals for quadratic forms of interest. We find in Italian social security
records that leave-out estimates of a variance decomposition in a two-way fixed
effects model of wage determination yield substantially different conclusions
regarding the relative contribution of workers, firms, and worker-firm sorting
to wage inequality than conventional methods. Monte Carlo exercises corroborate
the accuracy of our asymptotic approximations, with clear evidence of
non-normality emerging when worker mobility between blocks of firms is limited. | Leave-out estimation of variance components | 2018-06-05 07:59:27 | Patrick Kline, Raffaele Saggio, Mikkel Sølvsten | http://arxiv.org/abs/1806.01494v2, http://arxiv.org/pdf/1806.01494v2 | econ.EM |
28,850 | em | Autonomous ships (AS) used for cargo transport have gained a considerable
amount of attention in recent years. They promise benefits such as reduced crew
costs, increased safety and increased flexibility. This paper explores the
effects of a faster increase in technological performance in maritime shipping
achieved by leveraging fast-improving technological domains such as computer
processors, and advanced energy storage. Based on historical improvement rates
of several modes of transport (Cargo Ships, Air, Rail, Trucking) a simplified
Markov-chain Monte-Carlo (MCMC) simulation of an intermodal transport model
(IMTM) is used to explore the effects of differing technological improvement
rates for AS. The results show that the annual improvement rates of traditional
shipping (Ocean Cargo Ships = 2.6%, Air Cargo = 5.5%, Trucking = 0.6%, Rail =
1.9%, Inland Water Transport = 0.4%) improve at lower rates than technologies
associated with automation such as Computer Processors (35.6%), Fuel Cells
(14.7%) and Automotive Autonomous Hardware (27.9%). The IMTM simulations up to
the year 2050 show that the introduction of any mode of autonomous transport
will increase competition in lower cost shipping options, but is unlikely to
significantly alter the overall distribution of transport mode costs. Secondly,
if all forms of transport end up converting to autonomous systems, then the
uncertainty surrounding the improvement rates yields a complex intermodal
transport solution involving several options, all at a much lower cost over
time. Ultimately, the research shows a need for more accurate measurement of
current autonomous transport costs and how they are changing over time. | A Quantitative Analysis of Possible Futures of Autonomous Transport | 2018-06-05 17:00:58 | Christopher L. Benson, Pranav D Sumanth, Alina P Colling | http://arxiv.org/abs/1806.01696v1, http://arxiv.org/pdf/1806.01696v1 | econ.EM |
28,851 | em | A standard growth model is modified in a straightforward way to incorporate
what Keynes (1936) suggests in the "essence" of his general theory. The
theoretical essence is the idea that exogenous changes in investment cause
changes in employment and unemployment. We implement this idea by assuming the
path for capital growth rate is exogenous in the growth model. The result is a
growth model that can explain both long term trends and fluctuations around the
trend. The modified growth model was tested using the U.S. economic data from
1947 to 2014. The hypothesized inverse relationship between the capital growth
and changes in unemployment was confirmed, and the structurally estimated model
fits fluctuations in unemployment reasonably well. | A Growth Model with Unemployment | 2018-06-11 23:29:04 | Mina Mahmoudi, Mark Pingle | http://arxiv.org/abs/1806.04228v1, http://arxiv.org/pdf/1806.04228v1 | econ.EM |
28,852 | em | This study provides the theoretical framework and empirical model for
productivity growth evaluations in agricultural sector as one of the most
important sectors in Iran's economic development plan. We use the Solow
residual model to measure the productivity growth share in the value-added
growth of the agricultural sector. Our time series data includes value-added
per worker, employment, and capital in this sector. The results show that the
average total factor productivity growth rate in the agricultural sector is
-0.72% during 1991-2010. Also, during this period, the share of total factor
productivity growth in the value-added growth is -19.6%, while it has been
forecasted to be 33.8% in the fourth development plan. Considering the
effective role of capital in the agricultural low productivity, we suggest
applying productivity management plans (especially in regards of capital
productivity) to achieve future growth goals. | The Role of Agricultural Sector Productivity in Economic Growth: The Case of Iran's Economic Development Plan | 2018-06-11 23:43:32 | Morteza Tahamipour, Mina Mahmoudi | http://arxiv.org/abs/1806.04235v1, http://arxiv.org/pdf/1806.04235v1 | econ.EM |
28,853 | em | Tariff liberalization and its impact on tax revenue is an important
consideration for developing countries, because they are increasingly facing
the difficult task of implementing and harmonizing regional and international
trade commitments. The tariff reform and its costs for Iranian government is
one of the issues that are examined in this study. Another goal of this paper
is, estimating the cost of trade liberalization. On this regard, imports value
of agricultural sector in Iran in 2010 was analyzed according to two scenarios.
For reforming nuisance tariff, a VAT policy is used in both scenarios. In this
study, TRIST method is used. In the first scenario, imports' value decreased to
a level equal to the second scenario and higher tariff revenue will be created.
The results show that reducing the average tariff rate does not always result
in the loss of tariff revenue. This paper is a witness that different forms of
tariff can generate different amount of income when they have same level of
liberalization and equal effect on producers. Therefore, using a good tariff
regime can help a government to generate income when increases social welfare
by liberalization. | Estimating Trade-Related Adjustment Costs in the Agricultural Sector in Iran | 2018-06-11 23:44:02 | Omid Karami, Mina Mahmoudi | http://arxiv.org/abs/1806.04238v1, http://arxiv.org/pdf/1806.04238v1 | econ.EM |
28,854 | em | We consider the relation between Sion's minimax theorem for a continuous
function and a Nash equilibrium in an asymmetric multi-players zero-sum game in
which only one player is different from other players, and the game is
symmetric for the other players. Then,
1. The existence of a Nash equilibrium, which is symmetric for players other
than one player, implies Sion's minimax theorem for pairs of this player and
one of other players with symmetry for the other players.
2. Sion's minimax theorem for pairs of one player and one of other players
with symmetry for the other players implies the existence of a Nash equilibrium
which is symmetric for the other players.
Thus, they are equivalent. | On the relation between Sion's minimax theorem and existence of Nash equilibrium in asymmetric multi-players zero-sum game with only one alien | 2018-06-17 04:11:55 | Atsuhiro Satoh, Yasuhito Tanaka | http://arxiv.org/abs/1806.07253v1, http://arxiv.org/pdf/1806.07253v1 | econ.EM |
28,855 | em | It is common practice in empirical work to employ cluster-robust standard
errors when using the linear regression model to estimate some
structural/causal effect of interest. Researchers also often include a large
set of regressors in their model specification in order to control for observed
and unobserved confounders. In this paper we develop inference methods for
linear regression models with many controls and clustering. We show that
inference based on the usual cluster-robust standard errors by Liang and Zeger
(1986) is invalid in general when the number of controls is a non-vanishing
fraction of the sample size. We then propose a new clustered standard errors
formula that is robust to the inclusion of many controls and allows to carry
out valid inference in a variety of high-dimensional linear regression models,
including fixed effects panel data models and the semiparametric partially
linear model. Monte Carlo evidence supports our theoretical results and shows
that our proposed variance estimator performs well in finite samples. The
proposed method is also illustrated with an empirical application that
re-visits Donohue III and Levitt's (2001) study of the impact of abortion on
crime. | Cluster-Robust Standard Errors for Linear Regression Models with Many Controls | 2018-06-19 18:48:50 | Riccardo D'Adamo | http://arxiv.org/abs/1806.07314v3, http://arxiv.org/pdf/1806.07314v3 | econ.EM |
28,856 | em | We study inference in shift-share regression designs, such as when a regional
outcome is regressed on a weighted average of sectoral shocks, using regional
sector shares as weights. We conduct a placebo exercise in which we estimate
the effect of a shift-share regressor constructed with randomly generated
sectoral shocks on actual labor market outcomes across U.S. Commuting Zones.
Tests based on commonly used standard errors with 5\% nominal significance
level reject the null of no effect in up to 55\% of the placebo samples. We use
a stylized economic model to show that this overrejection problem arises
because regression residuals are correlated across regions with similar
sectoral shares, independently of their geographic location. We derive novel
inference methods that are valid under arbitrary cross-regional correlation in
the regression residuals. We show using popular applications of shift-share
designs that our methods may lead to substantially wider confidence intervals
in practice. | Shift-Share Designs: Theory and Inference | 2018-06-20 21:57:10 | Rodrigo Adão, Michal Kolesár, Eduardo Morales | http://dx.doi.org/10.1093/qje/qjz025, http://arxiv.org/abs/1806.07928v5, http://arxiv.org/pdf/1806.07928v5 | econ.EM |
28,857 | em | In this paper, we explore the relationship between state-level household
income inequality and macroeconomic uncertainty in the United States. Using a
novel large-scale macroeconometric model, we shed light on regional disparities
of inequality responses to a national uncertainty shock. The results suggest
that income inequality decreases in most states, with a pronounced degree of
heterogeneity in terms of shapes and magnitudes of the dynamic responses. By
contrast, some few states, mostly located in the West and South census region,
display increasing levels of income inequality over time. We find that this
directional pattern in responses is mainly driven by the income composition and
labor market fundamentals. In addition, forecast error variance decompositions
allow for a quantitative assessment of the importance of uncertainty shocks in
explaining income inequality. The findings highlight that volatility shocks
account for a considerable fraction of forecast error variance for most states
considered. Finally, a regression-based analysis sheds light on the driving
forces behind differences in state-specific inequality responses. | The transmission of uncertainty shocks on income inequality: State-level evidence from the United States | 2018-06-21 17:57:45 | Manfred M. Fischer, Florian Huber, Michael Pfarrhofer | http://arxiv.org/abs/1806.08278v1, http://arxiv.org/pdf/1806.08278v1 | econ.EM |
28,858 | em | We propose a new class of unit root tests that exploits invariance properties
in the Locally Asymptotically Brownian Functional limit experiment associated
to the unit root model. The invariance structures naturally suggest tests that
are based on the ranks of the increments of the observations, their average,
and an assumed reference density for the innovations. The tests are
semiparametric in the sense that they are valid, i.e., have the correct
(asymptotic) size, irrespective of the true innovation density. For a correctly
specified reference density, our test is point-optimal and nearly efficient.
For arbitrary reference densities, we establish a Chernoff-Savage type result,
i.e., our test performs as well as commonly used tests under Gaussian
innovations but has improved power under other, e.g., fat-tailed or skewed,
innovation distributions. To avoid nonparametric estimation, we propose a
simplified version of our test that exhibits the same asymptotic properties,
except for the Chernoff-Savage result that we are only able to demonstrate by
means of simulations. | Semiparametrically Point-Optimal Hybrid Rank Tests for Unit Roots | 2018-06-25 10:03:48 | Bo Zhou, Ramon van den Akker, Bas J. M. Werker | http://dx.doi.org/10.1214/18-AOS1758, http://arxiv.org/abs/1806.09304v1, http://arxiv.org/pdf/1806.09304v1 | econ.EM |
28,859 | em | In this article we introduce a general nonparametric point-identification
result for nonseparable triangular models with a multivariate first- and second
stage. Based on this we prove point-identification of Hedonic models with
multivariate heterogeneity and endogenous observable characteristics, extending
and complementing identification results from the literature which all require
exogeneity. As an additional application of our theoretical result, we show
that the BLP model (Berry et al. 1995) can also be identified without index
restrictions. | Point-identification in multivariate nonseparable triangular models | 2018-06-25 22:36:39 | Florian Gunsilius | http://arxiv.org/abs/1806.09680v1, http://arxiv.org/pdf/1806.09680v1 | econ.EM |
28,860 | em | Historical examination of the Bretton Woods system allows comparisons to be
made with the current evolution of the EMS. | The Bretton Woods Experience and ERM | 2018-07-02 03:00:20 | Chris Kirrane | http://arxiv.org/abs/1807.00418v1, http://arxiv.org/pdf/1807.00418v1 | econ.EM |
28,861 | em | This paper describes the opportunities and also the difficulties of EMU with
regard to international monetary cooperation. Even though the institutional and
intellectual assistance to the coordination of monetary policy in the EU will
probably be strengthened with the EMU, among the shortcomings of the Maastricht
Treaty concerns the relationship between the founder members and those
countries who wish to remain outside monetary union. | Maastricht and Monetary Cooperation | 2018-07-02 03:01:08 | Chris Kirrane | http://arxiv.org/abs/1807.00419v1, http://arxiv.org/pdf/1807.00419v1 | econ.EM |
28,862 | em | This paper proposes a hierarchical modeling approach to perform stochastic
model specification in Markov switching vector error correction models. We
assume that a common distribution gives rise to the regime-specific regression
coefficients. The mean as well as the variances of this distribution are
treated as fully stochastic and suitable shrinkage priors are used. These
shrinkage priors enable to assess which coefficients differ across regimes in a
flexible manner. In the case of similar coefficients, our model pushes the
respective regions of the parameter space towards the common distribution. This
allows for selecting a parsimonious model while still maintaining sufficient
flexibility to control for sudden shifts in the parameters, if necessary. We
apply our modeling approach to real-time Euro area data and assume transition
probabilities between expansionary and recessionary regimes to be driven by the
cointegration errors. The results suggest that the regime allocation is
governed by a subset of short-run adjustment coefficients and regime-specific
variance-covariance matrices. These findings are complemented by an
out-of-sample forecast exercise, illustrating the advantages of the model for
predicting Euro area inflation in real time. | Stochastic model specification in Markov switching vector error correction models | 2018-07-02 11:36:11 | Niko Hauzenberger, Florian Huber, Michael Pfarrhofer, Thomas O. Zörner | http://arxiv.org/abs/1807.00529v2, http://arxiv.org/pdf/1807.00529v2 | econ.EM |
28,864 | em | This paper studies the identifying content of the instrument monotonicity
assumption of Imbens and Angrist (1994) on the distribution of potential
outcomes in a model with a binary outcome, a binary treatment and an exogenous
binary instrument. Specifically, I derive necessary and sufficient conditions
on the distribution of the data under which the identified set for the
distribution of potential outcomes when the instrument monotonicity assumption
is imposed can be a strict subset of that when it is not imposed. | On the Identifying Content of Instrument Monotonicity | 2018-07-04 19:25:35 | Vishal Kamat | http://arxiv.org/abs/1807.01661v2, http://arxiv.org/pdf/1807.01661v2 | econ.EM |
28,865 | em | This paper develops an inferential theory for state-varying factor models of
large dimensions. Unlike constant factor models, loadings are general functions
of some recurrent state process. We develop an estimator for the latent factors
and state-varying loadings under a large cross-section and time dimension. Our
estimator combines nonparametric methods with principal component analysis. We
derive the rate of convergence and limiting normal distribution for the
factors, loadings and common components. In addition, we develop a statistical
test for a change in the factor structure in different states. We apply the
estimator to U.S. Treasury yields and S&P500 stock returns. The systematic
factor structure in treasury yields differs in times of booms and recessions as
well as in periods of high market volatility. State-varying factors based on
the VIX capture significantly more variation and pricing information in
individual stocks than constant factor models. | State-Varying Factor Models of Large Dimensions | 2018-07-06 07:05:40 | Markus Pelger, Ruoxuan Xiong | http://arxiv.org/abs/1807.02248v4, http://arxiv.org/pdf/1807.02248v4 | econ.EM |
28,866 | em | The methods of new institutional economics for identifying the transaction
costs of trade litigations in Bulgaria are used in the current paper. For the
needs of the research, an indicative model, measuring this type of costs on
microeconomic level, is applied in the study. The main purpose of the model is
to forecast the rational behavior of trade litigation parties in accordance
with the transaction costs in the process of enforcing the execution of the
signed commercial contract. The application of the model is related to the more
accurate measurement of the transaction costs on microeconomic level, which
fact could lead to better prediction and management of these costs in order
market efficiency and economic growth to be achieved. In addition, it is made
an attempt to be analysed the efficiency of the institutional change of the
commercial justice system and the impact of the reform of the judicial system
over the economic turnover. The augmentation or lack of reduction of the
transaction costs in trade litigations would mean inefficiency of the reform of
the judicial system. JEL Codes: O43, P48, D23, K12 | Transaction costs and institutional change of trade litigations in Bulgaria | 2018-07-09 13:34:56 | Shteryo Nozharov, Petya Koralova-Nozharova | http://arxiv.org/abs/1807.03034v1, http://arxiv.org/pdf/1807.03034v1 | econ.EM |
Dataset Card for Economics Paper Dataset
Dataset Summary
The Economics Research Paper Dataset was designed to support the development of the LLaMA-2-Econ models, with a focus on Title Generation, Abstract Classification, and Question & Answer (Q&A) tasks. It comprises abstracts and titles of economics research papers, along with synthetic Q&A pairs derived from the abstracts, to facilitate training of large language models for economics-specific applications.
Dataset Description
Content: The dataset includes:
- Economics paper abstracts and titles.
Source: The data was collected using the arXiv API, with papers selected from the categories Econometrics (ec.EM), General Economics (ec.GN), and Theoretical Economics (ec.TH).
Volume:
- Total abstracts and titles: 6362
Intended Uses
This dataset is intended for training and evaluating language models specialized in:
- Generating titles for economics research papers.
- Classifying abstracts into sub-fields of economics.
- Answering questions based on economics paper abstracts.
Dataset Creation
Curation Rationale
The dataset was curated to address the lack of specialized tools and datasets for enhancing research within the economics domain, leveraging the potential of language models like LLaMA-2.
Source Data
Initial Data Collection and Normalization
Data was collected through the arXiv API, targeting papers within specified categories of economics. Titles and abstracts were extracted, and synthetic Q&A pairs were generated using a process that involved the GPT-3.5 Turbo model for contextual dialogue creation.
Licensing Information
The dataset is derived from arXiv papers. Users are advised to adhere to arXiv's terms of use.
- Downloads last month
- 38