Goto

Collaborating Authors

 econometrics


ReLU-Based and DNN-Based Generalized Maximum Score Estimators

Chen, Xiaohong, Gao, Wayne Yuan, Wen, Likang

arXiv.org Machine Learning

We propose a new formulation of the maximum score estimator that uses compositions of rectified linear unit (ReLU) functions, instead of indicator functions as in Manski (1975,1985), to encode the sign alignment restrictions. Since the ReLU function is Lipschitz, our new ReLU-based maximum score criterion function is substantially easier to optimize using standard gradient-based optimization pacakges. We also show that our ReLU-based maximum score (RMS) estimator can be generalized to an umbrella framework defined by multi-index single-crossing (MISC) conditions, while the original maximum score estimator cannot be applied. We establish the $n^{-s/(2s+1)}$ convergence rate and asymptotic normality for the RMS estimator under order-$s$ Holder smoothness. In addition, we propose an alternative estimator using a further reformulation of RMS as a special layer in a deep neural network (DNN) architecture, which allows the estimation procedure to be implemented via state-of-the-art software and hardware for DNN.


Robust Inference Methods for Latent Group Panel Models under Possible Group Non-Separation

Akgun, Oguzhan, Okui, Ryo

arXiv.org Machine Learning

This paper presents robust inference methods for general linear hypotheses in linear panel data models with latent group structure in the coefficients. We employ a selective conditional inference approach, deriving the conditional distribution of coefficient estimates given the group structure estimated from the data. Our procedure provides valid inference under possible violations of group separation, where distributional properties of group-specific coefficients remain unestablished. Furthermore, even when group separation does hold, our method demonstrates superior finite-sample properties compared to traditional asymptotic approaches. This improvement stems from our procedure's ability to account for statistical uncertainty in the estimation of group structure. We demonstrate the effectiveness of our approach through Monte Carlo simulations and apply the methods to two datasets on: (i) the relationship between income and democracy, and (ii) the cyclicality of firm-level R&D investment.


Provably Efficient Neural Estimation of Structural Equation Model: An Adversarial Approach

Neural Information Processing Systems

Structural equation models (SEMs) are widely used in sciences, ranging from economics to psychology, to uncover causal relationships underlying a complex system under consideration and estimate structural parameters of interest. We study estimation in a class of generalized SEMs where the object of interest is defined as the solution to a linear operator equation.



Export Reviews, Discussions, Author Feedback and Meta-Reviews

Neural Information Processing Systems

First provide a summary of the paper, and then address the following criteria: Quality, clarity, originality and significance. The paper introduces a GP-Vol model to flexibly capture the time-dependent changes in variance, and develops a new online algorithm for fully Bayesian inference under the model. The paper is clearly written, the developed inference method seems technically sound, and the presented results look promising. My opinion on the model itself, using a non-parametric approach such as using the GP prior on the transition function (as in the paper), seems, though, a bit an obvious way of extending the prior work developed in the finance area. So, I wouldn't put too high grade on the paper in terms of its originality.


Evaluating Large Language Model Capabilities in Assessing Spatial Econometrics Research

Arbia, Giuseppe, Morandini, Luca, Nardelli, Vincenzo

arXiv.org Artificial Intelligence

This paper investigates Large Language Models (LLMs) ability to assess the economic soundness and theoretical consistency of empirical findings in spatial econometrics. We created original and deliberately altered "counterfactual" summaries from 28 published papers (2005-2024), which were evaluated by a diverse set of LLMs. The LLMs provided qualitative assessments and structured binary classifications on variable choice, coefficient plausibility, and publication suitability. The results indicate that while LLMs can expertly assess the coherence of variable choices (with top models like GPT-4o achieving an overall F1 score of 0.87), their performance varies significantly when evaluating deeper aspects such as coefficient plausibility and overall publication suitability. The results further revealed that the choice of LLM, the specific characteristics of the paper and the interaction between these two factors significantly influence the accuracy of the assessment, particularly for nuanced judgments. These findings highlight LLMs' current strengths in assisting with initial, more surface-level checks and their limitations in performing comprehensive, deep economic reasoning, suggesting a potential assistive role in peer review that still necessitates robust human oversight.


Nonparametric Factor Analysis and Beyond

Zheng, Yujia, Liu, Yang, Yao, Jiaxiong, Hu, Yingyao, Zhang, Kun

arXiv.org Machine Learning

Nearly all identifiability results in unsupervised representation learning inspired by, e.g., independent component analysis, factor analysis, and causal representation learning, rely on assumptions of additive independent noise or noiseless regimes. In contrast, we study the more general case where noise can take arbitrary forms, depend on latent variables, and be non-invertibly entangled within a nonlinear function. We propose a general framework for identifying latent variables in the nonparametric noisy settings. We first show that, under suitable conditions, the generative model is identifiable up to certain submanifold indeterminacies even in the presence of non-negligible noise. Furthermore, under the structural or distributional variability conditions, we prove that latent variables of the general nonlinear models are identifiable up to trivial indeterminacies. Based on the proposed theoretical framework, we have also developed corresponding estimation methods and validated them in various synthetic and real-world settings. Interestingly, our estimate of the true GDP growth from alternative measurements suggests more insightful information on the economies than official reports. We expect our framework to provide new insight into how both researchers and practitioners deal with latent variables in real-world scenarios.


Autoencoder Enhanced Realised GARCH on Volatility Forecasting

Zhao, Qianli, Wang, Chao, Gerlach, Richard, Storti, Giuseppe, Zhang, Lingxiang

arXiv.org Artificial Intelligence

Realised volatility has become increasingly prominent in volatility forecasting due to its ability to capture intraday price fluctuations. With a growing variety of realised volatility estimators, each with unique advantages and limitations, selecting an optimal estimator may introduce challenges. In this thesis, aiming to synthesise the impact of various realised volatility measures on volatility forecasting, we propose an extension of the Realised GARCH model that incorporates an autoencoder-generated synthetic realised measure, combining the information from multiple realised measures in a nonlinear manner. Our proposed model extends existing linear methods, such as Principal Component Analysis and Independent Component Analysis, to reduce the dimensionality of realised measures. The empirical evaluation, conducted across four major stock markets from January 2000 to June 2022 and including the period of COVID-19, demonstrates both the feasibility of applying an autoencoder to synthesise volatility measures and the superior effectiveness of the proposed model in one-step-ahead rolling volatility forecasting. The model exhibits enhanced flexibility in parameter estimations across each rolling window, outperforming traditional linear approaches. These findings indicate that nonlinear dimension reduction offers further adaptability and flexibility in improving the synthetic realised measure, with promising implications for future volatility forecasting applications.


Causal Inference Tools for a Better Evaluation of Machine Learning

Soumm, Michaël

arXiv.org Artificial Intelligence

We present a comprehensive framework for applying rigorous statistical techniques from econometrics to analyze and improve machine learning systems. We introduce key statistical methods such as Ordinary Least Squares (OLS) regression, Analysis of Variance (ANOVA), and logistic regression, explaining their theoretical foundations and practical applications in machine learning evaluation. The document serves as a guide for researchers and practitioners, detailing how these techniques can provide deeper insights into model behavior, performance, and fairness. We cover the mathematical principles behind each method, discuss their assumptions and limitations, and provide step-by-step instructions for their implementation. The paper also addresses how to interpret results, emphasizing the importance of statistical significance and effect size. Through illustrative examples, we demonstrate how these tools can reveal subtle patterns and interactions in machine learning models that are not apparent from traditional evaluation metrics. By connecting the fields of econometrics and machine learning, this work aims to equip readers with powerful analytical tools for more rigorous and comprehensive evaluation of AI systems. The framework presented here contributes to developing more robust, interpretable, and fair machine learning technologies.


On LASSO Inference for High Dimensional Predictive Regression

Gao, Zhan, Lee, Ji Hyung, Mei, Ziwei, Shi, Zhentao

arXiv.org Machine Learning

About one century ago, quantitative analysis and forecasting services were made available to businesses and the general public (Dominguez et al., 1988). Irving Fisher (1867-1947) spearheaded this initiative, pioneering data-driven forecasting practices grounded in theoretical foundations. His renowned theories, namely the monetary quantity theory, inflation-deflation theory, and index theory, were supported by statistical evidence. For example, Fisher (1925) was one of the earliest attempts to understand the source of business cycles in association with the price level (inflation), and Fisher (1926) was the first to report "a statistical relation between unemployment and price changes", today better known as the Phillips curve. Statistical inference is crucial for integrating economic theory and empirical data, which has become a tradition upheld by successive generations of researchers. In today's era of big data, we have unprecedented access to a vast amount of digital information about the economy. Recent advancements in statistical inference have uncovered new empirical patterns in prediction practices using datasets with temporal features. This paper aims at a plain quest: in a high dimensional linear predictive regression model when the number of potential regressors is larger than the sample size, how to conduct asymptotically valid statistical inference for a regressor of particular interest. To the best of our knowledge, no paper has solved this question before.