일 | 월 | 화 | 수 | 목 | 금 | 토 |
---|---|---|---|---|---|---|
1 | 2 | 3 | 4 | 5 | ||
6 | 7 | 8 | 9 | 10 | 11 | 12 |
13 | 14 | 15 | 16 | 17 | 18 | 19 |
20 | 21 | 22 | 23 | 24 | 25 | 26 |
27 | 28 | 29 | 30 | 31 |
- 대학원
- convex
- 카이스트
- lccm
- Python
- 넘파이
- 젠센 부등식
- 볼록 함수
- choice model
- jensen's inequality
- regret-minimization
- 통계
- journal club
- 티스토리챌린지
- python#yaml#가상환경#파이썬
- Closed Form
- em알고리즘#expectation maximization#algorithm
- 연구
- EM 알고리즘
- latex#티스토리#tistory#대학원#논문#논문정리
- 논문 리뷰
- DCM
- Expectation Maximization
- 옌센 부등식
- 나비에 스토크스
- 윤리 및 안전
- discrete choice model
- 안전교육
- numpy
- 닫힌 해
- Today
- Total
대학원생 리암의 블로그
[Paper Review] Estimating Standard Errors in Finance Panel Data Sets : Comparing Approaches 본문
[Paper Review] Estimating Standard Errors in Finance Panel Data Sets : Comparing Approaches
liam0222 2025. 3. 28. 11:36
Overview
This paper deals with the issue of standard errors that frequently arise in empirical finance research. In essence, when there is dependency (correlation) in the data, the effective sample size diminishes, leading to an underestimation of the standard error. Consequently, this can cause inflated t-statistics, higher rejection rates, and an overstatement of research significance.
Review of Basic Preliminaries and Background Knowledge
Panel data refers to datasets that capture observations on multiple entities over several time periods. Standard error is a measure of the variability or precision of a sample statistic, indicating how much the estimated statistic would likely vary if the study were repeated with different samples. It is important to distinguish between standard deviation and standard error: the former measures variability within a sample, whereas the latter measures variability across samples. When moving to a panel data setting, the notion of effective sample size becomes critical. Although the total number of observations is \(N \times T\), where \(N\) is the number of entities and \(T\) is the number of time periods, the effective sample size is often smaller than \(N \times T\). This reduction occurs because observations within the same entity are typically correlated, meaning that each additional time period provides less independent information.
Mathematical Background
Linear Model and Variance of the Estimator
Consider a simple linear model. The Ordinary Least Squares (OLS) estimator is derived by minimizing the mean squared error of the predicted \(Y\), yielding : \( \hat{\beta} = \beta + \text{(error term)} \).
The variance of \( \hat{\beta} \) under i.i.d. assumptions is: \( \mathrm{Var}(\hat{\beta}) \propto \frac{\sigma_\epsilon^2}{\sigma_x^2 \times NT} \).
As \( \sigma_\epsilon^2 \) (the variance of the error term) increases, the variance of \( \hat{\beta}\) increases, and as the sample size \( NT \) grows, the variance of \( \hat{\beta} \) decreases. The inverse relationship with \( \sigma_x^2 \) arises because a larger spread of \( x \)-values yields more information for estimating \( \beta \).
Firm Fixed Effect
The error term can be decomposed into a firm-specific component \( \gamma_i \) and an idiosyncratic term \( \eta_{i,t} \). Similarly, the independent variable can be decomposed into a firm-specific component \( \mu_i \) and the remaining variation \( \nu_{i,t} \). The fractions of these variances due to the firm-specific component are denoted by \( \rho_x \) and \( \rho_\epsilon \). When these components are present, the variance of the OLS estimator rises. However, if at least one of the \( \rho \) values is zero, the variance matches the i.i.d. case. Simulations show that clustered standard errors (by firm) more accurately reflect the true standard error compared to OLS standard errors.
Fama--MacBeth Estimator
The Fama--MacBeth approach involves running cross-sectional regressions in each period and then averaging over time. This method partially mitigates time-varying fluctuations but can underestimate standard errors when there are firm-specific effects. The firm-specific components can cancel out in the variance calculation, leading to an underestimation of the standard error.
Newey--West Variance Estimator
The Newey--West estimator is used to adjust for autocorrelation. For each lag \( j \), the covariance term \(\varepsilon_t \varepsilon_{t-j} \) is multiplied by a weight \( \left[ 1 - \frac{j}{M+1} \right] \). This approach addresses autocorrelation but may not fully capture firm-specific effects unless properly adapted.
Time Fixed Effect
Time fixed effects can be analyzed similarly to firm fixed effects by exchanging \( N \) and \( T \). The OLS standard errors remain correct when there is no time effect in either the independent variable \(\mathrm{Var}(\zeta) = 0 \) or the residual \( \mathrm{Var}(\delta) = 0 \). As time effects in the independent variables and residuals grow, the OLS standard errors increasingly underestimate the true standard errors. In cases where both firm and time effects are present, including dummy variables for each time period can capture time effects, while clustering by firm addresses firm-level correlation. Another approach involves combining estimators for firm-level and time-level clustering, then subtracting a “white noise” term to avoid double-counting.
Firm and Time Fixed Effects
In a firm fixed-effects model, certain unobserved characteristics---such as managerial style or corporate culture---are assumed to remain constant within a firm but may vary across firms. These characteristics can influence both a firm’s leverage (e.g., debt-to-assets ratio) and its explanatory variables (e.g., profitability, firm size, tangibility of assets, and growth opportunities). Ignoring these time-invariant features can lead to omitted variable bias. Time fixed effects capture shocks that are constant among all entities but vary over time, such as macroeconomic events or policy changes. Surveys of empirical finance papers suggest that many studies do not account for these correlations and therefore overstate the statistical significance of their results. Approximately 42% of the studies surveyed did not adjust for this correlation in their standard errors.
Empirical Example
A study examined the determinants of capital structure using NYSE data from 1965 to 2003, regressing the debt to asset ratio on standard firm characteristics (e.g., profitability, firm size).
The results indicated the following:
- Clustered by Firm : Standard errors clustered by firm were significantly larger (3.1--3.5 times) than White standard errors. This expansion is partly due to high autocorrelation in both the firm’s profit margin and the residual.
- Clustered by Time : Time clustering had a smaller effect, indicating limited time-series correlation in the data.
- Clustered by Firm and Time : Standard errors were almost identical to those clustered by firm alone, suggesting that firm-level autocorrelation was the dominant factor.
These findings highlight the importance of choosing an appropriate clustering method based on the data structure. Neglecting within-cluster correlation can lead to underestimation of standard errors and inflated t-statistics.
Conclusion
Panel data often exhibit within-firm and/or within-time correlations, which reduce the effective sample size and lead to downward-biased standard errors when ignored. Such biases inflate t-statistics and overstate the significance of empirical results. In practice, one must carefully consider the nature of the data—whether firm, time, or both types of effects are present—to choose the most appropriate method for estimating standard errors. When correlations are not addressed, empirical findings may be misleading.
Researchers should adopt robust approaches such as clustering standard errors, using Fama--MacBeth with corrections for firm-specific effects, or employing GLS with a properly specified variance-covariance matrix. Temporary effects that decay over time can also be modeled through AR(1) processes. Overall, careful handling of correlations and robust estimation techniques are essential for producing reliable empirical findings.
Appendix
Generalized Least Squares (GLS)
GLS accounts for the known correlation structure of the residuals through the variance-covariance matrix \( \Omega \), thereby correcting for both heteroskedasticity and autocorrelation.
Adjusted Fama--MacBeth
This method inflates the standard errors by incorporating the degree of autocorrelation in the estimated coefficients over time, thus addressing firm-specific effects more effectively.
Temporary Time Effects
In many real-world datasets, time effects may dissipate over time. Such temporary effects can be modeled using an AR(1) process, where the residual in period \( t \) depends on the lagged residual from period \( t-1 \), weighted by a parameter \( \phi \). Incorporating temporary firm or time effects generally increases estimated standard errors compared to models that assume fixed effects, as autocorrelation in the residuals introduces additional variability. Simulation results show that ignoring within-cluster correlation leads to higher rejection rates (i.e., overstating statistical significance). When the number of firms increases (and thus the number of time periods decreases), clustering by time becomes more prone to understating standard errors. Conversely, clustering by firm becomes more relevant as the number of firms increases. Clustering by both firm and time consistently maintains lower rejection rates, underscoring the necessity of accounting for both dimensions when they are present.