๐Ÿšจ Limited Offer: First 50 users get 500 credits for free โ€” only ... spots left!
Econometrics Flashcards

Free Econometrics flashcards, exportable to Notion

Learn faster with 49 Econometrics flashcards. One-click export to Notion.

Learn fast, memorize everything, master Econometrics. No credit card required.

Want to create flashcards from your own textbooks and notes?

Let AI create automatically flashcards from your own textbooks and notes. Upload your PDF, select the pages you want to memorize fast, and let AI do the rest. One-click export to Notion.

Create Flashcards from my PDFs

Econometrics

49 flashcards

Econometrics is the application of statistical and mathematical methods to economic data for the purpose of testing theories, making forecasts, and evaluating policies.
A regression model is a statistical technique that describes the relationship between a dependent variable and one or more independent variables, allowing you to make predictions or test hypotheses.
The goal of forecasting in econometrics is to make predictions about future economic trends, variables or events based on past data and statistical models.
Explanatory models aim to understand the underlying causes and relationships between variables, while predictive models focus on accurately forecasting the dependent variable without necessarily explaining the causality.
Heteroskedasticity refers to the unequal variance of errors across observations in a regression model, violating the assumption of constant variance required for ordinary least squares estimation.
Multicollinearity occurs when two or more independent variables in a regression model are highly correlated, making it difficult to isolate their individual effects on the dependent variable.
A time series model is a statistical technique used to analyze and forecast data that is collected over time, taking into account trends, cycles, and seasonal patterns.
A panel data model combines cross-sectional data (data collected across individuals or entities) and time series data, while a cross-sectional model analyzes data at a single point in time.
Diagnostic tests in econometrics are used to check if the assumptions of a regression model are met, such as normality, homoskedasticity, and absence of multicollinearity, to ensure valid inferences.
Endogeneity occurs when an independent variable in a regression model is correlated with the error term, leading to biased and inconsistent estimates. It can arise due to omitted variables, measurement error, or simultaneous causality.
A fixed effects model assumes that the unobserved individual-specific effects are correlated with the independent variables, while a random effects model assumes that they are not.
Instrumental variable estimation is a technique used to address endogeneity by finding an external variable (instrument) that is correlated with the endogenous regressor but uncorrelated with the error term.
A dummy variable is a binary variable used in regression models to represent categorical data, taking on values of 0 or 1 to indicate the presence or absence of a particular characteristic.
Hypothesis testing in econometrics is used to determine whether a claim or theory about the relationship between variables is supported by the sample data, based on statistical significance.
A one-tailed test is used when the alternative hypothesis specifies the direction of the effect (e.g., greater than or less than), while a two-tailed test is used when the alternative hypothesis does not specify the direction.
A confidence interval is a range of values that is likely to contain the true population parameter with a certain level of confidence, based on the sample data and statistical assumptions.
A linear model assumes that the relationship between the dependent and independent variables is linear, while a non-linear model allows for more complex, non-linear relationships between the variables.
Stationarity is a property of a time series where the mean, variance, and covariance are constant over time, meaning that the statistical properties of the series do not change with time.
A unit root is a characteristic of a non-stationary time series, where the effects of shocks persist indefinitely, leading to unpredictable long-term behavior.
Cointegration is a statistical property of non-stationary time series, where a linear combination of the series is stationary, indicating a long-run equilibrium relationship between the variables.
Granger causality is a statistical concept used to determine if one time series can be used to forecast another, based on the idea that the cause must precede the effect in time.
A vector autoregressive (VAR) model is a type of time series model that treats all variables as endogenous and models each variable as a function of its own lags and the lags of the other variables.
AIC (Akaike Information Criterion) and BIC (Bayesian Information Criterion) are model selection criteria used to choose the best model among a set of candidates, with BIC imposing a stronger penalty for complexity than AIC.
A logit model is a type of regression model used when the dependent variable is binary (e.g., yes/no, success/failure), modeling the probability of an event occurring as a function of the independent variables.
Sample selection bias occurs when the sample used in an analysis is not representative of the population of interest, leading to biased and inconsistent estimates.
A Chow test is a statistical test used to determine whether the coefficients in a regression model are stable across different subsets of the data, indicating structural change or instability.
A random walk model is a type of time series model where the current value of a variable is equal to its previous value plus a random disturbance, implying that future values cannot be predicted from past values.
A reduced form model estimates the direct relationships between the endogenous and exogenous variables without imposing theoretical restrictions, while a structural model incorporates economic theory and imposes restrictions on the parameters.
The Durbin-Watson statistic is a measure used to detect the presence of autocorrelation (serial correlation) in the residuals of a regression model.
A probit model is a type of regression model used when the dependent variable is binary, modeling the probability of an event occurring as a function of the independent variables using the standard normal distribution.
A Box-Jenkins model, also known as an ARIMA (Autoregressive Integrated Moving Average) model, is a flexible time series forecasting technique that combines autoregressive and moving average components after differencing to achieve stationarity.
In-sample forecasting uses the same data set to estimate the model and generate forecasts, while out-of-sample forecasting uses a different data set for forecasting than the one used for model estimation, providing a more realistic evaluation of the model's predictive power.
Serial correlation, also known as autocorrelation, refers to the correlation between the residuals of a regression model at different time periods, violating the assumption of independent errors required for valid statistical inference.
Omitted variable bias occurs when a relevant variable is excluded from a regression model, leading to biased and inconsistent estimates of the coefficients, as the omitted variable's effect is absorbed into the error term.
A cross-sectional model analyzes data collected at a single point in time across different individuals or entities, while a panel data model combines cross-sectional data and time series data, allowing for the analysis of both cross-sectional and temporal variation.
A recursive model assumes that the relationships between the variables are one-directional (no feedback loops), while a non-recursive model allows for bidirectional or simultaneous relationships between the variables.
A seemingly unrelated regression (SUR) model is a system of regression equations where the error terms are correlated across equations, allowing for more efficient estimation by combining information from different equations.
A Monte Carlo simulation is a computational technique that uses random sampling to estimate the probability of different outcomes or to evaluate the performance of an econometric model under various scenarios.
A parametric model assumes a specific functional form and distribution for the data, while a non-parametric model makes fewer assumptions about the underlying distribution and functional form, relying more on the data itself.
A static model assumes that the relationships between variables do not change over time, while a dynamic model incorporates time lags or changes in the relationships over time, allowing for more realistic modeling of dynamic systems.
Residual analysis in econometrics involves examining the residuals (the difference between observed and predicted values) of a regression model to check for violations of the underlying assumptions, such as heteroskedasticity, autocorrelation, or non-normality.
A spatial econometrics model is a type of regression model that accounts for spatial dependence or spatial autocorrelation, where observations in close geographical proximity tend to be more similar than those further apart.
A hazard model is a type of econometric model used to analyze the timing of events or transitions, such as unemployment duration or firm survival, taking into account censored observations and time-varying covariates.
A quantile regression model estimates the relationship between the independent variables and specific quantiles (e.g., median, quartiles) of the dependent variable, rather than just the mean, providing a more comprehensive understanding of the distribution.
A structural VAR model imposes economic theory-based restrictions on the contemporaneous relationships between the variables, while a reduced-form VAR model treats all variables as endogenous without imposing such restrictions.
Impulse response analysis in VAR models is used to trace the effects of shocks or innovations in one variable on the future values of other variables in the system, providing insights into the dynamic relationships between variables.
A generalized method of moments (GMM) estimator is a widely used estimation technique in econometrics that allows for consistent and efficient parameter estimation even when the moment conditions (assumptions about the error term) are not fully specified or satisfied.
Balanced panel data refers to a dataset where all individuals or entities have observations for the same set of time periods, while unbalanced panel data includes individuals or entities with missing observations for some time periods.
The Hausman test is a statistical test used to choose between a fixed effects model and a random effects model in panel data analysis by testing for the presence of correlation between the individual effects and the regressors.