Eric has been working to build, distribute, and strengthen the GAUSS universe since 2012. He is an economist skilled in data analysis and software development. He has earned a B.A. and MSc in economics and engineering and has over 18 years of combined industry and academic experience in data analysis and research.
The posterior probability distribution is the heart of Bayesian statistics and a fundamental tool for Bayesian parameter estimation. Naturally, how to infer and build these distributions is a widely examined topic, the scope of which cannot fit in one blog. In this blog, we examine bayesian sampling using three basic, but fundamental techniques, importance sampling, Metropolis-Hastings sampling, and Gibbs sampling.
We use regression analysis to understand the relationships, patterns, and causalities in data. Often we are interested in understanding the impacts that changes in the dependent variables have on our outcome of interest. However, not all models provide such straightforward interpretations. Coefficients in more complex models may not always provide direct insights into the relationships we are interested in.
In this blog, we look more closely at the interpretation of marginal effects in three types of models:
Purely linear models.
Models with transformations in independent variables.
Models with transformations of dependent variables.
In this blog, we examine one of the fundamentals of panel data analysis, the one-way error component model. We cover the theoretical background of the one-way error component model, we examine the fixed-effects and random-effects models, and provide an empirical example of both.
When policy changes or treatments are imposed on people, it is common and reasonable to ask how those people have been impacted. This is a more difficult question than it seems at first glance. In today’s blog, we examine difference-in-differences (DD) estimation, a common tool for considering the impact of treatments on individuals.
GAUSS includes a plethora of tools for creating publication-quality graphics. Unfortunately, many people fail to use these tools to their full potential. Today we unlock five advanced GAUSS hacks for building beautiful graphics:
Using HSL, and Colorbrewer color palettes.
Controlling graph exports.
Changing the plot canvas size.
Annotating graphs with shapes, text boxes, and lines.
Using LaTeX for GAUSS legends, labels and text boxes.
In this blog, we extend our analysis of unit root testing with structural breaks to panel data. Using panel data unit roots tests found in the GAUSS tspdlib we consider if a panel of international current account balances collectively shows unit root behavior.
In this blog, we examine the issue of identifying unit roots in the presence of structural breaks. We will use the quarterly US current account to GDP ratio to compare results from a number of unit root test found in the GAUSS tspdlib library including the: Zivot-Andrews (1992) unit root test with a single structural break, Narayan and Popp (2010) unit root test with two structural breaks, Lee and Strazicich (2013, 2003) LM tests with one and two structural breaks, Enders and Lee Fourier (2012) ADF and LM tests.
Classical linear regression estimates the mean response of the dependent variable dependent on the independent variables. There are many cases, such as skewed data, multimodal data, or data with outliers, when the behavior at the conditional mean fails to fully capture the patterns in the data. In these cases, quantile regression provides a useful alternative to linear regression. Today we explore quantile regression and use the GAUSS quantileFit procedure to analyze Major League Baseball Salary data.
The GAUSS interface includes a number of often overlooked hotkeys and shortcuts. These features can help make programming more efficient and navigation seamless. In this blog I highlight my top five GAUSS hotkeys:
Quickly view data symbols using Ctrl+E.
Open floating command reference pages using Shift+F1.
Linear regression commonly assumes that the error terms of a model are independently and identically distributed (i.i.d) However, when datasets contain groups, the potential for correlated error terms within groups arises. In this blog, we explore how to remedy this issue with clustered error terms.