Synthetic Control

Video Statistics and Information

Video
Captions Word Cloud
Reddit Comments
Captions
(humming tone) >>Hi again students. This is our last class topic and it's one of my favorites. We're going to explore the method of synthetic control. This topic is not in our textbook, but I've made a separate reading available to go with this video. Many times we have questions about a single policy. For example, what was the economic cost to West Germany reunification? Or, what is the effect of California's high cigarette tax on smoking? Or, what is the effect of Oklahoma's Right to Work law on wages or employment? To answer these kinds of questions we need to know what would have happened without the policy intervention. Synthetic control does this by creating an artificial version of the policy in question and comparing actual outcomes to the outcomes predicted by this estimated synthetic control. The technique brings a lot of credibility and rigor to case studies and is being widely adopted in finance, economics, and political science. The method seeks to create a control that mimics the behavior of the outcome being studied in the pretreatment period and also matches important characteristics of the policy under study. We will be more specific about that later. For now, suppose we want to estimate the effect of Oklahoma's 2001 Right to Work law on state output by a synthetic control. We would first choose a set of potential control units. In our example, states without Right to Work laws would be a good choice. Then, we would choose a set of characteristics that we think effect state output. For example, education levels, population density, and investment. Then the method uses linear programming to choose a weighted average of some of the potential control units to create a synthetic Oklahoma from say 1980 to 2000, the year before the law passed. That weighted average is the synthetic control and is chosen both to track actual Oklahoma state output levels and have a similar profile for education, population density, and investment. To tell if we have a good control, we can compare how well it tracks actual Oklahoma and how closely it matches actual Oklahoma on the characteristics we have chosen. In a way, this is like matching except we have two criteria and we're looking to match over a relatively long period of time instead of at one particular point in time. Once we have a good synthetic control, we simply compare the path of Oklahoma after 2001 with a path of synthetic Oklahoma generated by our control. Assuming the only difference between the two is due to the Right to Work law, our estimated treatment effect of the law is simply the difference in the two post 2001 time paths. This is somewhat similar to a difference in difference approach. So to use this technique, we need a relatively long pre-treatment period, a set of reasonable potential control units, a set of characteristics we think are important in determining the outcome we are studying, and data on the outcome, and characteristics for the treated unit and all the potential control units. In the pre-treatment period, we estimate the synthetic control unit as described above using only pre-treatment date. Then, since the control is just weighted average of the other units, we can use those weights to create our counterfactual outcome in the treatment period. The estimated treatment effect is the difference between the behavior of the treated unit and that of it's synthetic control after the treatment is realized. Let's consider an example from the literature. In 2015, Abadie, Diamond, and Hainmueller published paper in the American Journal of Political Science that used synthetic control to estimate the effect reunification had on West Germany. Their pre-treatment period was 1960 to 1989. Their potential control units were Australia, Austria, Belgium, Denmark, France, Greece, Italy, Japan, the Netherlands, New Zealand, Norway, Portugal, Spain, Switzerland, the United Kingdom, and the United States. The outcome variable was real per-capita gross domestic product. The characteristics they chose to match on were inflation, industry share of value added, investment, schooling, and a measure of trade openness. Their estimated synthetic West Germany was 42% Austria, 16% Japan, 10% Netherlands, 10% Switzerland, and 22% the USA. Here's a graph showing the results. As can be seen, the pre-treatment fit is very close. The synthetic control tracks actual West Germany extremely well. After reunification though, West Germany underperforms the control by about 12% by the end of the period. Here is the table checking for matching on the economic characteristics. By comparing the first two columns it becomes clear that the synthetic control matches West Germany quite well on the chosen economic characteristics. So in this paper, we see a terrific example of what we want our synthetic control to do which is closely track the actual outcome variable pre-treatment and closely match on the chosen characteristic variables. If we have that, our work is done and the treatment effect is what it is. In the case of West Germany, we see the country was around 12% poorer in 2003 than it would have been without reunification. Like propensity score matching, synthetic control matches on observables meaning that time varying unobservables could confound the results. However, if the control is able to mimic the actual unit over a long pre-treatment period, then unobservables are not a big problem during that period. Which means maybe they aren't a big problem in the treatment period either. As described, the method does not provide a significance test on the treatment effect. This isn't necessarily a problem as case studies generally do not provide such tests. However, we can create a significance level via what is know as a permutation test. To do this test, we estimate a series of placebos. That is we pretend that the treatment occurred in one of the control units, create a synthetic for that unit in the pretreatment period, and then compare the fit of the actual and synthetic in the treatment period. We repeat this process for each of the control units. Obviously, since the treatment didn't happen in the control countries, we shouldn't see an effect. So if the bogus treatment effects are estimated in the control units are the same size or larger than the one we estimated for the treated units, we would infer that the estimated treatment effect is not really significant. It looks just like the noise generated by the placebo treatments. Under the assumption that the treatment is randomly assigned to a unit, we can use the rank of the size of our estimated treatment effect compared to the placebo effects as a significance test. In the paper on German reunification, they show the estimated effect for West Germany is larger than any of the estimated placebo effects. The probability of that happening is one 1/17th. Or 17 is equal to the total number of countries used, namely West Germany, and the 16 controls. 1/17th is equal to about 0.06 and that could be taken as the significance level of estimate of the cost to West Germany of the reunification. Case studies often lack rigor and synthetic control is a good way to make them more systematic and believable. To run a good synthetic control experiment, we need a long pre-treatment time period to establish the control. We need a set of potential control units that are similar to the treated units, and we need to ensure these potential controls do not receive the same treatment as our treated unit. We also need a good set of variables that help to explain the outcome we are studying. If the synthetic control outcome closely matches the treated unit's outcome during the pre-treatment period and matches the treated unit on the values on the predictor variables chosen, then we have a good model and can simply take the difference of outcomes between the synthetic and the actual treated unit during the treatment period as our estimated treatment effect. If we have our heart set on having a significance test, we can get one by performing a series of placebo test and creating the significance level for the estimated treatment effect based on the rank ordering of the sizes of the estimated effects, both actual and placebo. Well, that's about it class. I hope you've enjoyed and learned from this short course on quasi experimental methods. Always remember, friends don't let friends run OLS regressions to estimate treatment effects. (humming tones)
Info
Channel: Janux
Views: 10,007
Rating: 4.9849625 out of 5
Keywords: Janux, University of Oklahoma, OU, economics, econometrics, online learning, experimental, online course, kevin grier, nextthought
Id: 1PQfeDT8zXM
Channel Id: undefined
Length: 8min 9sec (489 seconds)
Published: Mon Aug 21 2017
Related Videos
Note
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.