Susan Athey on the Impact of Machine Learning on Econometrics and Economics (part 2)

I posted Susan Athey’s 2019 Luncheon address to the AEA and AFA in part 1 of this post.  (See here for her address).

Now, in part 2, I post details on the continuing education Course on Machine Learning and Econometrics from 2018 featuring the joint efforts of Susan Athey and Guido Imbens.

It relates to the basis for this blog where I advocate for economists in Data Science roles. As a crude overview, ML uses many of the techniques known by all econometricians and has grown out of data mining and exploratory data analysis, long avoided by economists precisely because such methods lead to in the words of Jan Kmenta “beating the data until it confesses.” ML also makes use of AI type methods that started with brute force methods noting that the computer can be set lose on a data set and in a brutish way try all possible models and predictions seeking a best statistical fit, but not necessarily the best economic fit since the methods ignore both causality and the ability to interpret the results with a good explanation.  

Economists historically have focused on models that are causal and provide the ability to focus on the explanation, the the ability to say why. Their model techniques are designed to test economic hypotheses about the problem and not just to get a good fit. 

To say we have discussed historically the opposite ends of a fair coin by setting up the effect of X–>y is not too far off. ML focuses on y and econometrics focus on X. The future is focusing on both, the need to focus good algos on what is y and the critical understanding of “why” which is the understanding of the importance of X.  

This course offered by the American Economic Association just about one year ago, represents the state of the art of the merger of ML and econometrics.  I offer it here (although you can go directly to the AEA website) so more can explore how economists need to incorporate the lessons of ML and of econometrics and help produce even stronger data science professionals. 

AEA Continuing Education Short Course: Machine Learning and Econometrics, Jan 2018

Course Presenters

Susan Athey is the Economics of Technology Professor at Stanford Graduate School of Business. She received her bachelor’s degree from Duke University and her PhD from Stanford. She previously taught at the economics departments at MIT, Stanford and Harvard. Her current research focuses on the  intersection of econometrics and machine learning.  As one of the first “tech economists,” she served as consulting chief economist for Microsoft Corporation for six years.

Guido Imbens is Professor of Economics at the Stanford Graduate School of Business. After graduating from Brown University Guido taught at Harvard University, UCLA, and UC Berkeley. He joined the GSB in 2012. Imbens specializes in econometrics, and in particular methods for drawing causal inferences. Guido Imbens is a fellow of the Econometric Society and the American Academy of Arts and Sciences. Guido Imbens has taught in the continuing education program previously in 2009 and 2012.

Two day course in nine parts - Machine Learning and Econometrics, Jan 2018

 Materials:

Course Materials (will attach to your Google Drive)

The syllabus is included in the course materials and carries links to 4 pages of readings which are copied and linked to the source articles below.

Webcasts:

View Part 1 – Sunday 4.00-6.00pm: Introduction to Machine Learning Concepts 

(a) S. Athey (2018, January) “The Impact of Machine Learning on Economics,” Sections 1-2. 

(b) H. R. Varian (2014) “Big data: New tricks for econometrics.” The Journal of Economic Perspectives, 28 (2):3-27.

(c) S. Mullainathan and J. Spiess (2017) “Machine learning: an applied econometric approach” Journal of Economic Perspectives, 31(2):87-106  

View Part 2 – Monday 8.15-9.45am: Prediction Policy Problems

(a) S. Athey (2018, January) “The Impact of Machine Learning on Economics,” Section   3. 

(b) S. Mullainathan and J. Spiess (2017) “Machine learning: an applied econometric approach.” Journal of Economic Perspectives, 31(2):87-106. 

 

View Part 3 – Monday 10.00-11.45am: Causal Inference: Average Treatment Effects

(a) S. Athey (2018, January) “The Impact of Machine Learning on Economics,” Section 4.0, 4.1. 

(b) A. Belloni, V. Chernozhukov, and C. Hansen (2014) “High-dimensional methods and inference on structural and treatment effects.” The Journal of Economic Perspectives, 28(2):29-50. 

(c) V. Chernozhukov, D. Chetverikov, M. Demirer, E. Duo, C. Hansen, W. Newey, and J. Robins (2017, December) “Double/Debiased Machine Learning for Treatment and Causal Parameters.” 

(d) S. Athey, G. Imbens, and S.Wager (2016) “Estimating Average Treatment Effects: Supplementary Analyses and Remaining Challenges.”  Forthcoming, Journal of the Royal Statistical Society-Series B.

View Part 4 – Monday 12.45-2.15pm: Causal Inference: Heterogeneous Treatment Effects

(a) S. Athey (2018, January) “The Impact of Machine Learning on Economics,” Section 4.2. 

(b) S. Athey, G. Imbens (2016) “Recursive partitioning for heterogeneous causal effects.” Proceedings of the National Academy of Sciences, 113(27), 7353-7360.

View Part 5 – Monday 2.30-4.00pm: Causal Inference: Heterogeneous Treatment E ects, Supplementary Analysis

(a) S. Athey (2018, January) “The Impact of Machine Learning on Economics,” Section 4.2, 4.4. 

(b) S. Athey, and G. Imbens (2017) “The State of Applied Econometrics: Causality and Policy Evaluation,” Journal of Economic Perspectives, vol 31(2):3-32.

(c) S. Wager and S. Athey (2017) “Estimation and inference of heterogeneous treatment effects using random forests.” Journal of the American Statistical Association 

(d) S. Athey, Tibshirani, J., and S.Wager (2017, July) “Generalized Random Forests

(e) S. Athey, and Imbens, G. (2015) “A measure of robustness to misspeci cation.” The American Economic Review, 105(5), 476-480.

View Part 6 – Monday 4.15-5.15pm: Causal Inference: Optimal Policies and Bandits

(a) S. Athey. (2018, January) “The Impact of Machine Learning on Economics,”
Section 4.3. 

(b) S. Athey and S. Wager (2017) “Efficient Policy Learning.”

(c) M. Dudik, D. Erhan, J. Langford, and L. Li, (2014) “Doubly Robust Policy
Evaluation and Optimization” Statistical Science, Vol 29(4):485-511.

(d) S. Scott (2010), “A modern Bayesian look at the multi-armed bandit,” Applied Stochastic Models in Business and Industry, vol 26(6):639-658.

(e) M. Dimakopoulou, S. Athey, and G. Imbens (2017). “Estimation Considerations in Contextual Bandits.” 

View Part 7 – Tuesday 8.00-9.15am: Deep Learning Methods

(a) Y. LeCun, Y. Bengio and G. Hinton, (2015) “Deep learning” Nature, Vol. 521(7553): 436-444.

(b) I. Goodfellow, Y. Bengio, and A. Courville (2016) “Deep Learning.” MIT Press.

(c) J. Hartford, G. Lewis, K. Leyton-Brown, and M. Taddy (2016) Counterfactual
Prediction with Deep Instrumental Variables Networks.” 

View Part 8 – Tuesday 9.30-10.45am: Classi cation

(a) L. Breiman, J. Friedman, C. J. Stone R. A. Olshen (1984) “Classi cation and
regression trees,” CRC press.

(b) I. Goodfellow, Y. Bengio, and A. Courville (2016) \Deep Learning.” MIT Press.

View Part 9 – Tuesday 11.00am-12.00pm: Matrix Completion Methods for Causal Panel Data Models

(a) S. Athey, M. Bayati, N. Doudchenko, G. Imbens, and K. Khosravi (2017) “Matrix Completion Methods for Causal Panel Data Models.” 

(b) J. Bai (2009), “Panel data models with interactive fi xed effects.” Econometrica, 77(4): 1229{1279.

(c) E. Candes and B. Recht (2009) “Exact matrix completion via convex optimization.” Foundations of Computational Mathematics, 9(6):717-730.

 

Thomas Sowell and Economic Literacy

Image found at twitter.com/nick_bunker/status/971021895406641153, click image to buy.

I am teaching an introductory economics course that is below the typical principles level and is far too often taught as a “baby principles course” or such a watered down version of a principles course to be unrecognizable as economics. In my humble opinion, teaching a survey course as the watered down version of what we teach as Principles of Economics is a type of malpractice and an ethical lapse. It is the economic equivalent of teaching math by focusing only on adding and subtracting while obscuring all higher order functions as “beyond y’all,” or “too complicated.” It is also like teaching history as a set of unrelated dates without any context, or with only biased context. presented.

There are many books that report to survey economics available to me as a professor, but my choice is Thomas Sowell’s Basic Economics: A Common Sense Guide to the Economy, now in its 5th edition. The main reasons are:

  1. The proper goal for this course is to enhance economic literacy as a life skill, 
  2. Supply and demand drawing students once escaping the classroom forget all they learned, and worse remember that to solve that economic problem requires graphics and mathematics they can’t recall,
  3. Focusing on the graphics and equations necessarily competes with learning historical context and logical thinking,  
  4. Learning to not be deceived by good sounding things that do not hold up to scrutiny requires persons who are brought up to think critically, logically with historical context and the lessons of evidence, and finally
  5. the book is inexpensive (about $18 at Amazon) and close to Open Source in prices.

I desire to teach as Sowell writes, without any mathematics, equations or economic graphs and without being highly political. 

Students need to remember the principles of economics and not the ability to solve a problem with supply and demand graphs. They need to be economically literate. I think this course should be taught at this level with a foundation of common sense, logic and evidence. Moreover, in this highly partisan politicized world too often politics gets in the way of economics, or worse, someone hears an economic principle and immediately associates it with a partisan position. Students need also be introduced to an many economic fallacies as possible.

My mantra for students these many years since I have been teaching is we will embrace:

Economic Reality and not Wishful Thinking.

The freedom to this approach is that we can talk about policy and stay out of politics. My son relates a conversation when he was majoring in economics a few years ago with a highly partisan politically motivated student and head of a local campus political group. He was asked how my son could live with such an (insert pejorative political label here) for a dad. My personal politics beliefs were just the opposite. So I win. I leave politics outside of the classroom and focus only on analyzing the policies in a logical manner using historical context and evidence. 


Of course, I relate in class from time to time that the left think this and the right thinks that, but I am careful not to impress my political view on the students. I do not care if they figure it out, but I model the unbiased approach as much as I can. In fact i tell them if they vote left or right I do not care. Indeed, if they apply the evidence, logic and historical context to the problem then I am confident they are an informed voter and can pull whatever lever they wish. That is not my call, my mission is to make them think before they “jump.”

Susan Athey on the Impact of Machine Learning on Econometrics and Economics (part 1)

Economists make great data scientists. In part, this is because they all are trained in the four pillars of data science (1) data acquisition, (2) Data manipulation, cleaning and management, (3) analysis and (4) reporting and visualization. Good programs make sure that the economics students are trained in all four areas. Economists have subject matter expertise that is wrapped in a formalized way of thinking and problem solving ability. Quick answer, why are economists so valuable in business? – They know how to solve problems and tell stories from the evidence. 

As to the analysis part of these pillars, economists are typically wrapped in causality and explanation of X in the y = f(X,e) model. Economists in forecasting become more interested in predicting y with less or much less on the factors X

When I talk to many I hear data science being linked only with Machine Learning as if ML and data science are synonyms. This is far from the truth, with Data Science being very broad and ML a specific way and in some cases a dominate way of approaching a data problem. ML is making its way into economic curriculum. So what is the role of ML in economics now and into the future and more particularly the role between econometrics and ML?

No one in the economics profession knows more about the intersection of economics and machine learning than Susan Athey who just last month gave an address to the American Economic Association and the American Finance Association.  I am posting this address so you may understand the current state. 

In part 2 of this post I link to her two day course offered in January 2018 on ML and Econometrics with Guido Imbens. 

The AEA/AFA address, Jan 2019

This video was captured at the joint luncheon for the American Economics Association and the American Finance Association that occurred at the January 2019 Annual Meetings in Atlanta. 

Susan Athey who is The Economics of Technology Professor of the Graduate School of Business at Stanford University delivers the address and is introduced by Ben Bernanke, former FED chair, now at the Brookings Institution.

external link: https://www.aeaweb.org/webcasts/2019/aea-afa-joint-luncheon-impact-of-machine-learning

Economists as engineers: A new Chapter

(0:59:25) “The AI and econometric theory need work, but they are not the main constraint…. Instead the success is going to depend on understanding the context, understanding the setting….

“(The economist can be) motivated by social science research about where should I be spending my time, where should I be intervening? (Economists need) to use empirical work to help figure out what the best opportunities are.

Economists can help with defining measures of success.  We need to recognize that AI has billions of ways to optimize so we better be telling the algo the right thing. Those algos need to be constrained and informed by 

(1:01:22)  “Broadly, when economists return to their institutions that are building AI and data science initiatives that … the social scientists (she thinks)  are going to be more important than the  computer scientist in terms of what is the conceptual thing, what is the thing that makes something succeed or fail, that makes it screw up and have adverse consequences versus being really successful and impactful. We (economists) are going to need to join interdisciplinary teams and the evaluation will be embedded and not separable from the system. So that means we are going to have the opportunity to intervene in the world like we never have before. But it also comes great responsibility because we will be the people in the room who really can understand the good and the bad and make sure it happens in a safe way.””

 

Remembering Gary Becker

Rarely have I read the first few paragraphs of an article and felt that it captures the essence of the words to follow as well as I just have in Heckman, et al., Gary Becker Remembered. Perhaps this impact is because of the subject matter and my interest forged in the 1970s. Gary Becker made study of economics at The Ohio State University so exciting for me and my student colleagues such as Randy King, Tim Carr and Patricia Shields.  

It was a period where the 2nd edition of Human Capital had just come out, when we had read mimeograph copies of his Woytinsky lecture and struggled with his Theory of the Allocation of Time. Generally, his work and thought permeating all we did and studied. He made so much sense and conveyed so much wisdom.  It was a time we were working for the National Longitudinal Surveys under Herb Parnes, taking micro and labor from him, Belton Fleisher, Don Parsons and Ed Ray. All who made my interest in labor and labor econometrics all the more deep. Additionally, George Rhodes and Jerry Thursby pushed me econometrically and I had all that wonderful access to the early waves of the original NLS cohorts. It was a great time, although I did not necessarily think so at the time being a typical graduate student with little time to consider words like “great time.” It was at least exciting and rewarding, and over the next 40 years fulfilling.  


 

We who were running massive number of wage equations had him ever in our thoughts. We built on his foundation while stretching the model this way and that, but always on his foundation. While I have long ceased running wage regressions, my students in econometrics regularly do as they practice the techniques and at the root of their work is making sure they who are not all labor economic oriented, understand that the foundation on which they learn and build is Beckers, making them read the now basic work that was so vivid in the 70s. Whether they apply techniques from Ronald Oaxaca or Jim Heckman or others, it is Becker with whom they must first contend. He was an “intellectual giant” always the scientist, depending on the evidence from the causal link of theory to data and analysis and on to testable conclusion. Come to think of it, few others can boast of such a tight link between the economic model and the econometric model and testable results over and over that confirms his brilliance.

Here is the first three paragraphs of that remembrance from:
 

 James HeckmanEdward LazearKevin Murphy.  Gary Becker Remembered, Journal of Political Economy. October 2018Vol. 126Issue S1Pages S1-S6 Accessed at https://www.journals.uchicago.edu/doi/full/10.1086/698751 on Jan 1, 2019.

“Gary Becker was an intellectual giant. No one had a greater impact on broadening economics and making its impact felt throughout the social sciences than Becker. Indeed, Milton Friedman once described Gary Becker as the most important social scientist of the second half of the twentieth century.

“For those of us who knew him, he was the most creative thinker we ever encountered. It was his astounding imagination that made many of his early critics think of him as a heretic. They were correct: he was a heretic much like Luther, Copernicus, and Galileo, who transformed their worlds, just as he transformed economics. He brought a rigorous and insightful approach to issues that were viewed as inherently noneconomic. Eventually, he won over the economics profession, detractors and all, who eventually became converts.

“Becker was a scientist in the true sense of the word. He believed that economics was useful only if it explained and helped to improve the world. He practiced what he preached and carefully analyzed all of the social problems he addressed. He was innovative yet rigorous, open to new thought yet disciplined in sticking to the established rules of analysis. Most importantly, he extended the boundaries of economics to much of social science.” 

Read the rest at https://www.journals.uchicago.edu/doi/full/10.1086/698751

SAS Certificate in Economic Data Analytics

The Department of Economics is preparing economists with strong econometrics and programming skills  to prepare them to be data analysts and data scientists, ready to serve and lead business, governmental and not-profit institutions with their data analytic needs. Economists make the best data scientists, and in a team environment are a critical part of the data science team.  This owes to the rigorous analytical training that economists go through in programming, statistics, mathematics, econometrics and forecasting, but more importantly in their education in critical thinking and  problem solving in an environment that is both theoretical way and backed by scientific data evidence. In other words, economists stand out to employers. We have many testimonials that when economists are hired they move up in the organization quickly. The challenge is often getting employers to consider economists in the range of disciplines they hire.  One way we are using is to recognize our students for a particular skill, that of knowing SAS programming. 

The Economic Data Analytics Certificate of Completion is jointly sponsored by the SAS Institute and the Department of Economics  at The University of Akron. 

Spring 2018 SAS Certificate Presentation

The jointly offered certificate is based on student’s successful completion of the economics major core coursework in data analytics along with economic theory at either the undergraduate of graduate level.  Each honoree also successfully complete a major research study using SAS as the primary tool of their data analytic project. These data projects, most often the students senior research project, are a valuable asset to the student to help convince employer of their ability and interest in research. 

Importance of Economic Analysis to Data Science

Economists are unique in their approach to data analytics due to their unique training. They are story tellers, they are able to articulate problems and to take a rigorous approach to the solution of those problems. They understand the importance to understand the question: “why?”

Economists own causality and understand observational data, some are even said to be obsessed with the data generating process. Economists have a a strong linear regression toolkit and are well familiar with the processes that push beyond that level of rigor. Economists understand minimizing and maximizing of objective functions under constraint.

David Autor on Changing Patterns of Work

Are earnings differences between males and females due to discrimination? A typical approach is to compare the earnings of women to that of men and try to control for typical understandable differences such as education level and location and other factors. Perhaps education levels and location interact and an interaction term is introduced into our model. However, the largest assumption is that once we define our variables, homogeneity rules, that education is homogeneous, e.g., HS graduation means the same for all groups, over all times and locations). I see Autor’s lecture pointing out this heterogeneity and disputing the assumption that all persons are products of the same data generating process. He takes this on and at least for me smashes my initial biases. To be fair, this is my reading of his efforts, he does not utter the word heterogeneity at all, but I don’t think he needed to, not everyone in the audience are econometricians and the implicit heterogeneity problem is taken on directly. I will be sharing this lecture with my data analytic students as a great example of exploratory data analysis that allows a masterfully told story through complex preparation and easy to understand visuals.

The Richard T. Ely lecture at the 2019 American Economic Association meetings was presented by David H. Autor (MIT, NBER) comparing Work of the Past with the Work of the Future. Motivated in part by the “remarkable rise of wage inequality by education,” “vast increase in supply of educated workers,” and a “key observation (of the) polarization of work” that while “most occupational reallocation is upward,” “the occupational mobility is almost exclusively downward” for non-college workers, Autor proceeds to give rise to answers to the questions surrounding

  1. Diverging earnings and diverging job tasks
  2. The changing geography of work and wages
  3. The changing geography of workers, and
  4. What and where is the work of the future.

The visual presentation makes his data exploration very understandable and are masterfully done. He truly paints a picture that emerges from a vast amount of data that is entertaining and informative. This is well worth the 47 minutes and may actually challenge your preconceived thinking as to the nature of inequality in earnings. It is not as simple as one may think and he perfectly illustrates without ever uttering the word that data heterogeneity when ignored leads to false and inescapable conclusions.

Work of the past, work of the future. Richard T. Ely Lecture, AEA meetings, Atlanta, January 4, 2019.

Click on the above image and you will be well rewarded if you want to see a story told with strong graphics, proving to me anyway, that deep diving into data and presenting simple graphics (although complex in their creation) is a most effective way to communicate. A couple of examples of the graphics:

What if we do an econometric analysis of earnings between men and women using current data and a similar analysis from the 1980s. Can you see how this graph as one of many in Autor’s presentation might create havoc in comparing the result in the 1980s to one current? Watch the presentation, plenty of more visuals like this one.

SAS Coding, Problem Based Learning and preparing economists for data science careers: frustration to elation

Published on October 16, 2018 on LinkedIn

I love coding, but I love copying others code even more. There is a great SAS resource that is amazingly helpful, run by Lex Jansen (link below). I had a need and found it in a 2003 SUGI28 presentation: “Paper 118-28-Renaming All Variables in a SAS® Data Set Using the Information from PROC SQL’s Dictionary Tables,” by Prasad Ravi accessed at http://www2.sas.com/proceedings/sugi28/118-28.pdf

My need was to merge data from multiple years of the Index of Economic Freedom (https://www.heritage.org/index) across all countries where all variable names were the same. A simple merge overwrites the common variable name with the last data merged. So my creation of a clasroom problem/team based learning exercise by merging 2013 and 2018 data was much frustrated. Sure, I could go into the Excel and manually change every variable name in every sheet, but really that is what coding is supposed to before. The elegant macro by Prasad Ravi named “rename” worked and I modified it to change the specific prefix used (from NEWNAME_ to my choice) and to protect certain key variables (such as ID) for merging.

I changed %macro rename(lib,dsn); to %macro rename(lib,dsn,prefix,protect); and all is well. So the 2013 table will have prefix Y2013_ and I can protect the first 4 variables of country id, country name, webname and region which are (or should be) common in all years of the data. My specific change is limited to the inclusion of prefix and protect as illustrated in the PROC DATASET part of the rename macro shown here:

proc datasets library=&LIB; modify &DSN; rename %do i=&protect+1 %to &num_vars; &&var&i=&prefix&&var&i. %end;

and in the call %macro rename(lib,dsn,prefix,protect);

My economic data analytic students will be “thrilled” (or I will always think they should have been) as we move from EDA in Time Series to EDA in cross-section data. It is an introductory undergraduate class so nothing beyond simple presentations and simple statistics can be used and many do not yet have their statistics completed. After this class they will start down the multivariate inference, model selection and specification path. My class is to teach them SAS use in Economic Data Analytics and prepare for them a platform to stand on so any further foray into Economic Data Science is possible. At a minimum they will take required courses in econometrics and economic forecasting and encouraged to load up on other analytical electives.

Each problem they have to solve in their teams cover both the pillars of data science (aquisition, management and manipulation, analysis and reporting) and the pilars of applied econometrics (of problem articulation, data cleaning and model specification, hat tip to Peter Kennedy, p. 361 of his guide to econometrics, 6e.). The students just finished a time series problem/team based learning exercise requiring the merging of multiple datasets and a problem that requires extraoridinary articulation. The question was “Do deficits go up under republicians and down under democrats?” A moment of reflection beyond your personal knee-jerk reaction reveals that problem to have many facets and to the students discontent, there is no one simple answer. Dashed are their hopes that the first piece of analyitical effort will reveal truth and much dective work must ensue. The next step is into a problem/team based investigation of the role of economic freedom and progress throughout the countries of the world. I have until Monday to come up with the problem statement!

About Lex Jansen: If you do not know about the tremendous SAS resource at https://www.lexjansen.com then you should! I found a lot of help there after searching documentation and sas.support.com without finding the specific help I needed. Apologies to others offering similar approaches to a macro renaming, I do not know if the Ravi paper is the first or the best. What I know is it worked for me. Thanks to Prasad Ravi for writing it and to Lex Jansen for storing it. 

I mentionned applied econometrics and data science in the above and have to pause and thank two friends who I met each personally only once, but communicated with on these topics. Peter Kennedy has passed and I miss not asking him about my next brainstorm. He came and spoke to our university at my invitation in 2004. I was so thoroughly convinced of his applied econometric processes that his book (which I used from its first edition) and our conversations have effected my teaching to this day. While I had used problem based learning in my classes it was Peter who encouraged it as the single best way to teach applied econometrics. As much as the students in the midst of a PBL process “hate” it, the number of students feeding back their succcess in their careers and saying how it was the PBL that granted them success is my reward. Peter also writes in he book that there is a “world of difference between applied and theoretical econometrics …” that most university econometric “teaching is technique oriented rather than problem oriented.” He goes on to say that teaching applied and not theoretical econometrics is hard and brings the teacher no professional prestige. To this I can attest, but my product is not professional articles, but students in very high level data-analytic and sata-science positions.

As to data science it was work by Ken Stanford (@eKENomics), who I first talked to when he was at SAS, that convinced me that economists make great data scientists. Like Peter he confirmed much that I was already doing and teaching and helped me refine my approach to my teaching. He also was responsible for helping our students qualify for a SAS Certificate in Economic Data Analytics offered jointly by our department and SAS. In a pure self-serving manner I cite Ken’s contribution in a paper on our department’s website here. Ken serves on our Economics Advisory Board and has moved from SAS to Dataiku. His encouragement to our faculty at the 2017 NABE conference pushed us to forge ahead with our deeper curriculum changes into preparing economists for data science roles. Now if we can just get the administration to buy into our vision!

The test of time is how well these current students do in their careers, an answer we may have to wait years to know.

Importance of Economic Analysis to Data Science

I wrote this a few years ago (about 2014) and it still appears on my university department’s website. My thinking has evolved since then, but it is instructive where my thought was then.

Economists are unique in their approach to data analytics due to their unique training. They are story tellers, they are able to articulate problems and to take a rigorous approach to the solution of those problems. They understand the importance to understand the question: “why?”

Economists own causality and understand observational data, some are even said to be obsessed with the data generating process. Economists have a a strong linear regression toolkit and are well familiar with the processes that push beyond that level of rigor. Economists understand minimizing and maximizing of objective functions under constraint.

Read more:  Economists in Data Analytics

Blog Launch

Launching EconDataScience

Forty years ago I was sitting in a hotel room in New York City being interviewed for a tenured position at The University of Akron. At my on-campus interview I was given the challenge of rebuilding the graduate curriculum of Statistics and Econometrics. Revised in my image, they included an emphasis on Applied Econometrics, hands-on computing with real data sets and code level programming in SAS. 

Economists have always in my lifetime been sophisticated economic data analysts and I have taken note as the term and title of data science and data scientist has arisen that economists cover a vast amount of the territory of what is data science and in a way always have. 

One year ago, I organized for our Department and College, the first Data Science Day held at The University of Akron featuring two of our MA alumni that held the title of data scientist. A survey of our graduates found that all of our graduates alumni and about half of  our undergraduates were in data analytic positions and many in data scientist or data science positions. 

That economists make  good data scientists will be the subject of many posts of this blog.  I will also comment on issues of economic literacy which I regard as a crisis in our world.  I am less likely to comment on policy than on principles, the former being befuddled with all manner of things and the latter more inviolate, removed from opinion and evidenced based.