Abstract
The link between policy analysis and data science is more delicate than it may appear. A new policy, by de_nition, will change the underlying data generating model, rendering classi_cation or supervised learning inapplicable. Perhaps eliciting causal relations from observational data is the correct framework for estimating policy impact. However, there are substantial gaps between the theory, practice and feasibility of causal models. In this paper we argue that transduction, a form of inference where we reason from speci _c training instances to speci_c test instances, may provide an appropriate framework for evidence-based policy analysis. In particular, we will demonstrate that the matrix completion problem, introduced in the data science community for making predictions in recommendation systems, can be a powerful tool for both predicting and evaluating the impact of new policy changes.
Original language | English |
---|---|
Title of host publication | Accepted Papers for Workshop on Data Science for Social Good, Held in Conjunction with KDD 2014, 24 August 2014, New York, USA |
Publisher | ACM |
Number of pages | 5 |
Publication status | Published - 2014 |
Event | Workshop on Data Science for Social Good - Duration: 24 Aug 2014 → … |
Conference
Conference | Workshop on Data Science for Social Good |
---|---|
Period | 24/08/14 → … |