July 11, 2016

This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

When is big data too big? Making data-based models comprehensible

Credit: Mary Ann Liebert, Inc., publishers
× close
Credit: Mary Ann Liebert, Inc., publishers

Data-driven mathematical modeling is having an enormous impact on the ability to organize and describe very large data sets, and make inferences and predictions about populations and situations based on sampling data. However, as these models become increasingly complex, the ability of users to understand and apply them represents a growing challenge. The article "A Framework for Considering Comprehensibility in Modeling", which describes this emerging dilemma and a strategy for developing solutions, is published in Big Data.

Michael Gleicher, University of Wisconsin-Madison, defines comprehensibility as "the ability of the various stakeholders to understand relevant aspects of the modeling process." He suggests that comprehensibility should be a key goal in model development. However, as models become more sophisticated, tradeoffs may be inevitable—even between understandability and accuracy—in some cases, improving comprehensibility may help achieve other goals in modeling.

"Gleicher provides a holistic framework of comprehensibility that considers what the various stakeholders in a data project do and don't understand easily and their need for comprehensibility," says Big Data Editor-in-Chief Vasant Dhar, Professor at the Stern School of Business and the Center for Data Science at New York University. "More broadly, the article highlights comprehensibility from a human-centric standpoint, identifying the role and needs of humans in complex projects."

More information: Michael Gleicher, A Framework for Considering Comprehensibility in Modeling, Big Data (2016). DOI: 10.1089/big.2016.0007

Load comments (0)