News

When is big data too big? Making data-based models comprehensible

12 Jul 2016
When is big data too big? Making data-based models comprehensible

Data-driven mathematical modelling is having an enormous impact on the ability to organise and describe very large data sets, and make inferences and predictions about populations and situations based on sampling data.

However, as these models become increasingly complex, the ability of users to understand and apply them represents a growing challenge.

The article "A Framework for Considering Comprehensibility in Modeling", which describes this emerging dilemma and a strategy for developing solutions, is published in Big Data, the highly innovative, peer-reviewed journal from Mary Ann Liebert, Inc., publishers.

The article is available free for download on the Big Data website until August 8, 2016.

Michael Gleicher, University of Wisconsin-Madison, defines comprehensibility as "the ability of the various stakeholders to understand relevant aspects of the modelling process."

He suggests that comprehensibility should be a key goal in model development.

However, as models become more sophisticated, tradeoffs may be inevitable--even between understandability and accuracy--in some cases, improving comprehensibility may help achieve other goals in modelling.

"Gleicher provides a holistic framework of comprehensibility that considers what the various stakeholders in a data science project do and don't understand easily and their need for comprehensibility," says Big Data Editor-in-Chief Vasant Dhar, Professor at the Stern School of Business and the Center for Data Science at New York University. "More broadly, the article highlights comprehensibility from a human-centric standpoint, identifying the role and needs of humans in complex data science projects."

Source: Mary Ann Liebert, Inc