Data science pipelines view data mathematically (e.g., measures, values distribution) first and establish a context for it later relying on computed models that approximate insights and foresights of the phenomena they represent. What a data scientist really would want to be doing is looking at the whole data set in ways that tell her things and answer questions that she is not asking. The pipelines design and its results remain empirical and partially explicit on how statistical tools and computer technologies are used to identify meaningful patterns of information? How shall significant data correlations be interpreted? What is the role…

Genoveva Vargas-Solar

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store