NIST defines Big Data and Data Science

The National Institute of Standards and Technology (NIST) is attempting to create standards for Big Data. They just released the NIST Big Data interoperability framework, which is a huge set of documents aimed at creating standards around everything in big data from definitions to architectures.

Big Data Definitions

In case you are wondering, and I know you are, what are the definitions. The framework includes many more definitions.

Big Data consists of extensive datasets – primarily in the characteristics of volume, variety, velocity, and/or variability – that require a scalable architecture for efficient storage, manipulation, and analysis.

Data science is the empirical synthesis of actionable knowledge from raw data through the complete data lifecycle process.

Don’t like the definitions? Great, NIST would love to hear your opinions/comments. Comments are being collected until May 21, 2015.

The NIST Big Data interoperability framework is a massive work consisting of 7 volumes. All are open for comments.

  1. Definitions
  2. Taxonomies
  3. Use Case & Requirements
  4. Security and Privacy
  5. Architectures White Paper Survey
  6. Reference Architecture
  7. Standards Roadmap

The process to submit a comment appears rather old-school (hint: NIST, Github might be a good place to collect comments/edits), but it is not difficult.

Advertisements

3 thoughts on “NIST defines Big Data and Data Science”

Leave a Reply

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out /  Change )

Google+ photo

You are commenting using your Google+ account. Log Out /  Change )

Twitter picture

You are commenting using your Twitter account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s