Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

The National Data Service Consortium (NDSC) aims to enhance scientific reproduciblity in the Information Age by implementing interfaces, protocols, and standards within data relevant cyberinfrastructure tools/services towards allowing for an overall infrastructure enabling scientists to more easily search for, publish, link, and overall reuse digital data.  Specifically, NDSC has as its mission:

... advancing the frontiers of discovery and innovation by enabling open sharing of data and increase collaboration within and across fields and disciplines. Success will be achieved through coordinated and concentrated efforts, developing an open environment of federated, interoperable, and integrated national-scale services.

...

One might think of this in terms of the modern internet, which had its roots in the ARPANET effort ...

Image RemovedImage Added

 

 

... taking into account additional components addressing "Big Data" challenges within the internet, a sort of DATANET:

Image RemovedImage Added

The above is illustrative and not a comprehensive list of services.  There are a variety of components being actively explored and developed, some building on top of others, some not, some interacting with others, some not.  In essence, just like with every Internet component there are a variety of options for each Datanet component.  Unlike the internet components, however, these selections matter - they effect the user and limit what other components might be utilized/deployed.

...

... and in terms of the general public and broader impact, perhaps seeding a new kind of internet all can take advantage of:

  • A child might ask their Apple TV or Amazon Echo the question, "What is the weather going to be like in the afternoon... 1000 years from now?".  This will not result in an answer today, but theoretically could with several of the actively developed data components and interoperability between them.  For example, ecological models such as ED, SIPNET, or DALEC within workflow engines such as PEcAn could pull data from Ameriflux, NARR, BETYdb, DataONE, and NEON, transfer data via Globus, convert data to model input formats via BrownDog, run models on XSEDE resources, and returns a specific result such as temperature/forecast estimate (likely several results from several models).  The result can further include additional information such as "if you want to learn more" which links to summarized versions of publications for the utilized models and datasets with DOIs to published papers or datasets in Zenodo or SEAD or executable versions of the the tools themselves (possibly simplified) within HUBzero or Jupyter lab notebooks.

...