In this session, we consider the best practises for the long-term preservation of the data and other results associated with research across the preservation lifecycle, from the submission of data packages for preservation, to the access of data products. This includes the organisational structures, policies and standards adopted by data centres and archives to assure cost-effective preservation, together with risk management, uncertainty quantification, quality assessment and the evaluation of preservation capabilities. Further, we will consider novel architectures and tools used to realise different preservation strategies, and standards, tools and languages to capture the preservation context, including the preservation of data formats, the use of identifiers, metadata, semantics, data provenance, quality and uncertainty.
In this session, we consider activities that add value to archived data, facilitate their use or produce novel data services. Data archivists often focus most of their energy on creating well-formed, well-documented archives with the expectation that they will be available for the next 50 to 100 years. However, archived data are meaningless if they cannot be easily retrieved, understood, and used. As a result we would like to invite submissions from projects or archives who rising to the challenge of enhancing data in order to facilitate exploitation of data assets.
This session will consider new challenges, activities and research related to Virtual Research Environments or Collaborative Environments. While Massive data growth is calling for a new paradigm, with a shift towards “bring the user to the data", where scientists can bring their own code and run it where the data actually reside, instead of downloading the data and run their analysis on their computer. There is also an increased need for data, associated documentation and software long-term preservation and accessibility, for scientists to be able to re-run data analysis that was initially applied on the data. Last, scientists are now expecting to share not only their data, but also their software and the results of their research activities, and to work with their collaborators in an easy and effective manner, regardless of their location.
The purpose of this session is to examine existing practices and systems and highlight what has been learnt, including how to best benefit from collaboration between projects and/or disciplines. It will also look forward and attempt to understand how new developments and/or technologies and/or tomorrow's data volumes might influence or even constrain how things will be done in the future. Data preservation is not a static field: we wish to use this session to explore what we have learnt from previous migrations and to consider how to best prepare for the future, including potentially disruptive scenarios. We would also like to facilitate discussion on how different services involved in LTDP interplay and to use this opportunity to consider how we measure success and respond to requirements from funding agencies, such as those for F.A.I.R. data management.