Publication

Data Management Plans: the Importance of Data Management in the BIG-MAP Project[**]

Abstract

Open access to research data is increasingly important for accelerating research. Grant authorities therefore request detailed plans for how data is managed in the projects they finance. We have recently developed such a plan for the EU H2020 BIG-MAP project-a cross-disciplinary project targeting disruptive battery-material discoveries. Essential for reaching the goal is extensive sharing of research data across scales, disciplines and stakeholders, not limited to BIG-MAP and the European BATTERY 2030+ initiative but within the entire battery community. The key challenges faced in developing the data management plan for such a large and complex project were to generate an overview of the enormous amount of data that will be produced, to build an understanding of the data flow within the project and to agree on a roadmap for making all data FAIR (findable, accessible, interoperable, reusable). This paper describes the process we followed and how we structured the plan.

About this result
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.
Related concepts (32)
Data management
Data management comprises all disciplines related to handling data as a valuable resource. The concept of data management arose in the 1980s as technology moved from sequential processing (first punched cards, then magnetic tape) to random access storage. Since it was now possible to store a discrete fact and quickly access it using random access disk technology, those suggesting that data management was more important than business process management used arguments such as "a customer's home address is stored in 75 (or some other large number) places in our computer systems.
Big data
Big data primarily refers to data sets that are too large or complex to be dealt with by traditional data-processing application software. Data with many entries (rows) offer greater statistical power, while data with higher complexity (more attributes or columns) may lead to a higher false discovery rate. Though used sometimes loosely partly because of a lack of formal definition, the interpretation that seems to best describe big data is the one associated with a large body of information that we could not comprehend when used only in smaller amounts.
Data
In common usage and statistics, data (USˈdætə; UKˈdeɪtə) is a collection of discrete or continuous values that convey information, describing the quantity, quality, fact, statistics, other basic units of meaning, or simply sequences of symbols that may be further interpreted formally. A datum is an individual value in a collection of data. Data is usually organized into structures such as tables that provide additional context and meaning, and which may themselves be used as data in larger structures.
Show more
Related publications (39)

Data Champions Lunch Talks - Green Bytes: Data-Driven Approaches to EPFL Sustainability

Miguel Peon Quiros, Francesco Varrato, Chiara Gabella, Manuel Simon Paul Cubero-Castan

For this edition of the DC Lunch Talks series, the discussion centered around Data-Driven Approaches to sustainability at EPFL, a topic of significant relevance in the contemporary academic landscape. The event featured a series of short talks by experts w ...
2024

Diversity and neocolonialism in Big Data research: Avoiding extractivism while struggling with paternalism

Daniel Gatica-Perez

The extractive logic of Big Data-driven technology and knowledge production has raised serious concerns. While most criticism initially focused on the impacts on Western societies, attention is now increasingly turning to the consequences for communities i ...
Thousand Oaks2023

Shared metadata for data-centric materials science

Giovanni Pizzi, Ronald Earle Miller, Gian-Marco Rignanese, Carsten Baldauf, Matthias Scheffler, Tristan Bereau

The expansive production of data in materials science, their widespread sharing and repurposing requires educated support and stewardship. In order to ensure that this need helps rather than hinders scientific work, the implementation of the FAIR-data prin ...
Berlin2023
Show more
Related MOOCs (16)
Simulation Neurocience
Learn how to digitally reconstruct a single neuron to better study the biological mechanisms of brain function, behaviour and disease.
Simulation Neurocience
Learn how to digitally reconstruct a single neuron to better study the biological mechanisms of brain function, behaviour and disease.
Simulation Neurocience
Learn how to digitally reconstruct a single neuron to better study the biological mechanisms of brain function, behaviour and disease.
Show more

Graph Chatbot

Chat with Graph Search

Ask any question about EPFL courses, lectures, exercises, research, news, etc. or try the example questions below.

DISCLAIMER: The Graph Chatbot is not programmed to provide explicit or categorical answers to your questions. Rather, it transforms your questions into API requests that are distributed across the various IT services officially administered by EPFL. Its purpose is solely to collect and recommend relevant references to content that you can explore to help you answer your questions.