Skip to content

Releases: lnnrtwttkhn/talk-rdm

Version 5.0 of slides for a talk on an open and reproducible research workflow, prepared for the Open Science Initiative at the Department of Psychology at TU Dresden

20 Jan 11:00
019605e
Compare
Choose a tag to compare

Version 4.0 of talk on tools for an open and reproducible research workflow, prepared for the 5th Workshop on Research Data Management 2022 by the Max Planck Society

13 Sep 13:00
Compare
Choose a tag to compare

About

The following slides were presented during a talk prepared for the 5th RDM-Workshop 2022 on Research Data Management in the Max Planck Society at the Max Planck Institute for Human Cognitive and Brain Sciences in Leipzig, Germany from the 13th to 14th of September 2022. The talk was held in German (but slides were in English).

PDF

A PDF version of the slides can be found below. Please excuse minor issues resulting from HTML to PDF conversion.

Abstract (German)

Vorläufiger Titel: Forschung als Software Engineering - Lösung für alles?
Finaler Titel: Tools für einen offenen und reproduzierbaren Forschungsprozess

Unzureichende Reproduzierbarkeit, mangelnde Transparenz und ineffiziente Arbeitsabläufe - viele wissenschaftliche Disziplinen haben einen großen Innovationsbedarf. Zudem wird der Forschungsprozess immer digitaler, komputationaler und interaktiver. In meinem Vortrag argumentiere ich, dass technische Lösungen für diese Herausforderungen bereits vorhanden und in der professionellen Softwareentwicklung zu finden sind. Dazu zählen vor allem Tools für interaktive Versionskontrolle von Daten wie Git oder DataLad und Software-Container wie Docker. Ich diskutiere wie diese Tools wissenschaftliche Arbeitsabläufe transformieren und welche (vorranging nicht technischen) Hürden bei ihrer Implementierung bestehen. Ähnliche vorherige Vorträge finden sich auf https://lennartwittkuhn.com/talk-rdm/.

Version 3.0 of slides for a talk on a research workflow with DataLad, prepared for the Lifespan Neural Dynamics Group

20 Oct 20:09
Compare
Choose a tag to compare

The following slides were presented during a talk prepared for the Lifespan Neural Dynamics Group (LNDG) at the Max Planck Institute for Human Development on 20th of October 2021.

A PDF version of the slides can be found below.
Please excuse minor issues resulting from HTML to PDF conversion.

Version 2.0 of talk on an open and reproducible research workflow, prepared for the Open Science Ambassadors Day 2021

18 Oct 09:50
Compare
Choose a tag to compare

About

Version 2.0 of this talk, titled "Tools for an open and reproducible research workflow", was prepared for the "Open Science Ambassadors Day 2021" hosted by the Max Planck PhDnet and the Max Planck Digital Library (see details here) on 18th of October 2021.

Please find the .pdf version below (please excuse minor inconsistencies).

Version 1.0 of talk on a workflow for open and reproducible MRI studies, prepared for the Max Planck Digital Library

22 Jun 11:36
Compare
Choose a tag to compare

About

Version 1.0 of this talk, titled "Towards a workflow for open and reproducible fMRI studies", was prepared for the "Discussion Series: Human Research Data in Practice" hosted by the Max Planck Digital Library (see details here) on 22nd June 2021.

Abstract

Achieving computational reproducibility and accessible data sharing can be challenging, in particular for neuroimaging research that involves large amounts of heterogeneous data and code. Here, we showcase a workflow that combines several software tools to allow reproducibility and transparent sharing of code and data of a human fMRI study.

We recently published an open-access paper (Wittkuhn & Schuck, 2021, Nature Communications) together with the code, data and computational environments needed to reproduce the reported results. We shared > 10 datasets via GIN (G-Node Infrastructure) as modular version-controlled units, including fMRI data organized in BIDS format and derived data, such as pre-processed fMRI data and data quality metrics.

Research data was version-controlled using DataLad. Following the DataLad YODA principles, we nested datasets as modular units, allowing to better establish data provenance, i.e., a clear overview which code used which input data to produce which output data. Code that reproduced the analyses was integrated with additional documentation using RMarkdown notebooks. The notebooks were automatically executed using continuous integration on GitLab. In this process, data was retrieved from GIN using DataLad, the notebooks were rendered and deployed to a website (https://wittkuhn.mpib.berlin/highspeed/). Code execution was performed using software containers (Docker and Singularity) and virtual environments, allowing to reproduce the computational environment.