Overview of Scientific Workflows

Photograph of Scott Callaghan

Presenter

Scott Callaghan

Southern California Earthquake Center
University of Southern California

Presenter Bio


Callaghan is a research programmer at the Southern California Earthquake Center (SCEC) based in Los Angeles, though he works remotely from St. Louis, Missouri. He first got involved at SCEC as an undergraduate intern, then as a graduate student, and now as staff. He is the project lead on a software application which performs physics-based probabilistic seismic hazard analysis for California. This software typically runs on large HPC systems such as Blue Waters at NCSA and Titan at Oak Ridge Leadership Computing Facility. His research interests include scientific workflows and high throughput computing.

Abstract


I will present an overview of scientific workflows. I'll discuss what the community means by "workflows" and what elements make up a workflow. We'll talk about common problems that users might be facing, such as automation, job management, data staging, resource provisioning, and provenance tracking, and explain how workflow tools can help address these challenges. I'll present a brief example from my own work with a series of seismic codes showing how using workflow tools can improve scientific applications. I'll finish with an overview of high-level workflow concepts, with an aim to preparing users to get the most out of discussions of specific workflow tools and identify which tools would be best for them.


Session details

When:10:00 CST, March 8, 2017

Length of Session: One hour

Target Audience: Any researchers interested in potentially using scientific workflows in their work.

Prerequisites: None.


Training and Reference Materials: A recording of this topic from the Workflows Workshop of 2016 is available online.

Webinar video


If you don't see the video, reload the page or click here.

Webinar slides: Download