September 2004; revised March 2005 | Bertram Ludäscher†*, Ilkay Altintas†, Chad Berkley†, Dan Higgins†, Efrat Jaeger†, Matthew Jones†, Edward A. Lee§, Jing Tao†, Yang Zhao§
The paper "Scientific Workflow Management and the KEPLER System" by Bertram Ludäscher et al. discusses the evolution of scientific workflows and the KEPLER system, a community-driven, open-source project designed to support scientific workflows. The authors highlight the increasing importance of data and information in scientific disciplines, emphasizing the need for tools that facilitate the integration of distributed data and computational resources. They introduce the concept of scientific workflows, which are networks of analytical steps involving database access, data analysis, and computational tasks, and outline the characteristics and requirements of these workflows.
The paper presents several examples of scientific workflows, including promoter identification, mineral classification, and job scheduling, to illustrate the complexity and diversity of scientific workflows. It also compares scientific workflows with business workflows, noting their differences in focus, execution models, and design paradigms.
A key feature of KEPLER is its integration with web services, which allows seamless access to remote resources and services. The system includes a web service actor that can be instantiated from a URL, facilitating rapid prototyping and development of web service-based workflows. KEPLER also supports grid computing and other extensions, such as file transfer and job scheduling, to handle high-performance computing tasks.
The underlying PTOLEMY II system, on which KEPLER is based, employs an actor-oriented modeling paradigm, where workflows are composed of independent components called actors. This approach enhances reusability and flexibility, allowing for hierarchical modeling and behavioral polymorphism. The paper discusses the Process Network (PN) director, which provides a dataflow process network semantics, and the Synchronous Data-Flow (SDF) director, which optimizes scheduling and buffer management.
The authors conclude by outlining future research directions, including the need for higher-order constructs, third-party transfers, and semantic links to capture data semantics and improve workflow design. They emphasize the importance of making scientific workflows more reliable, fault-tolerant, and user-friendly, and highlight the potential of ontologies and checkpointing techniques to enhance workflow reproducibility and traceability.The paper "Scientific Workflow Management and the KEPLER System" by Bertram Ludäscher et al. discusses the evolution of scientific workflows and the KEPLER system, a community-driven, open-source project designed to support scientific workflows. The authors highlight the increasing importance of data and information in scientific disciplines, emphasizing the need for tools that facilitate the integration of distributed data and computational resources. They introduce the concept of scientific workflows, which are networks of analytical steps involving database access, data analysis, and computational tasks, and outline the characteristics and requirements of these workflows.
The paper presents several examples of scientific workflows, including promoter identification, mineral classification, and job scheduling, to illustrate the complexity and diversity of scientific workflows. It also compares scientific workflows with business workflows, noting their differences in focus, execution models, and design paradigms.
A key feature of KEPLER is its integration with web services, which allows seamless access to remote resources and services. The system includes a web service actor that can be instantiated from a URL, facilitating rapid prototyping and development of web service-based workflows. KEPLER also supports grid computing and other extensions, such as file transfer and job scheduling, to handle high-performance computing tasks.
The underlying PTOLEMY II system, on which KEPLER is based, employs an actor-oriented modeling paradigm, where workflows are composed of independent components called actors. This approach enhances reusability and flexibility, allowing for hierarchical modeling and behavioral polymorphism. The paper discusses the Process Network (PN) director, which provides a dataflow process network semantics, and the Synchronous Data-Flow (SDF) director, which optimizes scheduling and buffer management.
The authors conclude by outlining future research directions, including the need for higher-order constructs, third-party transfers, and semantic links to capture data semantics and improve workflow design. They emphasize the importance of making scientific workflows more reliable, fault-tolerant, and user-friendly, and highlight the potential of ontologies and checkpointing techniques to enhance workflow reproducibility and traceability.