
| | Besides the component identifier, the registry MUST provide sufficient information for a client to configure and execute the component on the platform using the JSS. \\ | Platform Application Registry | D5.2 | Split from previous requirement |
| | The identification mechanism used by the application registry MUST be applicable to composite applications (i.e. workflows). | Platform Application Registry \\ | D5.2 | |
| | Services (like the Data Connector API) MUST provide a client component (e.g. in the form of an executable Java archive) that can be added to a JSS workflow as a sequential pre/post processing activity. This component should for example support the transfer of a set of SCAPE Digital Objects to a configurable HDFS location. The client component should be configurable based on command-line parameters. | | D5.2 | |
| | It is the application developer’s responsibility to organize the output data of a calculation (say a file characterization) in a way that it can be loaded into a desired data sink (say a SCAPE digital object repository) using the provided data transfer client libraries (e.g. supporting to ingest of METS records into the repository). This can be achieved, for example, by implementing a MapReduce application that generates METS records (as output) using the SCAPE Digital Object Model (Java API) and the SCAPE METSFileFormat for Hadoop. | | D5.2 | |
| | In the context of SCAPE, there is a significant difference between the role of a “component developer” ensuring that a particular tool or piece of functionality (e.g. ImageMagick convert) is available on the parallel execution environment and the role of an “application/workflow developer” that implements a scenario/use-case on the platform based on the available components. It is the responsibility of the workflow developer to ensure the integration of the various components used in the workflow (like data source, data cleaning, processing, and data sink components).In the context of SCAPE, there is a significant difference between the role of a “component developer” ensuring that a particular tool or piece of functionality (e.g. ImageMagick convert) is available on the parallel execution environment and the role of an \\
10 \\
“application/workflow developer” that implements a scenario/use-case on the platform based on the available components. It is the responsibility of the workflow developer to ensure the integration of the various components used in the workflow (like data source, data cleaning, processing, and data sink components). | | D5.2 | |
| | The JSS must provide a job specification language that is capable to enact applications developed for the different Hadoop data analytic frameworks/languages utilized in SCAPE. | Job Submission Service | D5.2 | |
| | The JSS must also be able to enact sequential programs like those required to export data sets from a digital object repository. | Job Submission Service | D5.2 | |
| | The JSS must be integrated with the Platform Application Registry allowing a user to select an application based on its SCAPE Component identifier. | Job Submission Service | D5.2 | |
| | The JSS must provide means to specify workflows that are composed from multiple SCAPE Components and other (sequential) applications running on the cluster. | Job Submission Service | D5.2 | |
| | The JSS must be implemented as a REST-based service that provides the functionality described in the Platform Architecture (SCAPE deliverable D4.1). | Job Submission Service | D5.2 | |