All content

Search filter terms
Filter by category
Filter by type
Filter by tag
Filter by user
Filter by licence
Filter by group
Filter by wsdl
Results per page:
Sort by:
Showing 40 results. Use the filters on the left and the search box below to refine the results.
Category: Workflow Licence: by-sa Group: SCAPE

Workflow myExperiment Group Activity Report (1)

Thumb
Prepares a report (three CSV files) on the activity of a group on myExperiment.

Created: 2014-08-26

Credits: User Donal Fellows

Uploader

Workflow ARC to WARC Migration with CDX Index and w... (1)

Thumb
Workflow for migrating ARC to WARC and comparing the CDX index files (Linux). The workflow has an input port “input_directory” which is a local path to the directory containing the ARC files, and an input port “output_directory” which is the directory where the workflow outputs are created. The files in the input directory are migrated using the “arc2warc_migration_cli” tool service component to perform the migration. The “cdx_creator_arc” and “cdx_creator_warc” tool service components creat...

Created: 2014-07-09

Credits: User Sven

Uploader

Workflow TIF to JP2 file format migration with qual... (1)

Thumb
This workflow reads a textfile containing absolute paths to TIF image files and converts them to JP2 image files using OpenJPEG (https://code.google.com/p/openjpeg). Based on the input text file, the workflow creates a Taverna list to be processed file by file. A temporary directory is created (createtmpdir) where the migrated image files and some temporary tool outputs are stored. Before starting the actual migration, it is checked if the TIF input images are valid file format instances u...

Created: 2014-05-07 | Last updated: 2014-05-07

Credits: User Sven

Uploader

Workflow ARC to WARC Migration and CDX Index Compar... (1)

Thumb
Workflow for migrating ARC to WARC and comparing the CDX index files (Linux). The workflow has an input port “input_directory” which is a local path to the directory containing the ARC files, and an input port “output_directory” which is the directory where the workflow outputs are created. The files in the input directory are migrated using the “arc2warc_migration_cli” tool service component to perform the migration. The “cdx_creator_arc” and “cdx_creator_warc” tool service components creat...

Created: 2014-04-23

Credits: User Sven

Uploader

Workflow ARC2WARC Hadoop Job (1)

Thumb
Just a wrapper workflow for a Hadoop job converting ARC to WARC files.

Created: 2014-03-06

Credits: User Sven

Workflow Preservation Action Plan sampler (2)

Thumb
This is an example of the structure of a SCAPE Preservation Action Plan.

Created: 2014-03-05 | Last updated: 2014-03-05

Uploader

Workflow ToMaR HDFS Input Directory Processing (2)

Thumb
This workflow allows processing an HDFS input directory using ToMaR. The "hdfs_working_dir" input port is the HDFS input directory which containes the data to be processed by ToMaR. The "toolspec" input port contains the toolspec XML describing operations that can be used (see "operation" input port). The "operation" input port defines the operation to be used in the current ToMaR job execution (see "toolspec" input port, an operation port used here must be defined in the tool specificatio...

Created: 2014-03-04 | Last updated: 2014-03-11

Credits: User Sven

Workflow Slim Migrate And QA mp3 to Wav Using Hadoo... (4)

Thumb
This workflow migrates an input list (available on HDFS) of mp3 files (available on NFS) to wav files (in output directory on NFS) using an ffmpeg Hadoop job. The workflow then compares content of the original mp3 and the migrated wav by first converting the two files to wav using an mpg123 Hadoop job and the identity function respectively, and then using an xcorrSound waveform-compare Hadoop job. The needed Hadoop jobs are available from https://github.com/statsbiblioteket/scape-audio-qa-ex...

Created: 2014-02-21 | Last updated: 2014-06-30

Uploader

Workflow ONB Web Archive Fits Characterisation usin... (2)

Thumb
Hadoop based workflow for applying FITS on the files contained in ARC web archive container files and ingest the FITS output in a MongoDB using C3PO. Dependencies: - Spacip (https://github.com/shsdev/spacip) - Tomar (https://github.com/openplanets/tomar) - C3PO (https://github.com/peshkira/c3po) Parameters: - hdfs_input_path: Path to a directory which contains textfile(s) with absolute HDFS paths to ARC files - num_files_per_invokation: Number of items to be processed per invokation - fits...

Created: 2013-12-09 | Last updated: 2013-12-10

Credits: User Sven

Workflow MeasuresDocCombiner (2)

Thumb
Added namespace

Created: 2013-10-28 | Last updated: 2014-07-23

Results per page:
Sort by: