Workflows in this Topic



Workflow Cow-Human Ortholog Pathways and Gene annot... (2)

Thumb
This workflow searches for genes which reside in a QTL (Quantitative Trait Loci) region in the cow, Bos taurus. The workflow requires an input of: a chromosome name or number; a QTL start base pair position; QTL end base pair position. Data is then extracted from BioMart to annotate each of the genes found in this region. As the Cow genome is currently unfinished, the workflow subsequently maps the cow ensembl gene ids to human orthologues. Entrez and UniProt identifiers are then identified...

Created: 2007-10-03 | Last updated: 2009-12-03

Workflow dreseden-term-extraction1 (1)

Thumb
No description

Created: 2007-11-22

Credits: User Simon Jupp

Uploader

Workflow Termine Webservice (1)

Thumb
Termine is a service provided by the National Centre for Text Mining (NaCTeM) to assist in the discovery of terms in text. More information on the Termine service can be found here. This workflow represents the simplest method of using Termine. The input represents a text string with the output being an string containing a representation of the list of terms, with their C-Value scores (representing significance in the text), in a simple xml format. Other variations of this tools will be adde...

Created: 2008-05-19 | Last updated: 2008-05-19

Credits: User Brian Rea Network-member National Centre for Text Mining (NaCTeM)

Workflow EBI_Whatizit (1)

Thumb
Perform a text-mining analysis of an input text document using the EBI's Whatizit tool (http://www.ebi.ac.uk/webservices/whatizit/info.jsf). Whatizit provides a number of text-mining pipelines which can can detect various terms of biological interest in text documents. For example finding gene names and mapping them to UniProtKB identifiers, finding chemical terms and mapping them to ChEBI, etc.

Created: 2008-07-09

Credits: User Hamish McWilliam

Workflow extract_uniprot_embl_gi.xml (1)

Thumb
No description

Created: 2008-07-12 | Last updated: 2008-07-12

Credits: User Paul Fisher

Workflow Ensembl_id_2_Swissport_id.xml (1)

Thumb
No description

Created: 2008-07-12 | Last updated: 2008-07-12

Credits: User Paul Fisher

Workflow genes_from_probesets.xml (1)

Thumb
No description

Created: 2008-07-12 | Last updated: 2008-07-12

Credits: User Paul Fisher

Workflow genomic_sequence.xml (1)

Thumb
No description

Created: 2008-07-12 | Last updated: 2008-07-12

Credits: User Paul Fisher

Workflow Execute cmd line app - unix - /bin/ls -R o... (1)

Thumb
Note that this workflow only works on Unix systems. The get_temporary_directory beanshell returns the path to the directory in which temporary files are held. This value is passed to the create_and_populate_list beanshell which creates a list of '-R' and the path. The resultant list is passed to the args port of the Execute_cmd_line_app service. The command port of that service has been defaulted to '/bin/ls'. The execute_cmd_line_app service runs the /bin/ls -R command on the temporary ...

Created: 2008-09-27

Credits: User Alan Williams

Workflow color_pathway_by_elements (2)

Thumb
Color the objects (rectangles and circles on a pathway map) corresponding to the given 'element_id_list' with the specified colors and return the URL of the colored image. Example of inputs: pathway_id: path:bsu00010 element_list: (to add as list) 78 79 fg_color_list:(to add as list) red blue bg_color_list: (to add as list) green yellow

Created: 2008-09-30 | Last updated: 2008-10-01

Credits: User Franck Tanoh

Workflow SigWin-detector Config-Basic (2)

Thumb
Detects significant windows in a sequence. Input: A space-delimited file with (at least) one column containing the input sequence E= {E1, E2, …, EN}. A two-line header should precede the data. Input file example. Output: (1) A file containing the detected significant windows for each label. Each data row represents a stretch of consecutive significant windows. Column 1 gives the window size and columns 2 and 3 give the first and last significant windows in the stretch. Output file e...

Created: 2009-10-09 | Last updated: 2010-04-23

Credits: User Adambel

Uploader

Workflow caDSR metadata query in caGrid (1)

Thumb
This workflow shows the coordinated use of two services in CaGrid, i.e., the caDSR (Cancer Data Standards Repository) and EVS (Enterprise Vocabulary Services) services. caDSR is to define a comprehensive set of standardized metadata descriptors for cancer research terminology used in information collection and analysis. EVS provides resources and services to meet NCI needs for controlled terminology, and to facilitate the standardization of terminology and information systems across the Insti...

Created: 2008-12-05 | Last updated: 2008-12-05

Credits: User Wei Tan

Uploader

Workflow hierarchical microarray clustering (1)

Thumb
To illustrate our caGrid plug-in’s application, we tested it with a microarray hierarchical clustering workflow that involves services hosted at multiple institutions. Microarrays are a high-throughput technology used to measure the expression of tens of thousands of genes in different tissues or cells. Scientists represent the data from each microarray via a vector (profile) in which each element represents a gene’s expression level. They use clustering analysis to identify sim...

Created: 2008-12-05 | Last updated: 2008-12-05

Credits: User Wei Tan

Uploader

Workflow Using CQL to query protein sequence data (1)

Thumb
To query protein sequence infomation out of 3 caGrid data services: caBIO, CPAS and GridPIR. Scientific value To query protein sequence information out of 3 caGrid data services: caBIO, CPAS and GridPIR. To analyze a protein sequence from different data sources. Steps Querying CPAS and get the id, name, value of the sequence. Querying caBIO and GridPIR using the id or name obtained from CPAS.    

Created: 2008-12-05 | Last updated: 2009-07-14

Credits: User Wei Tan

Uploader

Workflow Lymphoma type prediction based on microar... (7)

Thumb
Scientific value Using gene-expression patterns associated with DLBCL and FL to predict the lymphoma type of an unknown sample. Using SVM (Support Vector Machine) to classify data, and predicting the tumor types of unknown examples. Steps Querying training data from experiments stored in caArray. Preprocessing, or normalize the microarray data. Adding training and testing data into SVM service to get classification result.

Created: 2010-05-11 | Last updated: 2010-05-11

Credits: User Wei Tan User Ravi User Stian Soiland-Reyes

Workflow Using CQL to query protein sequence data (1)

Thumb
To query protein sequence infomation out of 3 caGrid data services: caBIO, CPAS and GridPIR Adapted from http://www.myexperiment.org/workflows/600

Created: 2009-05-07

Credits: User Stian Soiland-Reyes

Attributions: Workflow Using CQL to query protein sequence data

Workflow group difference pseudopotential (1)

Thumb
This is a Nimrod plan file for a parameter scan of a group difference pseudopotential (GDP). This is the initial step in "paremeterization of a capping atom for hybrid quantum mechanics-molecule mechanics (QM/MM) calculations." Based on paper by Sudholt, W., Baldridge, K., Abramson, D., Enticott, C. and Garic, S., “Application of Grid computing to parameter sweeps and optimizations in molecular modeling”, Future Generation Computer Systems, 21 (2005), 27-35. Also appeare...

Created: 2009-05-14 | Last updated: 2009-05-16

Workflow FirstExampleWorkflow (1)

Thumb
This is my first test example of Taverna Workflow.

Created: 2009-06-24

Credits: User Jelena (Obradovic) Dreskai

Uploader

Workflow Term Extraction with NaCTeM's TerMine Tool (2)

Thumb
Pass in text and retrieve a list of terms discovered ranked by their importance within the text.

Created: 2009-07-23

Credits: User Brian Rea Network-member National Centre for Text Mining (NaCTeM)

Uploader

Workflow An example workflow with WSRF service (1)

Thumb
An example workflow with WSRF service The service used is http://sidgrid.ci.uchicago.edu:8100/wsrf/services/CounterService The service is included in Globus toolkit installation (ws-core). If the it is down you can host your own and modify the workflow definition file to point to your own. The workflow first creates a counter instance and adds the value 10 for two times to the same counter. Therefore the result should be 20.

Created: 2009-07-28

Credits: User Wei Tan User Stian Soiland-Reyes

Uploader

Workflow caDSR Data service query in caGrid (3)

Thumb
This workflow is used as an example in this wiki article: https://wiki.cagrid.org/display/knowledgebase/How+to+Create+CaGrid+Workflow+Using+Taverna+2   Tested with Taverna 2.1.2 as of 6/10/2010

Created: 2010-05-25 | Last updated: 2010-06-11

Credits: User Wei Tan

Uploader

Workflow caArray data retrieving (1)

Thumb
Query all the gene expression data in a caArray experiment. Returns a evenly divided gene expression data set with corresponding class information. They ca be later used as training and test data set in many classification algorithms.Query all the gene expression data in a caArray experiment. Returns a evenly divided gene expression data set with corresponding class information. They can be later used as training and test data set in many classification algorithms.

Created: 2009-11-23

Credits: User Wei Tan

Uploader

Workflow genePattern data preprocessing (2)

Thumb
preprocess data set using genePattern preProces service, the input should be in genePattern STATML format. Configuration parameters can be adjusted by changing the default preprocess data set using genePattern preProces service, the input should be in genePattern STATML format.preprocess data set using genePattern preProces service, the input should be in genePattern STATML format. Configuration parameters can be adjusted by changing the string constants.

Created: 2010-05-24 | Last updated: 2010-05-24

Credits: User Wei Tan

Uploader

Workflow Support-Vector-Machine (SVM) based data cl... (2)

Thumb
Support-Vector-Machine based data classificationSupport-Vector-Machine based data classification using genePattern SVM service, the input should be in genePattern STATML format.

Created: 2010-05-24 | Last updated: 2010-05-24

Credits: User Wei Tan

Workflow Clean plain text (ASCII) (1)

Thumb
This workflow will remove any XML-invalid and non-ASCII characters (e.g. for sending to the ASCII-only Termine service) from any text supplied to the input port. This is a workflow component, designed to be used as a nested workflow inside a larger text mining or text processing workflow.

Created: 2010-02-18 | Last updated: 2011-12-13

Credits: User James Eales

Workflow Clean plain text (1)

Thumb
This workflow will remove any XML-invalid characters (these characters often appear in the output of PDF to text software) from any text supplied to the input port. This is a workflow component, designed to be used as a nested workflow inside a larger text mining or text processing workflow.  

Created: 2010-02-18 | Last updated: 2011-12-13

Credits: User James Eales

Workflow Load plain text from directory (1)

Thumb
This workflow will automate the reading of a set of text files stored in a single directory (the path to which should be supplied as a single input value).  It will assume that the text files are saved using the default character encoding for the system that Taverna is running on.  This is a workflow component, designed to be used as a nested workflow inside a larger text mining or text processing workflow.  

Created: 2010-02-18 | Last updated: 2011-12-13

Credits: User James Eales

Workflow Load PDF from directory (1)

Thumb
This workflow will automate the reading of a set of PDF files stored in a single directory (the path to which should be supplied as a single input value). This is a workflow component, designed to be used as a nested workflow inside a larger text mining or text processing workflow.  

Created: 2010-02-19 | Last updated: 2011-12-13

Credits: User James Eales

Workflow PDF to plain text (1)

Thumb
This workflow will extract the plain text content of PDF files supplied to the input port.  You can connect the Load PDF from directory workflow to this workflows input. We recommend you send the output from this workflow to the Clean plain text workflow, because the PDF to text process can add characters into the text that are XML-invalid and therefore can not be sent to most services as plain text.  Another way round this problem is to encode the text as Base64 using the handy loc...

Created: 2010-02-19 | Last updated: 2011-12-13

Credits: User James Eales

Workflow Sentence splitting (1)

Thumb
This workflow will attempt to split up text into sentences, returning a list of sentences to the output port.  The sentence splitting service makes use of the OpenNLP sentence detector and has been trained to work on english text. This workflow can be used to provide input to the Termine with c-value threshold workflow. This is a workflow component, designed to be used as a nested workflow inside a larger text mining or text processing workflow.

Created: 2010-02-19 | Last updated: 2011-12-13

Credits: User James Eales

Workflow Termine with c-value threshold (1)

Thumb
This workflow accepts a list of sentences from a single document and returns the terms found by the TerMine web service. It also allows you to set a threshold c-value score so that only terms with a user-controlled probability (of being a real term) are returned as an output.   To get sentences to supply to this workflow you can use the sentence splitting workflow.  The TerMine service (used in this workflow) only accepts text in ASCII encoding, so you should also use the Clean p...

Created: 2010-02-19 | Last updated: 2011-12-13

Credits: User James Eales

Workflow Terms from collection of PDF files (2)

Thumb
This workflow will give you a set of candidate terms for each PDF document in a user-specified directory. You can also specify a c-value threshold that will restrict the terms to those with higher scores. This workflow was created using only nested workflows.  These workflow components work on their own and can be linked together to form more complex workflows such as this. You can view the text mining workflow components in this pack. If you receive errors when running this workflow t...

Created: 2010-02-19 | Last updated: 2011-12-13

Credits: User James Eales

Workflow Terms from collection of text files (1)

Thumb
This workflow will give you a set of candidate terms for each text file in a user-specified directory. You can also specify a c-value threshold that will restrict the terms to those with higher scores. This workflow was created using only nested workflows.  These workflow components work on their own and can be linked together to form more complex workflows such as this. You can view the text mining workflow components in this pack. If you receive errors when running this workflow then...

Created: 2010-02-22 | Last updated: 2011-12-13

Credits: User James Eales

Uploader

Workflow Get locations from postcode (1)

Thumb
 This workflow will return all the areas that correspond to a postcode. Enter the first part of the postcode only, e.g. SW19

Created: 2010-03-08 | Last updated: 2010-03-08

Credits: User Rory

Workflow Parse unique proteins from Blast file (1)

Thumb
The workflow parses uses the blast results to determine the unique proteins found in the target genome that have no similairty to the source genome. Using these unique protein ids, and the original target protein fasta file, a fasta file of unique proteins is created.

Created: 2010-03-19 | Last updated: 2010-03-19

Credits: User Ian Laycock Network-member nclteamc

Workflow Retrieve all data for all instruments for ... (3)

Thumb
This workflow looks up all instruments known to have worked during the requested periode of time (ICS), restructures results to make them fit instruments.xsd and requests data for this periode from DPAS. Result: VOTable with list of URLs to data.

Created: 2010-04-15 | Last updated: 2010-08-09

Credits: User Anja Le Blanc

Attributions: Workflow Extracting data from VOTable format by using XPath query

Workflow Image Mining with RapidMiner (1)

Thumb
This is an image mining process using the image mining Web service provided by NHRF within e-Lico. It first uploads a set of images found in a directory, then preprocesses the images and visualizes the result. Furthermore, references to the uploaded images are stored in the local RapidMiner repository so they can later be used for further processing without uploading images a second time.

Created: 2010-04-28 | Last updated: 2012-01-16

Uploader

Workflow Query caArray data service and retrieving ... (2)

Thumb
need to install Taverna 2 caGrid integration suite from http://www.mcs.anl.gov/~wtan/t2/ and get a cagrid Dorian account (see http://wiki.cagrid.org/display/caGrid13/Home)

Created: 2010-05-24 | Last updated: 2010-05-24

Credits: User Wei Tan

Uploader

Workflow Federated query using DCQL and credential ... (2)

Thumb
CDS_Activity issues an EPR of the delegated credential. FQP uses this EPR to fetch the actual delegated credential from CDS and uses it to invoke multiple data services (the query activity) on behalf of the invoker. CDS_Activity issues an EPR of the delegated credential. FQP uses this EPR to fetch the actual delegated credential from CDS and uses it to invoke multiple data services (the query activity) on behalf of the invoker. Need to install Taverna 2 caGrid integration suite from http://ww...

Created: 2010-05-11 | Last updated: 2010-11-05

Credits: User Wei Tan

Uploader

Workflow Invoke a secured caGrid service: caTissue (2)

Thumb
No description

Created: 2010-05-24 | Last updated: 2010-11-05

Credits: User Wei Tan

Uploader

Workflow A simple CQL query workflow in caGrid (1)

Thumb
1. CQL is a language to query data from caGrid/caBIG services. This workflow is tested with Taverna 2.1.2 and the caGrid Workflow Suite downloadable from http://www.mcs.anl.gov/~wtan/t2/. 2.More information regarding CQL can be found from http://wiki.cagrid.org/display/dataservices. 3. Sample input (95) is provided in the workflow. It is to query all the hybridization data within a microarray experiment whose id is 95.

Created: 2010-05-24 | Last updated: 2010-05-25

Credits: User Wei Tan

Workflow CamelCases (1)

Thumb
this process splits up camelcases

Created: 2010-06-02

Uploader

Workflow Connect to twitter and analyze the key words (1)

Thumb
Hi All, This workflow connects RapidMiner to Twitter and downloads the timeline. It then creates a wordlist from the tweets and breaks them into key words that are mentioned in the tweets. You can then visualize the key words mentioned in the tweets. This workflow can be further modified to review various key events that have been talked about in the twitterland. Do let me know your feedback and feel free to ask me any questions that you may have. Shaily web: http://advanced-analyti...

Created: 2010-07-26 | Last updated: 2010-07-26

Uploader

Workflow Testing caBIG workflow (1)

Thumb
An example workflow to test the installation of caGrid Workflow Suite and the submission of a workflow to a caGrid workflow service.

Created: 2010-08-17 | Last updated: 2010-08-17

Credits: User Wei Tan

Uploader

Workflow Principle Component Analysis (PCA) over mi... (1)

Thumb
Principle Component Analysis (PCA) over microarray data. Data is uploaded through caGrid transfer ultility.

Created: 2010-09-10 | Last updated: 2010-09-10

Credits: User Wei Tan

Uploader

Workflow From PDF to lemmatized text (1)

Thumb
This workflow uses the web service stationed in JSI (IJS Slovenia), which is based on Matjaž Juršič's LemmaGen - lemmatization engine. The workflow accepts a PDF file as an input an uses James Eales's wrokflows to preprocess the data. The workflow interactively asks the user of which language is the text, since the lemmatization process is language based. The output is a string in Taverna Workbench.

Created: 2010-09-16 | Last updated: 2012-01-18

Credits: User Netr User James Eales

Attributions: Workflow PDF to plain text Workflow Clean plain text

Workflow Lemmatization (3)

Thumb
The workflow lemmatizes the text in the input port. Takes text as input and returns (language dependent) lemmatized text as output. All the words in the resulting text are in the same order as in the original text, but they are transformed to their dictionary form. The workflow asks for the language of lemmatization. Currently, 12 languages are supported: en,sl,ge,bg,cs,et,fr,hu,ro,sr,it,sp.

Created: 2010-12-17 | Last updated: 2010-12-23

Credits: User Petra Kralj Novak

Attributions: Workflow Select from a list of possible web service parameter values

Workflow miRNA GFF to entrez gene (1)

Thumb
This workflow reads a GFF file of miRNA cooridinates and uses BioMart to search human ensemble genes for the gene that codes for the miRNA. The workflow returns a list of miRNAid, chromosome, start, stop, strand, entrez gene id, gene name, gene strand. Example input file here: ftp://mirbase.org/pub/mirbase/CURRENT/genomes/hsa.gff

Created: 2011-01-26 | Last updated: 2012-01-11

Workflow Random recommender (1)

Thumb
This process does a random item recommendation; for a given item ID, from the example set of items, it randomly recommends a desired number of items. The purpose of this workflow is to produce a random recommendation baseline for comparison with different recommendation solutions, on different retrieval measures. The inputs to the process are context defined macros: %{id} defines an item ID for which we would like to obtain recommendation and %{recommender_no} defines the required number of ...

Created: 2011-03-15 | Last updated: 2011-03-15

Workflow Item to item similarity matrix -based reco... (1)

Thumb
This process executes the recommendation based on item to item similarity matrix. The inputs to the process are context defined macros: %{id} defines an item ID for which we would like to obtain recommendation and %{recommender_no} defines the required number of recommendations. The process internally uses an item to item similarity matrix written in pairwise form (id1, id2, similarity). The process essentially filters out appearances of the required ID in both of the columns of the pairwis...

Created: 2011-03-15 | Last updated: 2011-03-15

Workflow Content based recommender (1)

Thumb
This process is a special case of the item to item similarity matrix based recommender where the item to item similarity is calculated as cosine similarity over TF-IDF word vectors obtained from the textual analysis over all the available textual data. The inputs to the process are context defined macros: %{id} defines an item ID for which we would like to obtain recommendation and %{recommender_no} defines the required number of recommendations. The process internally uses an example set of...

Created: 2011-03-15 | Last updated: 2011-03-15

Workflow Collaborative filtering recommender (1)

Thumb
This process executes a collaborative filtering recommender based on user to item score matrix. This recommender predicts one user’s score on some of his non scored items based on similarity with other users. The inputs to the process are context defined macros: %{id} defines an item ID for which we would like to obtain recommendation and %{recommender_no} defines the required number of recommendations and %{number_of_neighbors} defines the number of the most similar users taken into a...

Created: 2011-03-15 | Last updated: 2012-03-06

Workflow GRASS-GIS orchestration using pyWPS (2)

Thumb
Generic workflow that run r.watershed, with auxiliary services: r.math and geotiff2png. Watershed accumulation is calculated from DEM using r.watershed, the accumulation result is then filtered using r.math with equation:output=(if(a>10,a,null())) Generic workflow that run r.watershed, with auxiliary services: r.math and geotiff2png. Watershed accumulation is calculated from DEM using r.watershed, the accumulation result is then filtered using r.math with equation: output=(if(a>10,...

Created: 2011-04-18 | Last updated: 2011-04-25

Credits: User Jorgejesus

Workflow Agglomerative clustering of a GEO dataset ... (2)

Thumb
No description

Created: 2011-04-28 | Last updated: 2011-12-13

Credits: User James Eales

Workflow Distance heatmap of GEO dataset produced b... (2)

Thumb
No description

Created: 2011-04-28 | Last updated: 2011-12-13

Credits: User James Eales

Workflow Classification of GEO assays using RapidAn... (2)

Thumb
No description

Created: 2011-05-04 | Last updated: 2011-12-13

Credits: User James Eales

Workflow Content based recommender system template (1)

Thumb
As an input, this workflow takes two distinct example sets: a complete set of items with IDs and appropriate textual attributes (item example set) and a set of IDs of items our user had interaction with (user example set). Also, a macro %{recommendation_no} is defined in the process context, as a required number of outputted recommendations. The first steps of the workflow are to preprocess those example sets; select only textual attributes of item example set, and set ID roles on both of th...

Created: 2011-05-05 | Last updated: 2011-05-09

Credits: User Matko Bošnjak User Ninoaf

Attributions: Blob Datasets for the pack: RCOMM2011 recommender systems workflow templates

Workflow User-based collaborative filtering recomme... (1)

Thumb
The workflow for user-based collaborative filtering, takes only one example set as an input: a user-item matrix, where the attributes denote item IDs, and rows denote users. If a user i has rated an item j with a score s, the matrix will have the value s written in i-th row and j-th column. In the context of the process we define the ID of the user %{id}, desired number of recommendations %{recommendation_no}, and the number of neighbors used in ca...

Created: 2011-05-05 | Last updated: 2011-05-09

Credits: User Matko Bošnjak User Ninoaf

Attributions: Blob Datasets for the pack: RCOMM2011 recommender systems workflow templates

Workflow Item-based collaborative filtering recomme... (1)

Thumb
The workflow for item-based collaborative filtering receives a user-item matrix for its input, and the same context defined macros as the user-based recommender template, namely %{id}, %{recommendation_no}, and %{number_of_neighbors}. Although this process is in theory very similar to user-based technique, it differs in several processing steps since we are dealing with an item-user matrix, the transposed user-item example set. The first step of the workflow, after declaring zero values miss...

Created: 2011-05-05 | Last updated: 2011-05-09

Credits: User Matko Bošnjak User Ninoaf

Attributions: Blob Datasets for the pack: RCOMM2011 recommender systems workflow templates

Workflow One sentence per line (1)

Thumb
This workflow accepts a plain text input and provides a single text document per input containing one sentence per line.  Newline characters are removed from the original input. The OpenNLP sentence splitter is used to split the text, this is provided by University of Manchester Web Services.

Created: 2011-05-06 | Last updated: 2011-12-13

Credits: User James Eales

Uploader

Workflow LSI content based recommender system template (1)

Thumb
This workflow performs LSI text-mining content based recommendation. We use SVD to capture latent semantics between items and words and to obtain low-dimensional representation of items. Latent Semantic Indexing (LSI) takes k greatest singular values and left and right singular vectors to obtain matrix  A_k=U_k * S_k * V_k^T. Items are represented as word-vectors in the original space, where each row in matrix A represents word-vector of particular item. Matrix U_k, on the other hand ...

Created: 2011-05-06 | Last updated: 2011-05-09

Credits: User Ninoaf User Matko Bošnjak

Attributions: Workflow Content based recommender system template Blob Datasets for the pack: RCOMM2011 recommender systems workflow templates

Uploader

Workflow SVD user-based collaborative filtering rec... (1)

Thumb
This workflow takes user-item matrix A as a input. Then it calculates reduced SVD decomposition A_k by taking only k greatest singular values and corresponding singular vectors. This worfkflow calculates recommendations and predictions for particular user %{id} from matrix A. Particular row %{id} is taken from original matrix A and replaced with %{id} row in A_k matrix. Predictions are made for %{id} user based on another users A_k. Note: This workflow uses R-script operator with R library ...

Created: 2011-05-09 | Last updated: 2011-05-09

Credits: User Ninoaf User Matko Bošnjak

Attributions: Workflow User-based collaborative filtering recommender system template Blob Datasets for the pack: RCOMM2011 recommender systems workflow templates

Workflow Sequence Format Conversion (1)

Thumb
Demonstrate the use of the format conversion tool. Converts a DNA sequence in fasta format into plain, genbank, and phylip formats.

Created: 2011-06-28 | Last updated: 2011-06-28

Credits: User Carol Lushbough

Uploader

Workflow Statistics example (1)

Thumb
This workflow allows one to find the average and standard deviation of a set of values.

Created: 2011-09-21 | Last updated: 2011-09-21

Credits: User Wdsnellg

Uploader

Workflow Concatenated List of Values (1)

Thumb
This workflow takes multiple lists of values and finds the standard deviation and average of the provided values. It then concatenates each value into a single string.

Created: 2011-09-21 | Last updated: 2011-09-21

Credits: User tpacurtis

Uploader

Workflow New York vs. New Jersey Broccoli Production (1)

Thumb
Uses the SpreadsheetImport service to import data from an Excel spreadsheet. The workflow imports two files and generates a single graph comparing the information from the two spreadsheets.

Created: 2011-09-21 | Last updated: 2011-09-21

Credits: User tpacurtis User Alan Williams

Attributions: Workflow Spreadsheet Import Example Blob NY Broccoli Production Blob NJ Broccoli Production

Uploader

Workflow Simple addition Kepler workflow expanded (1)

Thumb
This Kepler allows one to perform mathematical equations, including addition, subtraction, muliplication, and division. 

Created: 2011-09-27 | Last updated: 2011-09-27

Credits: User Wdsnellg

Workflow FFmpeg convert audio2aac (REST) (1)

Thumb
Converts supported audio files to AAC using FFmpeg through a REST webservice.

Created: 2012-01-16 | Last updated: 2012-01-16

Credits: User Rui Castro

Workflow Iterate through datasets (1)

Thumb
This is a dataset iteration workflow. It is a part of Experimentation workflow for recommender extension. Loop FIles operator iterates through datasets from a specified directory using read aml operator. Only datasets specified with a proper regular expression are considered. Train and test data filenames must correspond e.g (train1.aml, test1.aml). In each iteration Loop Files calles specified operator testing workflow with Execute subprocess operator. Informations about training and t...

Created: 2012-01-29

Credits: User Matej Mihelčić User Matko Bošnjak

Workflow Metafeature extraction (1)

Thumb
This is a metafeature extraction workflow used in Experimentation workflow for recommender extension operators. This workflow extracts metadata from the train/test datasets (user/item counts, rating count, sparsity etc). This workflow is called from the operator testing workflow using Execute Process operator.

Created: 2012-01-29 | Last updated: 2012-01-30

Credits: User Matko Bošnjak

Workflow Operator testing workflow (1)

Thumb
This workflow is used for operator testing. It joins dataset metafeatures with execution times and performanse measures of the selected recommendation operator. In the Extract train and Extract test Execute Process operator user should open Metafeature extraction workflow. In the Loop Operator train/test data are used to evaluate performanse of the selected operator. Result is remebered and joined with the time and metafeature informations. This workflow can be used both for Item Recommend...

Created: 2012-01-29

Credits: User Matej Mihelčić User Matko Bošnjak

Uploader

Workflow Semantic clustering (with alpha-clustering... (1)

Thumb
The workflow uses RapidMiner extension named RMonto (http://semantic.cs.put.poznan.pl/RMonto/) to perform clustering of SPARQL query results based on chosen semantic similarity measure. The measure used in this particualr workflow is a kernel that exploits membership of clustered individuals to OWL classes from a background ontology ("Epistemic" kernel from [1]). Since the semantics of the backgound ontology is used in this way, we use the name "semantic clustering". This ...

Created: 2012-01-29 | Last updated: 2012-01-30

Workflow Recommender workflow (1)

Thumb
This is a main online update experimentation workflow. It consists of three Execute Process operators. First operator executes model training workflow. Second operator executes online updates workflow for multiple query update sets. The last operator executes performance testing and comparison workflow. Final performance results are saved in an Excel file.

Created: 2012-01-29

Credits: User Matej Mihelčić

Workflow Model saving workflow (1)

Thumb
This workflow trains and saves a model for a selected item recommendation operator.

Created: 2012-01-29 | Last updated: 2012-01-30

Credits: User Matej Mihelčić

Workflow Model testing workflow (1)

Thumb
This workflow measures performance of three models. Model learned on train data and upgraded using online model updates. Model learned on train data + all query update sets. Model learned on train data only.

Created: 2012-01-29

Credits: User Matej Mihelčić

Workflow Data iteration workflow (1)

Thumb
This is a data iteration workflow used to iterate throug query update sets.

Created: 2012-01-29

Credits: User Matej Mihelčić User Matko Bošnjak

Workflow Model update workflow (1)

Thumb
This is a Model update workflow called from data iteration workflow on every given query set. In the Loop operator model and current training set are retrieved from the repository. Model update is performed on a given query set creating new model. Model and updated train set are saved in the repository.

Created: 2012-01-29 | Last updated: 2012-01-29

Credits: User Matej Mihelčić User Matko Bošnjak

Workflow Model saving workflow (RP) (1)

Thumb
This workflow trains and saves model for a selected rating prediction operator.

Created: 2012-01-29 | Last updated: 2012-01-30

Credits: User Matej Mihelčić

Workflow Model testing workflow (RP) (1)

Thumb
This workflow measures performance of three models. Model learned on train data and upgraded using online model updates. Model learned on train data + all query update sets. Model learned on train data only.

Created: 2012-01-29 | Last updated: 2012-01-30

Credits: User Matej Mihelčić

Workflow recommender workflow (RP) (1)

Thumb
This is a main online update experimentation workflow. It consists of three Execute Process operators. First operator executes model training workflow. Second operator executes online updates workflow for multiple query update sets. The last operator executes performance testing and comparison workflow. Final performance results are saved in an Excel file.

Created: 2012-01-29

Credits: User Matej Mihelčić

Workflow Model update workflow (RP) (1)

Thumb
This is a Model update workflow called from data iteration workflow on every given query set. In the Loop operator model and current training set are retrieved from the repository. Model update is performed on a given query set creating new model. Model and updated train set are saved in the repository.

Created: 2012-01-29 | Last updated: 2012-01-30

Credits: User Matej Mihelčić

Workflow Data iteration workflow (RP) (1)

Thumb
This is a data iteration workflow used to iterate throug query update sets.

Created: 2012-01-29

Credits: User Matej Mihelčić User Matko Bošnjak

Uploader

Workflow document clustering (1)

Thumb
Tesla Experiment Clustering wikipedia articles of different topics Evaluation of the results with the purity measure and rand index

Created: 2012-02-15 | Last updated: 2012-02-17

Credits: User AGeduldig

Workflow VAMDC VALD query with SME processing (2)

Thumb
VAMDC workflow that queries the VALD Atomic database in Moscow and runs SME (Spectroscopy Made Easy) VAMDC Taverna Plugin: http://voparis-twiki.obspm.fr/twiki/bin/view/VAMDC/TavernaUserGuide

Created: 2012-02-24 | Last updated: 2012-09-12

Workflow VAMDC VALD query with SME processing (4)

Thumb
VAMDC workflow that queries the VALD Atomic database in Moscow and runs SME (Spectroscopy Made Easy) VAMDC Taverna Plugin: http://voparis-twiki.obspm.fr/twiki/bin/view/VAMDC/TavernaUserGuide

Created: 2012-03-07 | Last updated: 2012-11-29

Credits: User Stormshadow Network-member Virtual Atomic and Molecular Data Centre - VAMDC

Workflow Workflow to Query TAP-XSAMS node. (3)

Thumb
Workflow to just one TAP-XSAMS Node. This workflow queries the Hitran Node for MoleculeChemicalName of 'HF' VAMDC Taverna Plugin: http://voparis-twiki.obspm.fr/twiki/bin/view/VAMDC/TavernaUserGuide

Created: 2012-03-07 | Last updated: 2012-09-12

Credits: User Stormshadow

Workflow VAMDC to query on ChemicalName 'OR' Stoich... (4)

Thumb
Workflow to query just one TAP-XSAMS Node. This workflow queries the Hitran Node for MoleculeChemicalName of 'HF' and StoichimetricFormula of 'HV' Demonstrates building a query for two Restrictables. VAMDC Taverna Plugin: http://voparis-twiki.obspm.fr/twiki/bin/view/VAMDC/TavernaUserGuide

Created: 2012-03-07 | Last updated: 2012-09-12

Credits: User Stormshadow Network-member Virtual Atomic and Molecular Data Centre - VAMDC

Workflow VAMDC Query to Two Nodes using HTML Consum... (2)

Thumb
Workflow to two TAP-XSAMS Node and see results in a HTML view using a Consumer Service. This workflow queries the Hitran and CDMS Node for MoleculeStoichiometricForumua of 'CO' VAMDC Taverna Plugin: http://voparis-twiki.obspm.fr/twiki/bin/view/VAMDC/TavernaUserGuide

Created: 2012-03-07 | Last updated: 2012-09-12

Credits: User Stormshadow Network-member Virtual Atomic and Molecular Data Centre - VAMDC

Workflow VAMDC SpectCol tool using Full VAMDC-Tap Q... (2)

Thumb
Runs the SpectCol tool bundled with the plugin and can handle a list of URLS. This version asks for the full vamdc URL. SpectCol information: SPECTCOL is a tool that implements VAMDC standards, r etrieve relevant information from different databases such as CDMS, HITRAN, BASECOL, and can upload local files. All transfer of data between the client and the databases use the VAMDC-XSAMS schema. The spectroscopic and collisional information is combined and useful outputs (ascii or xsams) are ...

Created: 2012-03-07 | Last updated: 2012-09-12

Credits: User Stormshadow

Workflow VAMDC Query to Two Nodes using HTML Consum... (2)

Thumb
Workflow to two TAP-XSAMS Node and see results in a HTML view using a Consumer Service. This workflow queries the BaseCol and CDMS Node for MoleculeStoichiometricForumua of 'CO' VAMDC Taverna Plugin: http://voparis-twiki.obspm.fr/twiki/bin/view/VAMDC/TavernaUserGuide **NOTE an update to the plugin will need to be released later to handle prefixes on certain query nodes and their restrictables. For instance BaseCol expects 'Target.MoleculeStoichiometricFormula' instead of just 'MoleculeSto...

Created: 2012-03-07 | Last updated: 2012-09-12

Credits: User Stormshadow Network-member Virtual Atomic and Molecular Data Centre - VAMDC

Workflow Hybrid recommendation system (1)

Thumb
This is one hybrid recommendation system combining linear regression recommender, created using RapidMiner core operators, and Recommender extension multiple collaborative filtering and attribute based operators.

Created: 2012-05-17

Credits: User Matej Mihelčić User Matko Bošnjak

Workflow Get concept information (5)

Thumb
The workflow takes a (list of) concept ID(s) as input and returns the profile, ID, definition and name of the concept.

Created: 2012-07-06 | Last updated: 2014-07-14

Credits: User Kristina Hettne User Reinout van Schouwen User Marco Roos User Martijn Schuemie Network-member BioSemantics