Workflows QuickStart Part 2 - Configure workflow to run on your own data
FollowWelcome to the Workflows QuickStart Tutorial Part 2 - You'll learn how to set up the workflow in the configuration form to analyze two single entities of genomic data from the data table. When you run the workflow, Terra will generate a set of the two samples that you can use for further analysis.
The tutorial uses the same file format conversion workflow from Part 1, where everything was already set up to run. In Part 2, we'll walk through the workflow setup from scratch in the UI.
Contents
1. Set up general workflow options in configuration form
2. Set inputs (from the data table)
3. Set up workspace-level inputs
- Using direct file paths for inputs
4. Write output file paths to the data table
5. Launch and monitor workflow analysis
What to expect when your workflow finishes
|
|
---|---|
What you will learn How much will it cost? How long will it take? |
Right click to open the tutorial demo in a new tab
1. Set (confirm) workflow options
Verify default options
Most of the defaults are fine (in order from the top left down): you don't need to change the version of the workflow (dropdown menu), you want to run on inputs defined by the data table (radio button), and the "Root entity type" is "sample" (see the Step 1 dropdown in the screenshot below).
Step 2 Select Data (highlighted above in orange)
Notice that the workflow will run on all samples in the workspace data table by default and generate a set of that group of samples. This is useful if you want to run this or a downstream workflow on the same group of samples.
To choose some particular subset of samples, or to change the name Terra gives the set of samples, click on the blue "Select Data" button. You'll be taken to the Select Data form:
1.1. Select the "Choose specific rows to process" radio button and select both samples in the data table by clicking on the squares at left. This would be the place where you could select particular rows of data.
1.2. Give your set a memorable name
1.3. Confirm your selection by clicking the "OK" button
The next step tells the workflow exactly to find inputs in the "root entity" data table.
2. Set up workflow Inputs to read from data table

You need to give the WDL instructions about where to find each variable in the table. You will do this with a particular format in the attribute field.
2.1. Go to the first required variable that is blank - "InputCram".
2.2. Start typing this.
in the "Attributes" column and you will see a drop-down menu appear.
|
|
---|---|
The "this.something" format tells the workflow "go to the root entity type table and For example, The drop-down includes all the columns available in the "sample" data table. |
2.3. Select "this.CRAM" from the dropdown
2.4. Go to the next blank variable, "SampleName" and follow the same process
The SampleName variable corresponds to the "sample_id" option in the dropdown: it's the unique identifier for that sample from the first column of the sample table).
The "this.sample_id" format tells the WDL to find the value for the
SampleName variable in the sample_id column of the root entity table
2.5. Save your inputs by clicking the blue "Save" button at the top right of the form.
Next you'll configure workspace-level attributes, like reference or index FASTA files. You'll do this in the Inputs section of the configuration form, too.
|
|
---|---|
There is a dedicated place in the workspace data page for workspace-level metadata. For programmers, think of these as your global variables. They are values that don’t |
3. Set up workspace-level inputs
3.1. Go to the first workspace-level variable, "RefIndex"
3.2. Start typing "workspace." in the Attributes field
The "workspace.something" format tells the WDL the value for this
variable is in the "something" row of the "workspace data" table.
3.3. Select the right attribute from the dropdown
(hint - use the variable name in the variable column to figure out what to choose in the dropdown)
3.4. Repeat for every required workspace-level variable
3.5. Save your setup by clicking the blue "Save" button at the top right of the form
Using direct references to inputs in Google bucket (instead of the table)
"gs://url-to-file-in-bucket"
4. Write output file paths to the data table
4.1. Start in the Outputs tab
4.2. For the first output variable, "outputBai", go to the attribute field and type in "this." + a name for your output file.
In the dropdown, you will see all the columns from the "sample" data table. This includes the original columns in the data table (i.e. the ones you used for Input variables) as well as the columns the workflow created for the outputs generated in Part 1.
Try typing in a different output name for this run!
4.3. Save your output attributes by selecting the blue button at the top right. The Run Analysis button should turn blue (if it doesn't, you might need to go back and fill in an attribute, or save).
5. Launch and monitor analysis
5.2. Click Launch to finalize your submission
Notice that this will submit two analyses, one for each sample. When your jobs are submitted, you'll be redirected to the Job History page to monitor your submission. To see job status updates, refresh the page.
What to expect when your workflow completes
Hopefully your submission will succeed. If so, congratulations on setting up the workflow and running on your own data!!
Your Job History tab (once you refresh) will look like this:
And your sample data table will look like this (notice the additional columns for the generated BAM and BAM index files)
Notice the additional "sample_set" table! This is a group that includes the two samples in your "sample" table.
You can choose to run this analysis again or to run a downstream analysis on this group by choosing the set when you choose the data to run on (we'll do that in Part 3!).
|
|
---|---|
If you click on the "2 samples" you will see that this column includes the
|
![]() |
|
---|---|
You've completed Part 2 of the Workflows QuickStartNext up: Part 3 - Run downstream analysis on output data |
Comments
0 comments
Please sign in to leave a comment.