Advanced Taskflow for Data Integration
Advanced Taskflow for Data Integration
Integration tasks
Prakash Jain
Informatica Global Customer Support
Housekeeping Tips
➢ The session will include a webcast and then your questions will be answered live at the end of the presentation
➢ All dial-in participants will be muted to enable the speakers to present without interruption
➢ Questions can be submitted to “All Panelists" via the Q&A option and we will respond at the end of the presentation
➢ The webinar is being recorded and will be available to view on our INFASupport YouTube channel and Success Portal.
The link will be emailed as well.
➢ Please take time to complete the post-webinar survey and provide your feedback and suggestions for upcoming topics.
Bootstrap trial and Enriched Customer Product Learning Informatica Tailored training and
POC Customers Onboarding Paths and Weekly Concierge with content
experience Expert Sessions Chatbot integrations recommendations
Example 1: You want to run two mapping tasks in sequence. However, you do not want to run
the second mapping task as soon as the first task ends if you get a warning on the
first mapping task. Instead, you want to run the second mapping task to run after two hours. You
can create a task flow to orchestrate this scenario.
Example 2: You may want to re-try your Data Integration task few times if they fail due to some
network glitch or other issue. In this case, you can use taskflow for performing this retry
attempts before notifying the end user if they still fail.
You can use the Command Task outputs like status, Exit code etc to orchestrate subsequent tasks in the taskflow.
For example, you can wait for a file to arrive at a particular location and then consume the file in a subsequent step.
Ingestion Task
Use an Ingestion Task step to leverage a file ingestion task for taskflow orchestration. Select an existing file ingestion
task.
You might want to perform data integration operations after moving files to an intermediate location and before
transferring the files to the target. In this case, you can use the Ingestion Task step in conjunction with the Data Task
step.
10 © Informatica. Proprietary and Confidential.
Taskflow Steps
Subtaskflow
Use a Subtaskflow step to embed and reuse an existing taskflow. You can configure input fields to provide
input when you run the taskflow. You can also enable fault handling to determine the reason for a taskflow
failure.
Decision
Use a Decision step when you want a taskflow to take different paths based on the value of a specific field.
Parallel Paths
Use a Parallel Paths step when you want a taskflow to run multiple items at the same time. For example,
you can run three mapping tasks simultaneously. The taskflow runs all items in the Parallel Paths step and
then moves to the next step.
Jump
Use a Jump step when you want to jump from one part of the taskflow to another.
End
Use an End step to define the HTTP status code that must be used when a taskflow completes.
11 © Informatica. Proprietary and Confidential.
Taskflow Steps
Wait
Use a Wait step when you want to pause taskflow execution for a specific duration.
Throw
Use a Throw step to catch a fault, return the fault details, and prevent the execution of the
subsequent steps in a taskflow. The Throw step is an interrupting step, which means that if a
fault occurs, the Throw step stops the execution of the taskflow and sets the taskflow status to
failed.
Use this template when you need a basic canvas with a Start step and an End step. You can add other steps at any point on the canvas on using any of the task templates.
Parallel Tasks
Use Parallel Tasks template if your major requirement is to run two or more data integration tasks in parallel. You start with a taskflow that contains a Start step, a Parallel Paths step, and
an End step.
Use Parallel Tasks with Decision template if your major requirement is to run two or more data integration tasks in parallel and then make a decision based on the outcome of any task. You
start with a taskflow that contains a Start step, a Parallel Paths step, a Decision step, and an End step.
Sequential Tasks
Use Sequential Tasks template if your major requirement is to run two data integration tasks, one after the other. You start with a taskflow that contains a Start step, two Data Task steps,
and an End step.
Use Sequential Tasks with Decision template if your major requirement is to run two Data Integration tasks sequentially and then make a decision based on the output of either task. You
start with a taskflow that contains a Start step, two Data Task steps, a Decision step, and an End step.
Single Task
Use Single Task template if your major requirement is to run one data integration task on a daily or weekly schedule, for example. You start with a taskflow that contains a Start step, a Data
Task step, and an End step.
REST/SOAP
• If you select the REST/SOAP binding type, you can run the taskflow by using a REST or SOAP endpoint. You can use the Allowed Groups
and the Allowed Users fields to define the user groups and users who can run a published taskflow as an API.
Event
• If you select the Event binding type, the taskflow is invoked when the specified event occurs. For example, a taskflow can be invoked
upon an event such as arrival of a file in a file system. The Event Source Name field is available where you can select the file listener
that you created for the event.
Taskflow access:
If the taskflow uses the REST/SOAP binding type, you can define the user groups and users who can run a published taskflow as an API in the Allowed
Groups and the Allowed Users fields.
• Use the Input Fields to add fields that a taskflow uses at the beginning of a step.
• When you select the binding type for a taskflow as Event and select a file listener, Data Integration creates an
input field to store details of the files that arrived, were updated, or were deleted as part of the file listener event.
Temporary fields
• Create temporary fields for use in a taskflow step. A taskflow uses temporary fields internally. Temporary fields
do not appear in the input or output of a taskflow.
Advanced properties
You can configure a taskflow to suspend on a fault that occurs at the taskflow level. You can also configure the
taskflow to send an email notification to specified recipients when it is suspended on a fault.
You can use a taskflow to pass input and in-out parameters to a task. The mapping must be created with
input parameters or in-out parameters.
Input parameters
• Input parameter is a placeholder for a value or values in a mapping. You define the value of the
parameter when you configure the mapping task.
• You can use a taskflow to override the following subset of mapping input parameters:
--Source object. Changes the object that the mapping task reads from.
-Source connection. Changes the connection that the mapping task uses to read from the source.
-Target connection. Changes the connection that the mapping task uses to write to the target.
-Target object. Changes the object that the mapping task writes to.
• To add a field, click the Fields tab, drill down to the field that you want to
• To add an operator, click an operator in the Operators section. You can also
• •To add a function, click the Functions tab, drill down to the function you want to
• •To add a comment, enter the comment in the Expression section with the
following syntax:(:<comment>:).
Example:
To run a taskflow from the taskflow designer, open the taskflow and click Run in the upper-right part of the
page.
API
To run a taskflow as an API, you must first publish the taskflow as a service, and then run it. When you
publish a taskflow, Data Integration generates the service URL and the SOAP service URL. You can use these
endpoint URLs to run the taskflow as an API.
You can invoke a taskflow through a connector file listener. Within the taskflow, define the binding type as
Event and select the connector file listener as the event source. When you publish the taskflow, the taskflow
subscribes to the connector file listener that is defined in it.
Schedule
• To run a taskflow on a schedule, create a schedule in Administrator and associate the taskflow with the
schedule.
23 © Informatica. Proprietary and Confidential.
Running a Taskflow as API
• To run a taskflow as an API, you must first publish the taskflow as a service.
• When you publish a taskflow, Data Integration generates the service URL and the SOAP service URL.
• You can use these endpoint URLs to run the taskflow as an API.
• You can dynamically provide input parameters for the tasks that the taskflow contains and perform
orchestration.
• You receive the taskflow run ID as the response.
• The service URL uses the following format:
<Informatica Intelligent Cloud Services URL>/active-bpel/rt/<API_name>
The SOAP service URL uses the following format:
<Informatica Intelligent Cloud Services URL>/active-bpel/soap/<API_name>
• Use the run ID to monitor the taskflow run in the My Jobs page.