- Release notes
- Before you begin
- Getting started
- Integrations
- Managing access
- Working with process apps
- Creating apps
- Loading data
- Uploading data
- Retrieving the SQL Server database parameters
- Setting up a SQL Server account for data upload using an extractor
- Loading data using Theobald Xtract Universal
- Customizing process apps
- Data transformations
- TemplateOne app template
- Purchase to Pay app template
- Order to Cash app template
- Basic troubleshooting guide
Process Mining
Create a job
The input data must meet the format as required for the app template you are using to create your process app. Refer to App Templates for details.
Follow these steps to create the extraction job.
- Create a new CData Sync job.
- Enter a descriptive name for the job in the Name field. For example, ServiceNow_to_AzureBlob or ServiceNow_to_SQLServer.
- Select the source connection created in Create a source connection the source connection from the Source list.
- Select the destination connection created in Create a destination connection from the Destination list.
- Make sure the option Standard is selected as the Replication Type and select Add Job.
-
Go to the Task tab and select Add Tasks. Enable the Custom Query option.
- Create a Custom Query using the replication queries for your app template and select Add Tasks.
Tip:
You can copy the replication queries from the documentation for you app template that can be accessed from the list of App templates. Refer to App Templates.
- Edit the job Settings on the Overview tab. Enable the checkbox
Parallel Processing
and enter8
in the Worker Pool field to improve loading speed.- Locate the Destination schema entry and copy the schema retrieved in Create a Destination.
- Configure the Incremental Replication to make sure the Replicate Interval and Replicate Interval Unit are set so the resulting period is equal or larger than the extraction period. For example, if you want to extract data for a period of a year, the Replicate Interval and Replicate Interval Unit must reflect at least a year period.
- Go to the Advanced tab in the Settings panel and edit the Replicate Options. Enable Drop Table to prevent the data to be appended to the table.
- Save your changes.
Follow these steps to edit the post-job event to call the data ingestion API.
- Go to the Events tab in the Job Settings panel.
- Edit the Post-Job Event section and add the ingestion API call from the code block below ensuring to replace the
value
of the "http.url" with the End of upload API. Refer toRetrieving the SQL Server database parameters.<api:set attr="http.url" value="https://my-uipath-server.com/default/defaulttenant/processMining_/api/v4.0/apps/98dfd1d5-9e42-4f0e-9a0a-8945629f01b3/transform/unauthenticated"/> <api:call op="httpPost" in="http"/>
<api:set attr="http.url" value="https://my-uipath-server.com/default/defaulttenant/processMining_/api/v4.0/apps/98dfd1d5-9e42-4f0e-9a0a-8945629f01b3/transform/unauthenticated"/> <api:call op="httpPost" in="http"/> - Save your changes.
- Select Jobs and locate the destination job.
- Run the job to test if the job runs correctly.
You can add additional statements to obtain logging information on the API call. Follow these steps.
- Add the following statements between the lines previously added in the Post-Job event.
<api:set attr="http.verbosity" value="5"/>
<api:set attr="http.logfile" value="D:/mydir/cdata_log.txt"/>
-
Edit the job Settings on the Overview tab and select Verbose from the Logfile Verbosity list box.
- Run the job and check the log file that is created.