User Community Service Desk Downloads
If you can't find the product or version you're looking for, visit support.ataccama.com/downloads

Post-Processing Plans

Results of DQ evaluation can be sent to ONE Desktop and used there as an input for plans consisting of multiple post-processing steps that adapt your data to the desired form. Or, you can export these records into ONE Data, an external database, or a custom business intelligence tool.

When using the JDBC Writer step as an output in your post-processing plan, it is possible to write to the following locations:

  • The JDBC driver of the data source of your catalog item in your monitoring project. That is, if the catalog item is stored in the PostgreSQL database, you are only able to write into the PostgreSQL database in your post-processing plan.

  • Data sources defined in the Global Runtime Configuration (the JDBC data source should be specified even if it is already declared in Data Processing Engine, or DPE).

    When it comes to post-processing, the following requirements need to be fulfilled:

    • The <datasource> element must include the driverClass attribute instead of driverName.

    • The driver class must be in the DQC classpath.

    • Post-processing components referencing the data source must use the same name of the data source as specified in the runtime configuration.

If any of the catalog item attribute names (column names) present in a monitoring project is a duplicate of the system input names, then the post-processing plan for that catalog item fails.

The names that must be avoided are: score, valid_rules, invalid_rules, or invalid_rules_explanation. Renaming the column names in the source database can resolve this issue.

Prerequisites

  • ONE Desktop.

  • Connection to ONE (see Connect to Ataccama ONE Platform).

    Editing server details, such as the server name, after the catalog item has been created can later cause issues during profiling. In that case, the recommended approach is to specify additional parameters, for example, the new server name, through the global runtime configuration.

    This configuration is provided as a base64-encoded string to the Data Processing Module (DPM) configuration. For more information, see DPM Configuration, section Global runtime.

  • A preconfigured monitoring project with rules applied.

Create a post-processing plan

  1. In ONE Desktop, navigate to Monitoring Projects.

  2. Find the required project and right-click the project name.

  3. Select New > Post processing plan.

  4. In Create a new post processing plan, enter the following:

    400
    1. Name: For the purpose of our example, we are using My Post Processing Plan.

    2. Catalog Item: The catalog item in the monitoring project for which you want to create the post-processing plan.

    3. Output Location: The output location for CSV files produced from post-processing plans. For example, export/${date}/${catalogItemName}/.

  5. Click Finish.

  6. Navigate to Components.

  7. Right-click the newly-created component and select Edit component.

  8. Remove the Trash step and replace it with the Text File Writer step.

  9. Configure the Text File Writer step:

    1. File Name: Output.csv

    2. Encoding: UTF-8

    3. Compression: NONE

    4. Line Separator: \r\n

    5. Field Separator: ;

    6. String Qualifier: "

    7. String Qualifier Escape: "

    8. Write Header: enabled

      400
      It is not possible to use variables in the Text File Writer for File Name as variables are not supported in post-processing plans.
  10. Select OK

  11. Right-click the component and select Publish.

  12. In Publish Component, change the State to Implemented.

    400
  13. Select Publish.

  14. Return to ONE and run monitoring for the respective monitoring project.

  15. Navigate to the Export tab.

  16. Download the exported report (Output.csv).

All monitoring project rule instances are present in post-processing plans, either in the valid_rules or invalid_rules column.

The dimension result mapping to the passed or failed state is used for assigning the rule to valid_rules and invalid_rules columns respectively. For more information, see Data Quality Dimensions.

Whether or not a dimension is contributing to overall quality or not does not impact whether it is shown in invalid samples. invalid_rules_explanation contains only the failed results.

Was this page useful?