Archiving break records from Pushdown jobs

This section shows you how to set up the archive break records feature for Pushdown jobs.

To automatically write break records to a database or schema in your data source, you first need to configure your connection's output location and enable Archive Break Records from the Settings modal on the Explorer page.

Prerequisites

Specify the archive location for the connection

  1. Open Connection Management and edit your connection.
  2. In the Edit Connection dialog box, select Archive Break Records.
  3. Specify the schema to archive break records in the Archive Break Records Location field. Keep in mind the following tips for selecting the schema:

Example screenshot...archive break records connection setup example

Create the job

  1. From Explorer, connect to a Pushdown data source.
  2. Optionally, assign a Link ID to a column in the Select Columns step.
  3. Important If you specify a link ID, the column you assign as the link ID should not contain NULL values and its values should be unique, most commonly the primary key. Composite primary key is also supported.

  4. In the lower left corner, click Cogwheel icon Settings.
  5. The Settings dialog box appears.
  6. Select the Archive Break Records switch.
  7. The switches for the available archive options, such as Archive Rules Break Records, become selectable.

    Note By default, enabling Archive Break Records turns on any available archive option. The availability of these options depends on the data source.

  8. Optionally, disable any of the archive options for which you do not want to send break records to the source system.
  9. Optionally, enter the name of alternative destination schema in the Source Output Schema field, to create tables for break records storage instead of the schema provided in the connection.
  10. Tip This can be either the database.schema or the schema and requires write access to the source output schema location.

    Example screenshot...archive break records settings

  11. Click Save.
  12. Set up and run your DQ job.
  13. When a record breaks, its metadata exports automatically to the data source.

Limitations

Troubleshooting

If you receive an out-of-memory error during a job, such as java.lang.OutOfMemoryError: Java heap space, you may need to allocate more memory to the owl-web pod. Apply the following settings to the helm values.yaml file and restart the services.

Copy
global:
 web:
  resources:
   requests:
    memory: "500Mi"
    cpu: "100m"
   limits:
    memory: "4Gi"
    cpu: "2000m"