User Community Service Desk Downloads

Pipeline Monitoring

Pipeline monitoring gives you visibility into the execution of your data pipelines directly within Ataccama ONE - Agentic Data Trust Platform. By connecting your orchestration tools, you can track job runs, detect failures, and investigate alerts without leaving the platform.

What pipeline monitoring provides

When you connect an orchestrator, Ataccama ONE receives execution metadata for every pipeline run. This enables you to:

  • Track job executions — See which jobs are running, completed, or failed across your data ecosystem.

  • Detect failures early — Identify failed or aborted jobs as soon as they occur.

  • Investigate alerts and findings in context — View job history, run durations, and failure patterns alongside your data assets.

  • Trigger alerts — Automatically create alerts when jobs fail, and notify the right people through your preferred channels.

How it works

Pipeline monitoring uses the OpenLineage standard to collect execution metadata from your orchestration tools.

When you configure an orchestrator connection:

  1. Your orchestration tool (Airflow, dbt, Dagster) sends OpenLineage events to ONE.

  2. Events are processed and associated with the connection you created.

  3. Job executions appear in the Pipeline monitoring module, where you can monitor status and investigate issues.

  4. If configured, alerts are triggered for job failures and sent to the appropriate audiences.

Supported orchestrators

Pipeline monitoring currently supports:

  • Apache Airflow with OpenLineage

  • dbt Core with OpenLineage

  • Dagster with OpenLineage

Prerequisites

Before you can monitor pipelines, you need:

Administrative access to your orchestrator

You must be able to configure OpenLineage settings in your Airflow or dbt environment, including setting endpoint URLs and API keys.

A supported orchestration tool

See Supported orchestrators for the list of currently supported tools.

Appropriate permissions

You need sufficient permissions in ONE to create orchestrator connections and view pipeline monitoring data. For details, see alerts:manage-observability-permissions.adoc.

Get started

Setting up pipeline monitoring involves three steps: creating a connection in ONE, configuring your orchestrator to send events, and optionally setting up alerts for job failures.

Step 1: Create an orchestrator connection

Create a connection in ONE to receive events from your orchestration tool.

  1. Navigate to Data Observability > Pipeline monitoring.

  2. Select Add connection.

  3. Provide a display name and select your connection type (Airflow with OpenLineage, dbt Core with OpenLineage, or Dagster with OpenLineage).

  4. Generate an API key and copy the OpenLineage endpoint URL.

For detailed instructions, see Connect to an Orchestrator.

Step 2: Configure your orchestrator

Use the API key and endpoint URL from the previous step to configure your orchestration tool.

Once configured, your orchestrator begins sending execution events to ONE.

To be notified when jobs fail, configure alert notifications that match pipeline monitoring findings.

  1. Navigate to Alerts from the quick access menu.

  2. Select the Notifications tab.

  3. Create a notification policy with:

    • Finding type: Select Job failure, Job aborted, or both under Pipeline monitoring.

    • Severity: Choose which severity levels should trigger notifications.

    • Audiences: Select who should receive notifications.

For detailed instructions, see Get Notifications for Alerts.

Start by creating a single notification policy for all critical pipeline failures. You can refine your notification strategy as you learn which alerts are most valuable to your team.

Monitor your pipelines

Once your orchestrator is connected and sending events, you can monitor pipeline executions in two places.

Pipeline monitoring

Navigate to Data Observability > Pipeline monitoring to see all your orchestrator connections and their job executions.

Pipeline monitoring page showing orchestrator connection cards with job counts and last event times

Each connection card displays the number of pipeline jobs tracked, the connection type, when it was created, and how recently the last event was received. Select View details to open the connection detail page.

Connection detail page showing pipeline jobs table with run status

The detail page shows connection information and a table of all pipeline jobs tracked by this connection. For each job, you can see the last run start and end times, duration, and status (Completed, Failed, or Running). Select a job to view its full run history and related alerts.

Pipeline job detail page showing run duration chart over time

Alerts feed

Navigate to Alerts to see pipeline-related alerts alongside other findings. Filter by finding type (Job failure, Job aborted) to focus on pipeline issues.

Each alert provides:

  • Details about the failed job and when it occurred

  • A pipeline view showing affected assets and downstream impact

  • Resolution tracking to document how issues were addressed

For more information about working with alerts, see Investigate Alerts.

Next steps

Was this page useful?