Skip to main content

Metadata Exporter Application

The Metadata Exporter application enables organizations to export Data Quality (DQ) test results (pass/fail flags, rule IDs, asset metadata, timestamps) and profile data into downstream analytical or data storage systems such as Snowflake, BigQuery, and Databricks.

Overview

This capability supports:
  • Feeding downstream dashboards (Power BI, Tableau).
  • Triggering alerting and remediation workflows.
  • Historical tracking and versioning of data quality scores (DQI).

Key Features

  • Supports Snowflake, Databricks, and BigQuery destinations.
  • Manual or scheduled export runs.
  • Selective export of data quality events or profile data.

How to Access

  1. Go to Settings > Applications > Metadata Exporter.
Metadata Exporter Navigation
  1. Review available tabs: Schedule, Configuration, Recent Runs.
Metadata Exporter Tabs

Configuration Options

Ingestion Runner

Defines the agent responsible for executing the ingestion pipeline.
  • Example: Collate SaaS Agent

Connection Configuration

Establishes connectivity to your export destination (Snowflake, BigQuery, Databricks). Configuration

Snowflake Configuration

FieldDescription
Service TypeSnowflake
UsernameSnowflake user login
PasswordUser password (optional if using private key)
AccountSnowflake account identifier (e.g., AAAAA-99999)
RoleSnowflake role to assume (e.g., ACCOUNTADMIN)
DatabaseTarget database (e.g., OBS_ANALYTICS)
WarehouseTarget virtual warehouse (e.g., COMPUTE_WH)
Query TagOptional tagging for traceability
Private Key & PassphraseFor key-pair auth (optional, secure)
Advanced Option
  • Client Session Keep Alive: Useful for long-running exports.

BigQuery Configuration

FieldDescription
Service TypeMust be BigQuery
Project IDGCP project where the BigQuery dataset resides
Dataset IDTarget dataset where the metadata will be exported
Table NameDestination table name (BQ table to export metadata to)
Service Account JSONContents of the service account key in JSON format with write access
LocationBigQuery region (e.g., us-central1)
Ensure the service account has the BigQuery Data Editor and BigQuery Job User roles.

Databricks Configuration

FieldDescription
Service TypeMust be Databricks
Host URLDatabricks workspace URL (e.g., https://<region>.azuredatabricks.net)
TokenPersonal Access Token (PAT) for API authentication
Cluster IDTarget cluster where jobs will run
Database NameTarget database within the Databricks environment
Schema NameSchema (if applicable)
Table NameDestination table to store metadata
Path (Optional)DBFS path or external location (if exporting to files instead of a table)
Requirements
  • The token must have workspace-wide read/write access.
  • The cluster must have access to the target database or mount location.

Export Range

Defines the temporal scope of the data to be exported.
FieldDescription
Range Type (exportRange.rangeType)Options: ALL, LATEST, or DATE_RANGE
Interval (exportRange.interval)Used with DATE_RANGE (e.g., 7)
Unit (exportRange.unit)Time unit for the interval (e.g., days, hours)
Event TypesSelect which types of DQ events to export (All, or specific types)
BackfillEnable to process historical data on first run
Export Range

Table Configuration

Specifies the target table where exported metadata will be written.
FieldDescription
Database Name (tableConfiguration.databaseName)e.g., OBS_ANALYTICS
Schema Name (tableConfiguration.schemaName)e.g., OBS_DATA
Table Name (tableConfiguration.tableName)e.g., COLLATE_METADATA
Table Configuration

Scheduling

Configure how often the metadata export runs:
  • Manual: Click Run Now on the Schedule tab.
  • Scheduled: Set up periodic exports (feature roadmap).
Scheduling

Monitoring Runs

Under the Recent Runs tab:
  • View status: Success or Failed.
  • Check run time, duration, logs, and the configuration used.
A successful export shows Status: Success, with details on execution duration and timestamps. Monitoring Runs

Metadata Exporter Application Parameters

KeyDescription
exportRange.rangeTypeDefines range (ALL, LATEST, DATE_RANGE)
exportRange.intervalInterval number for DATE_RANGE
exportRange.unitTime unit (days, hours)
eventTypesEvent types to export
BackfillBoolean, historical data processing
tableConfiguration.databaseNameTarget DB
tableConfiguration.schemaNameTarget schema
tableConfiguration.tableNameTarget table