๐ Metadata Exporter Documentation
Overview
The Metadata Exporter is a configurable application that enables organizations to export Data Quality (DQ) test results โ such as pass/fail flags, rule IDs, asset metadata, timestamps โ and profile data into downstream analytical or data storage systems like Snowflake, BigQuery (BQ), and Databricks.
This functionality enables:
- Feeding downstream dashboards (e.g., Power BI, Tableau)
- Triggering alerting and remediation workflows
- Historical tracking and versioning of data quality scores (DQI)
Key Features
Destination Support:
- Snowflake
- Databricks
- BigQuery
Export Cadence Options:
- Manual
- Scheduled
Selective Export:
- Ability to decide what event to export (data quality or profile data)
Navigation
To configure the Metadata Exporter:
- Go to:
Settings > Applications > Metadata Exporter

Metadata Exporter Navigation
Youโll find the following tabs:
Schedule
Configuration
Recent Runs

Metadata Exporter Tabs
Configuration Options
1. Ingestion Runner
Defines the agent responsible for executing the ingestion pipeline.
Example: Collate SaaS Agent
2. Connection Configuration
Establishes connectivity to your export destination (e.g., Snowflake, BigQuery, Databricks).

Configuration
Snowflake Configuration
Field | Description |
---|---|
Service Type | Snowflake |
Username | Snowflake user login |
Password | User password (optional if using private key) |
Account | Snowflake account identifier (e.g., AAAAA-99999 ) |
Role | Snowflake role to assume (e.g., ACCOUNTADMIN ) |
Database | Target database (e.g., OBS_ANALYTICS ) |
Warehouse | Target virtual warehouse (e.g., COMPUTE_WH ) |
Query Tag | Optional tagging for traceability |
Private Key & Passphrase | For key-pair auth (optional, secure) |
Advanced Option:
- Client Session Keep Alive โ Useful for long-running exports
BigQuery (BQ) Configuration
Field | Description |
---|---|
Service Type | Must be BigQuery |
Project ID | GCP project where the BigQuery dataset resides |
Dataset ID | Target dataset where the metadata will be exported |
Table Name | Destination table name (BQ table to export metadata to) |
Service Account JSON | Contents of the service account key in JSON format with write access |
Location | BigQuery region (e.g., us-central1 ) |
Security Note: Ensure the service account has the BigQuery Data Editor and BigQuery Job User roles.
Databricks (DBX) Configuration
Field | Description |
---|---|
Service Type | Must be Databricks |
Host URL | Databricks workspace URL (e.g., https://<region>.azuredatabricks.net ) |
Token | Personal Access Token (PAT) for API authentication |
Cluster ID | Target cluster where jobs will run |
Database Name | Target database within the Databricks environment |
Schema Name | Schema (if applicable) |
Table Name | Destination table to store metadata |
Path (Optional) | DBFS path or external location (if exporting to files instead of a table) |
Requirements:
- The token must have workspace-wide read/write access.
- The cluster must have access to the target database or mount location.
3. Export Range
Defines the temporal scope of the data to be exported.
Field | Description |
---|---|
Range Type (exportRange.rangeType ) | Options: ALL , LATEST , or DATE_RANGE |
Interval (exportRange.interval ) | Used with DATE_RANGE (e.g., 7 ) |
Unit (exportRange.unit ) | Time unit for the interval (e.g., days , hours ) |
Event Types | Select which types of DQ events to export (All , or specific types) |
Backfill | Enable to process historical data on first run |

Export Range
4. Table Configuration
Specifies the target table where exported metadata will be written.
Field | Description |
---|---|
Database Name (tableConfiguration.databaseName ) | e.g., OBS_ANALYTICS |
Schema Name (tableConfiguration.schemaName ) | e.g., OBS_DATA |
Table Name (tableConfiguration.tableName ) | e.g., COLLATE_METADATA |

Table Configuration
Scheduling
Configure how often the metadata export runs:
- Manual: Click
Run Now
on the Schedule tab - Scheduled: Setup periodic exports (feature roadmap)

Scheduling
Monitoring Runs
Under the Recent Runs tab:
- View status: Success or Failed
- Check:
- Run time
- Duration
- Logs for troubleshooting
- Config used during run
A successful export shows the Status: Success, with details on execution duration and timestamps.

Monitoring Runs
MetadataExporterApplication Parameters (Developer Reference)
Key | Description |
---|---|
exportRange.rangeType | Defines range (ALL , LATEST , DATE_RANGE ) |
exportRange.interval | Interval number for DATE_RANGE |
exportRange.unit | Time unit (days , hours ) |
eventTypes | Event types to export |
Backfill | Boolean, historical data processing |
tableConfiguration.databaseName | Target DB |
tableConfiguration.schemaName | Target schema |
tableConfiguration.tableName | Target table |