๐ Metadata Exporter Documentation
Overview
The Metadata Exporter is a configurable application that enables organizations to export Data Quality (DQ) test results โ such as pass/fail flags, rule IDs, asset metadata, timestamps โ and profile data into downstream analytical or data storage systems like Snowflake, BigQuery (BQ), and Databricks.
This functionality enables:
- Feeding downstream dashboards (e.g., Power BI, Tableau)
 - Triggering alerting and remediation workflows
 - Historical tracking and versioning of data quality scores (DQI)
 
Key Features
Destination Support:
- Snowflake
 - Databricks
 - BigQuery
 
Export Cadence Options:
- Manual
 - Scheduled
 
Selective Export:
- Ability to decide what event to export (data quality or profile data)
 
Navigation
To configure the Metadata Exporter:
- Go to: 
Settings > Applications > Metadata Exporter 

Metadata Exporter Navigation
Youโll find the following tabs:
ScheduleConfigurationRecent Runs

Metadata Exporter Tabs
Configuration Options
1. Ingestion Runner
Defines the agent responsible for executing the ingestion pipeline.
Example: Collate SaaS Agent
2. Connection Configuration
Establishes connectivity to your export destination (e.g., Snowflake, BigQuery, Databricks).

Configuration
Snowflake Configuration
| Field | Description | 
|---|---|
| Service Type | Snowflake | 
| Username | Snowflake user login | 
| Password | User password (optional if using private key) | 
| Account | Snowflake account identifier (e.g., AAAAA-99999) | 
| Role | Snowflake role to assume (e.g., ACCOUNTADMIN) | 
| Database | Target database (e.g., OBS_ANALYTICS) | 
| Warehouse | Target virtual warehouse (e.g., COMPUTE_WH) | 
| Query Tag | Optional tagging for traceability | 
| Private Key & Passphrase | For key-pair auth (optional, secure) | 
Advanced Option:
- Client Session Keep Alive โ Useful for long-running exports
 
BigQuery (BQ) Configuration
| Field | Description | 
|---|---|
| Service Type | Must be BigQuery | 
| Project ID | GCP project where the BigQuery dataset resides | 
| Dataset ID | Target dataset where the metadata will be exported | 
| Table Name | Destination table name (BQ table to export metadata to) | 
| Service Account JSON | Contents of the service account key in JSON format with write access | 
| Location | BigQuery region (e.g., us-central1) | 
Security Note: Ensure the service account has the BigQuery Data Editor and BigQuery Job User roles.
Databricks (DBX) Configuration
| Field | Description | 
|---|---|
| Service Type | Must be Databricks | 
| Host URL | Databricks workspace URL (e.g., https://<region>.azuredatabricks.net) | 
| Token | Personal Access Token (PAT) for API authentication | 
| Cluster ID | Target cluster where jobs will run | 
| Database Name | Target database within the Databricks environment | 
| Schema Name | Schema (if applicable) | 
| Table Name | Destination table to store metadata | 
| Path (Optional) | DBFS path or external location (if exporting to files instead of a table) | 
Requirements:
- The token must have workspace-wide read/write access.
 - The cluster must have access to the target database or mount location.
 
3. Export Range
Defines the temporal scope of the data to be exported.
| Field | Description | 
|---|---|
Range Type (exportRange.rangeType) | Options: ALL, LATEST, or DATE_RANGE | 
Interval (exportRange.interval) | Used with DATE_RANGE (e.g., 7) | 
Unit (exportRange.unit) | Time unit for the interval (e.g., days, hours) | 
| Event Types | Select which types of DQ events to export (All, or specific types) | 
| Backfill | Enable to process historical data on first run | 

Export Range
4. Table Configuration
Specifies the target table where exported metadata will be written.
| Field | Description | 
|---|---|
Database Name (tableConfiguration.databaseName) | e.g., OBS_ANALYTICS | 
Schema Name (tableConfiguration.schemaName) | e.g., OBS_DATA | 
Table Name (tableConfiguration.tableName) | e.g., COLLATE_METADATA | 

Table Configuration
Scheduling
Configure how often the metadata export runs:
- Manual: Click 
Run Nowon the Schedule tab - Scheduled: Setup periodic exports (feature roadmap)
 

Scheduling
Monitoring Runs
Under the Recent Runs tab:
- View status: Success or Failed
 - Check:
- Run time
 - Duration
 - Logs for troubleshooting
 - Config used during run
 
 
A successful export shows the Status: Success, with details on execution duration and timestamps.

Monitoring Runs
MetadataExporterApplication Parameters (Developer Reference)
| Key | Description | 
|---|---|
exportRange.rangeType | Defines range (ALL, LATEST, DATE_RANGE) | 
exportRange.interval | Interval number for DATE_RANGE | 
exportRange.unit | Time unit (days, hours) | 
eventTypes | Event types to export | 
Backfill | Boolean, historical data processing | 
tableConfiguration.databaseName | Target DB | 
tableConfiguration.schemaName | Target schema | 
tableConfiguration.tableName | Target table |