Can you monitor the data quality in different clouds safely?
Multi Cloud Data Observability
Multi Cloud Data Observability
Observe data quality issues in multiple clouds from one place
DQO was designed as a distributed data monitoring platform where the data monitoring agents may run independently. Deploy the data monitoring agent in the monitored cloud (Azure, AWS, GCP) and let it talk to the DQO master node. The agent will efficiently execute data quality rules close to the data source.
- Analyze data quality across multiple clouds
- Run data monitoring agents close to the source to avoid latency
- Analyze data quality of on-premise data sources by deploying the agent in the corporate network
No data movement
No data movement
Avoid moving data across the Internet and between clouds to run data quality checks. Run the quality checks close to the data source.
DQO data quality checks run a query on the data source and retrieve metrics such as aggregated row count grouped by a business column. Limited metadata that is extracted avoids data leakage of sensitive data or high-volume data movements.
- Limit the data movements across clouds
- Avoid the risk of data leakage
- Run the data quality checks where the data is located
Shared Data Quality Rules
Shared Data Quality Rules
Run the same set of approved data quality checks on different databases and data lakes available on different clouds.
SQL engines available in different clouds use slightly different SQL dialects. DQO supports customized SQL templates for data quality queries that are adapted for different query engines. The list of quality checks simply references a built-in check that has multiple vendor-specific implementations.
- Analyze the quality of data using the same set of rules for multiple clouds
- Safely migrate between clouds by defining data quality checks to be executed on both the old and new cloud environments
- Define additional business-specific data quality rules with vendor-specific SQL templates