GCP BigQuery audit logs
In this mode, the package will monitor all the jobs that written to a GCP BigQuery Audit logs table instead of using INFORMATION_SCHEMA.JOBS one.
tip
To get the best out of this mode, you should enable the should_combine_audit_logs_and_information_schema setting to combine both sources.
More details on the related page.
At the time of writing, there are 2 versions of the audit logs table. More details in the repository from Google.
dbt-bigquery-monitoring supports only the v2 version
It's likely that if you use v2, you will have a table named cloudaudit_googleapis_com_data_access in your audit dataset.
To enable the "cloud audit logs mode", you'll need to define explicitly mandatory settings to set in the dbt_project.yml file:
vars:
enable_gcp_bigquery_audit_logs: true
gcp_bigquery_audit_logs_storage_project: 'my-gcp-project'
gcp_bigquery_audit_logs_dataset: 'my_dataset'
gcp_bigquery_audit_logs_table: 'cloudaudit_googleapis_com_data_access'
# should_combine_audit_logs_and_information_schema: true # Optional, default to false but you might want to combine both sources