diff --git a/providers/src/airflow/providers/google/cloud/hooks/bigquery.py b/providers/src/airflow/providers/google/cloud/hooks/bigquery.py index 159a8f3f639b5..bce89e9e4184b 100644 --- a/providers/src/airflow/providers/google/cloud/hooks/bigquery.py +++ b/providers/src/airflow/providers/google/cloud/hooks/bigquery.py @@ -2095,7 +2095,7 @@ async def create_job_for_partition_get( query_request = { "query": "SELECT partition_id " f"FROM `{project_id}.{dataset_id}.INFORMATION_SCHEMA.PARTITIONS`" - + (f" WHERE table_id={table_id}" if table_id else ""), + + (f" WHERE table_name='{table_id}'" if table_id else ""), "useLegacySql": False, } job_query_resp = await job_client.query(query_request, cast(Session, session)) diff --git a/providers/tests/google/cloud/hooks/test_bigquery.py b/providers/tests/google/cloud/hooks/test_bigquery.py index b0e7f8efb2098..ee0f904bb94b6 100644 --- a/providers/tests/google/cloud/hooks/test_bigquery.py +++ b/providers/tests/google/cloud/hooks/test_bigquery.py @@ -1604,7 +1604,7 @@ async def test_create_job_for_partition_get_with_table(self, mock_job_instance, expected_query_request = { "query": "SELECT partition_id " f"FROM `{PROJECT_ID}.{DATASET_ID}.INFORMATION_SCHEMA.PARTITIONS`" - f" WHERE table_id={TABLE_ID}", + f" WHERE table_name='{TABLE_ID}'", "useLegacySql": False, } await hook.create_job_for_partition_get(