Create Amazon S3 BigLake tables
This document describes how to create an Amazon Simple Storage Service (Amazon S3) BigLake table. A BigLake table lets you to use access delegation to query data in Amazon S3. Access delegation decouples access to the BigLake table from access to the underlying data store.
For information about how data flows between BigQuery and Amazon S3, see Data flow when querying data.
Before you begin
Ensure that you have a connection to access Amazon S3 data.
Required roles
To get the permissions that you need to create an external table,
ask your administrator to grant you the
BigQuery Admin (roles/bigquery.admin
) IAM role on your dataset.
For more information about granting roles, see
Manage access.
This predefined role contains the permissions required to create an external table. To see the exact permissions that are required, expand the Required permissions section:
Required permissions
The following permissions are required to create an external table:
-
bigquery.tables.create
-
bigquery.connections.delegate
You might also be able to get these permissions with custom roles or other predefined roles.
Create a dataset
Before you create an external table, you need to create a dataset in the supported region. Select one of the following options:Console
Go to the BigQuery page.
- In the Explorer pane, select the project where you want to create the dataset.
- Expand the View actions option and click Create dataset.
- On the Create dataset page, specify the following details:
- For Dataset ID enter a unique dataset name.
- For Data location choose a supported region.
- Optional: To delete tables automatically, select the Enable table expiration checkbox and set the Default maximum table age in days. Data in Amazon S3 is not deleted when the table expires.
- Optional: Expand the Advanced options section and select the
following options:
- If you want to use a customer-managed encryption key, then select the Customer-managed encryption key (CMEK) option. By default, BigQuery encrypts customer content stored at rest by using a Google-managed key.
- If you want to use default collation, then select the Enable default collation option.
- Click Create dataset.
SQL
Use the CREATE SCHEMA
DDL statement.
The following example create a dataset in the aws-us-east-1
region:
In the Google Cloud console, go to the BigQuery page.
In the query editor, enter the following statement:
CREATE SCHEMA mydataset OPTIONS ( location = 'aws-us-east-1');
Click
Run.
For more information about how to run queries, see Running interactive queries.
bq
In a command-line environment, create a dataset using the bq mk
command:
bq --location=LOCATION mk \ --dataset \ PROJECT_ID:DATASET_NAME
The --project_id
parameter overrides the default project.
Replace the following:
LOCATION
: the location of your datasetFor information about supported regions, see Locations. After you create a dataset, you can't change its location. You can set a default value for the location by using the
.bigqueryrc
file.PROJECT_ID
: your project IDDATASET_NAME
: the name of the dataset that you want to createTo create a dataset in a project other than your default project, add the project ID to the dataset name in the following format:
PROJECT_ID:DATASET_NAME
.
Java
Before trying this sample, follow the Java setup instructions in the
BigQuery quickstart using
client libraries.
For more information, see the
BigQuery Java API
reference documentation.
To authenticate to BigQuery, set up Application Default Credentials.
For more information, see
Set up authentication for a local development environment.
Create BigLake tables on unpartitioned data
Select one of the following options:
Console
Go to the BigQuery page.
In the Explorer pane, expand your project, and then select a dataset.
In the Dataset info section, click
Create table.On the Create table page, in the Source section, do the following:
- For Create table from, select Amazon S3.
- For Select S3 path, enter a URI pointing to the Amazon S3
data in the format
s3://BUCKET_NAME/PATH
. ReplaceBUCKET_NAME
with the name of the Amazon S3 bucket; the bucket's region should be the same as the dataset's region. ReplacePATH
with the path that you would like to write the exported file to; it can contain one wildcard*
. - For File format, select the data format in Amazon S3. Supported formats are AVRO, PARQUET, ORC, CSV, and JSONL (Newline delimited JSON).
In the Destination section, specify the following details:
- For Dataset, choose the appropriate dataset.
- In the Table field, enter the name of the table.
- Verify that Table type is set to External table.
- For Connection ID, choose the appropriate connection ID from the drop-down. For information about connections, see Connect to Amazon S3.
In the Schema section, you can either enable schema auto-detection or manually specify a schema if you have a source file. If you don't have a source file, you must manually specify a schema.
To enable schema auto-detection, select the Auto-detect option.
To manually specify a schema, leave the Auto-detect option unchecked. Enable Edit as text and enter the table schema as a JSON array.
Click Create table.
SQL
To create a BigLake table, use the
CREATE EXTERNAL TABLE
statement with the
WITH CONNECTION
clause:
In the Google Cloud console, go to the BigQuery page.
In the query editor, enter the following statement:
CREATE EXTERNAL TABLE DATASET_NAME.TABLE_NAME WITH CONNECTION `AWS_LOCATION.CONNECTION_NAME` OPTIONS ( format = "DATA_FORMAT", uris = ["S3_URI"]);
Replace the following:
DATASET_NAME
: the name of the dataset you createdTABLE_NAME
: the name you want to give to this tableAWS_LOCATION
: an AWS location in Google Cloud (for example, `aws-us-east-1`)CONNECTION_NAME
: the name of the connection you createdDATA_FORMAT
: any of the supported BigQuery federated formats (such asAVRO
orCSV
)S3_URI
: a URI pointing to the Amazon S3 data (for example,s3://bucket/path
)
Click
Run.
For more information about how to run queries, see Running interactive queries.
Example:
CREATE EXTERNAL TABLE awsdataset.awstable WITH CONNECTION `aws-us-east-1.s3-read-connection` OPTIONS (format="CSV", uris=["s3://s3-bucket/path/file.csv"]);
bq
Create a table definition file:
bq mkdef \ --source_format=DATA_FORMAT \ --connection_id=AWS_LOCATION.CONNECTION_NAME \ S3_URI > table_def
Replace the following:
DATA_FORMAT
: any of the supported BigQuery federated formats (such asAVRO
orCSV
).S3_URI
: a URI pointing to the Amazon S3 data (for example,s3://bucket/path
).AWS_LOCATION
: an AWS location in Google Cloud (for example,aws-us-east-1
).CONNECTION_NAME
: the name of the connection you created.
Next, create the BigLake table:
bq mk --external_table_definition=table_def DATASET_NAME.TABLE_NAME
Replace the following:
DATASET_NAME
: the name of the dataset you created.TABLE_NAME
: the name you want to give to this table.
For example, the following command creates a new BigLake table,
awsdataset.awstable
, which can query your Amazon S3 data that's stored
at the path s3://s3-bucket/path/file.csv
and has a read connection in the
location aws-us-east-1
:
bq mkdef \ --autodetect \ --source_format=CSV \ --connection_id=aws-us-east-1.s3-read-connection \ s3://s3-bucket/path/file.csv > table_def bq mk --external_table_definition=table_def awsdataset.awstable
API
Call the tables.insert
method
API method, and create an
ExternalDataConfiguration
in the Table
resource
that you pass in.
Specify the schema
property or set the
autodetect
property to true
to enable schema auto detection for
supported data sources.
Specify the connectionId
property to identify the connection to use
for connecting to Amazon S3.
Java
Before trying this sample, follow the Java setup instructions in the BigQuery quickstart using client libraries. For more information, see the BigQuery Java API reference documentation.
To authenticate to BigQuery, set up Application Default Credentials. For more information, see Set up authentication for a local development environment.
Create BigLake tables on partitioned data
You can create a BigLake table for Hive partitioned data in Amazon S3. After you create an externally partitioned table, you can't change the partition key. You need to recreate the table to change the partition key.
To create a BigLake table based on Hive partitioned data, select one of the following options:
Console
Go to the BigQuery page.
In the Explorer pane, expand your project and select a dataset.
Click
View actions, and then click Create table. This opens the Create table pane.In the Source section, specify the following details:
For Create table from, select one of the following options:
- Amazon S3
- Azure Blob Storage
Provide the path to the folder, using wildcards. For example:
- For Amazon S3:
s3://mybucket/*
- For Blob Storage:
azure://mystorageaccount.blob.core.windows.net/mycontainer/*
The folder must be in the same location as the dataset that contains the table you want to create, append, or overwrite.
- For Amazon S3:
From the File format list, select the file type.
Select the Source data partitioning checkbox, and then specify the following details:
- For Select Source URI Prefix, enter the
URI prefix. For example,
s3://mybucket/my_files
. - Optional: To require a partition filter on all queries for this table, select the Require partition filter checkbox. Requiring a partition filter can reduce cost and improve performance. For more information, see Requiring predicate filters on partition keys in queries.
In the Partition inference mode section, select one of the following options:
- Automatically infer types: set the partition schema
detection mode to
AUTO
. - All columns are strings: set the partition schema
detection mode to
STRINGS
. - Provide my own: set the partition schema detection mode to
CUSTOM
and manually enter the schema information for the partition keys. For more information, see Provide a custom partition key schema.
- Automatically infer types: set the partition schema
detection mode to
- For Select Source URI Prefix, enter the
URI prefix. For example,
In the Destination section, specify the following details:
- For Project, select the project in which you want to create the table.
- For Dataset, select the dataset in which you want to create the table.
- For Table, enter the name of the table that you want to create.
- For Table type, verify that External table is selected.
- For Connection ID, select the connection that you created earlier.
In the Schema section, you can either enable schema auto-detection or manually specify a schema if you have a source file. If you don't have a source file, you must manually specify a schema.
To enable schema auto-detection, select the Auto-detect option.
To manually specify a schema, leave the Auto-detect option unchecked. Enable Edit as text and enter the table schema as a JSON array.
To ignore rows with extra column values that don't match the schema, expand the Advanced options section and select Unknown values.
Click Create table.
SQL
Use the
CREATE EXTERNAL TABLE
DDL statement:
In the Google Cloud console, go to the BigQuery page.
In the query editor, enter the following statement:
CREATE EXTERNAL TABLE `PROJECT_ID.DATASET.EXTERNAL_TABLE_NAME` WITH PARTITION COLUMNS ( PARTITION_COLUMN PARTITION_COLUMN_TYPE, ) WITH CONNECTION `PROJECT_ID.REGION.CONNECTION_ID` OPTIONS ( hive_partition_uri_prefix = "HIVE_PARTITION_URI_PREFIX", uris=['FILE_PATH'], format ="TABLE_FORMAT" );
Replace the following:
PROJECT_ID
: the name of your project in which you want to create the table—for example,myproject
DATASET
: the name of the BigQuery dataset that you want to create the table in—for example,mydataset
EXTERNAL_TABLE_NAME
: the name of the table that you want to create—for example,mytable
PARTITION_COLUMN
: the name of the partitioning columnPARTITION_COLUMN_TYPE
: the type of the partitioning columnREGION
: the region that contains the connection—for example,us
CONNECTION_ID
: the name of the connection—for example,myconnection
HIVE_PARTITION_URI_PREFIX
: hive partitioning uri prefix–for example:s3://mybucket/
azure://mystorageaccount.blob.core.windows.net/mycontainer/
FILE_PATH
: path to the data source for the external table that you want to create—for example:s3://mybucket/*.parquet
azure://mystorageaccount.blob.core.windows.net/mycontainer/*.parquet
TABLE_FORMAT
: the format of the table that you want to create—for example,PARQUET
Click
Run.
For more information about how to run queries, see Running interactive queries.
Examples
The following example creates a BigLake table over partitioned data in Amazon S3. The schema is autodetected.
CREATE EXTERNAL TABLE `my_dataset.my_table` WITH PARTITION COLUMNS ( sku STRING, ) WITH CONNECTION `us.my-connection` OPTIONS( hive_partition_uri_prefix = "s3://mybucket/products", uris = ['s3://mybucket/products/*'] );
The following example creates a BigLake table over partitioned data in Blob Storage. The schema is specified.
CREATE EXTERNAL TABLE `my_dataset.my_table` ( ProductId INTEGER, ProductName, STRING, ProductType, STRING ) WITH PARTITION COLUMNS ( sku STRING, ) WITH CONNECTION `us.my-connection` OPTIONS( hive_partition_uri_prefix = "azure://mystorageaccount.blob.core.windows.net/mycontainer/products", uris = ['azure://mystorageaccount.blob.core.windows.net/mycontainer/*'] );
bq
First, use the
bq mkdef
command to
create a table definition file:
bq mkdef \ --source_format=SOURCE_FORMAT \ --connection_id=REGION.CONNECTION_ID \ --hive_partitioning_mode=PARTITIONING_MODE \ --hive_partitioning_source_uri_prefix=URI_SHARED_PREFIX \ --require_hive_partition_filter=BOOLEAN \ URIS > DEFINITION_FILE
Replace the following:
SOURCE_FORMAT
: the format of the external data source. For example,CSV
.REGION
: the region that contains the connection—for example,us
.CONNECTION_ID
: the name of the connection—for example,myconnection
.PARTITIONING_MODE
: the Hive partitioning mode. Use one of the following values:AUTO
: Automatically detect the key names and types.STRINGS
: Automatically convert the key names to strings.CUSTOM
: Encode the key schema in the source URI prefix.
URI_SHARED_PREFIX
: the source URI prefix.BOOLEAN
: specifies whether to require a predicate filter at query time. This flag is optional. The default value isfalse
.URIS
: the path to the Amazon S3 or the Blob Storage folder, using wildcard format.DEFINITION_FILE
: the path to the table definition file on your local machine.
If PARTITIONING_MODE
is CUSTOM
, include the partition key schema
in the source URI prefix, using the following format:
--hive_partitioning_source_uri_prefix=GCS_URI_SHARED_PREFIX/{KEY1:TYPE1}/{KEY2:TYPE2}/...
After you create the table definition file, use the
bq mk
command to
create the BigLake table:
bq mk --external_table_definition=DEFINITION_FILE \ DATASET_NAME.TABLE_NAME \ SCHEMA
Replace the following:
DEFINITION_FILE
: the path to the table definition file.DATASET_NAME
: the name of the dataset that contains the table.TABLE_NAME
: the name of the table you're creating.SCHEMA
: specifies a path to a JSON schema file, or specifies the schema in the formfield:data_type,field:data_type,...
. To use schema auto-detection, omit this argument.
Examples
The following example uses AUTO
Hive partitioning mode for Amazon S3
data:
bq mkdef --source_format=CSV \
--connection_id=us.my-connection \
--hive_partitioning_mode=AUTO \
--hive_partitioning_source_uri_prefix=s3://mybucket/myTable \
--metadata_cache_mode=AUTOMATIC \
s3://mybucket/* > mytable_def
bq mk --external_table_definition=mytable_def \
mydataset.mytable \
Region:STRING,Quarter:STRING,Total_sales:INTEGER
The following example uses STRING
Hive partitioning mode for Amazon S3
data:
bq mkdef --source_format=CSV \
--connection_id=us.my-connection \
--hive_partitioning_mode=STRING \
--hive_partitioning_source_uri_prefix=s3://mybucket/myTable \
s3://mybucket/myTable/* > mytable_def
bq mk --external_table_definition=mytable_def \
mydataset.mytable \
Region:STRING,Quarter:STRING,Total_sales:INTEGER
The following example uses CUSTOM
Hive partitioning mode for
Blob Storage data:
bq mkdef --source_format=CSV \
--connection_id=us.my-connection \
--hive_partitioning_mode=CUSTOM \
--hive_partitioning_source_uri_prefix=azure://mystorageaccount.blob.core.windows.net/mycontainer/{dt:DATE}/{val:STRING} \
azure://mystorageaccount.blob.core.windows.net/mycontainer/* > mytable_def
bq mk --external_table_definition=mytable_def \
mydataset.mytable \
Region:STRING,Quarter:STRING,Total_sales:INTEGER
API
To set Hive partitioning using the BigQuery API, include the
hivePartitioningOptions
object in the ExternalDataConfiguration
object when you create the table definition file.
To create a BigLake table, you must also specify a value
for the connectionId
field.
If you set the hivePartitioningOptions.mode
field to CUSTOM
, you must
encode the partition key schema in the
hivePartitioningOptions.sourceUriPrefix
field as follows:
s3://BUCKET/PATH_TO_TABLE/{KEY1:TYPE1}/{KEY2:TYPE2}/...
To enforce the use of a predicate filter at query time, set the
hivePartitioningOptions.requirePartitionFilter
field to true
.
Query BigLake tables
For more information, see Query Amazon S3 data.
View resource metadata
You can view the resource metadata withINFORMATION_SCHEMA
views. When you query the
JOBS_BY_*
,
JOBS_TIMELINE_BY_*
, and
RESERVATION*
views, you must specify the query's processing location
that is collocated with the table's region. For information about BigQuery Omni
locations, see Locations. For all
other system tables, specifying the query job location is optional.
For information about the system tables that BigQuery Omni supports, see Limitations.
To query JOBS_*
and RESERVATION*
system tables, select one of the following
methods to specify the processing location:
Console
Go to the BigQuery page.
If the Editor tab isn't visible, then click
Compose new query.Click More > Query settings. The Query settings dialog opens.
In the Query settings dialog, for Additional settings > Data location, select the BigQuery region that is collocated with the BigQuery Omni region. For example, if your BigQuery Omni region is
aws-us-east-1
, specifyus-east4
.Select the remaining fields and click Save.
bq
Use the --location
flag to set the job's processing location to the
BigQuery region that is
collocated with the BigQuery Omni region.
For example, if your BigQuery Omni region is aws-us-east-1
,
specify us-east4
.
Example
bq query --use_legacy_sql=false --location=us-east4 \
"SELECT * FROM region-aws-us-east-1.INFORMATION_SCHEMA.JOBS limit 10;"
bq query --use_legacy_sql=false --location=us-east4 \
"SELECT * FROM region-aws-ap-northeast-2.INFORMATION_SCHEMA.JOBS limit 10;"
API
If you are running jobs programmatically,
set the location argument to the BigQuery region
that is collocated with the BigQuery Omni region.
For example, if your BigQuery Omni region is aws-us-east-1
,
specify us-east4
.
VPC Service Controls
You can set up an egress policy that restricts access to only specified external cloud resources from within a VPC Service Controls perimeter. For more information, see Set up VPC Service Controls for BigQuery Omni.
Limitations
For a full list of limitations that apply to BigLake tables based on Amazon S3 and Blob Storage, see Limitations.
What's next
- Learn about BigQuery Omni.
- Use the BigQuery Omni with AWS lab.
- Learn about BigLake tables.
- Learn how to export query results to Amazon S3.