Reference documentation and code samples for the BigQuery API class Google::Cloud::Bigquery::ExtractJob::Updater.
Yielded to a block to accumulate changes for an API request.
Inherits
Methods
#cancel
def cancel()
#compression=
def compression=(value)
Sets the compression type. Not applicable when extracting models.
-
value (String) — The compression type to use for exported
files. Possible values include
GZIP
andNONE
. The default value isNONE
.
#delimiter=
def delimiter=(value)
Sets the field delimiter. Not applicable when extracting models.
-
value (String) — Delimiter to use between fields in the
exported data. Default is
,
.
#format=
def format=(new_format)
Sets the destination file format. The default value for
tables is csv
. Tables with nested or repeated fields cannot be
exported as CSV. The default value for models is ml_tf_saved_model
.
Supported values for tables:
csv
- CSVjson
- Newline-delimited JSONavro
- Avro
Supported values for models:
ml_tf_saved_model
- TensorFlow SavedModelml_xgboost_booster
- XGBoost Booster
- new_format (String) — The new source format.
#header=
def header=(value)
Print a header row in the exported file. Not applicable when extracting models.
-
value (Boolean) — Whether to print out a header row in the
results. Default is
true
.
#labels=
def labels=(value)
Sets the labels to use for the job.
-
value (Hash) —
A hash of user-provided labels associated with the job. You can use these to organize and group your jobs.
The labels applied to a resource must meet the following requirements:
- Each resource can have multiple labels, up to a maximum of 64.
- Each label must be a key-value pair.
- Keys have a minimum length of 1 character and a maximum length of 63 characters, and cannot be empty. Values can be empty, and have a maximum length of 63 characters.
- Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. All characters must use UTF-8 encoding, and international characters are allowed.
- The key portion of a label must be unique. However, you can use the same key with multiple resources.
- Keys must start with a lowercase letter or international character.
#location=
def location=(value)
Sets the geographic location where the job should run. Required except for US and EU.
- value (String) — A geographic location, such as "US", "EU" or "asia-northeast1". Required except for US and EU.
require "google/cloud/bigquery" bigquery = Google::Cloud::Bigquery.new dataset = bigquery.dataset "my_dataset" table = dataset.table "my_table" destination = "gs://my-bucket/file-name.csv" extract_job = table.extract_job destination do |j| j.location = "EU" end extract_job.wait_until_done! extract_job.done? #=> true
#refresh!
def refresh!()
#reload!
def reload!()
#rerun!
def rerun!()
#use_avro_logical_types=
def use_avro_logical_types=(value)
Indicate whether to enable extracting applicable column types (such
as TIMESTAMP
) to their corresponding AVRO logical types
(timestamp-micros
), instead of only using their raw types
(avro-long
).
Only used when #format
is set to "AVRO"
(#avro?
).
- value (Boolean) — Whether applicable column types will use their corresponding AVRO logical types.
#wait_until_done!
def wait_until_done!()