BigQuery API - Class Google::Cloud::Bigquery::ExtractJob::Updater (v1.42.0)

Reference documentation and code samples for the BigQuery API class Google::Cloud::Bigquery::ExtractJob::Updater.

Yielded to a block to accumulate changes for an API request.



def cancel()


def compression=(value)

Sets the compression type. Not applicable when extracting models.

  • value (String) — The compression type to use for exported files. Possible values include GZIP and NONE. The default value is NONE.


def delimiter=(value)

Sets the field delimiter. Not applicable when extracting models.

  • value (String) — Delimiter to use between fields in the exported data. Default is ,.


def format=(new_format)

Sets the destination file format. The default value for tables is csv. Tables with nested or repeated fields cannot be exported as CSV. The default value for models is ml_tf_saved_model.

Supported values for tables:

Supported values for models:

  • ml_tf_saved_model - TensorFlow SavedModel
  • ml_xgboost_booster - XGBoost Booster
  • new_format (String) — The new source format.


def header=(value)

Print a header row in the exported file. Not applicable when extracting models.

  • value (Boolean) — Whether to print out a header row in the results. Default is true.


def labels=(value)

Sets the labels to use for the job.

  • value (Hash) —

    A hash of user-provided labels associated with the job. You can use these to organize and group your jobs.

    The labels applied to a resource must meet the following requirements:

    • Each resource can have multiple labels, up to a maximum of 64.
    • Each label must be a key-value pair.
    • Keys have a minimum length of 1 character and a maximum length of 63 characters, and cannot be empty. Values can be empty, and have a maximum length of 63 characters.
    • Keys and values can contain only lowercase letters, numeric characters, underscores, and dashes. All characters must use UTF-8 encoding, and international characters are allowed.
    • The key portion of a label must be unique. However, you can use the same key with multiple resources.
    • Keys must start with a lowercase letter or international character.


def location=(value)

Sets the geographic location where the job should run. Required except for US and EU.

  • value (String) — A geographic location, such as "US", "EU" or "asia-northeast1". Required except for US and EU.
require "google/cloud/bigquery"

bigquery =
dataset = bigquery.dataset "my_dataset"
table = dataset.table "my_table"

destination = "gs://my-bucket/file-name.csv"
extract_job = table.extract_job destination do |j|
  j.location = "EU"

extract_job.done? #=> true


def refresh!()
Alias Of: #reload!


def reload!()


def rerun!()


def use_avro_logical_types=(value)

Indicate whether to enable extracting applicable column types (such as TIMESTAMP) to their corresponding AVRO logical types (timestamp-micros), instead of only using their raw types (avro-long).

Only used when #format is set to "AVRO" (#avro?).

  • value (Boolean) — Whether applicable column types will use their corresponding AVRO logical types.


def wait_until_done!()