Class TransferConfig (3.7.2)

TransferConfig(mapping=None, *, ignore_unknown_fields=False, **kwargs)

Represents a data transfer configuration. A transfer configuration contains all metadata needed to perform a data transfer. For example, destination_dataset_id specifies where data should be stored. When a new transfer configuration is created, the specified destination_dataset_id is created when needed and shared with the appropriate data source service account.

.. _oneof: https://proto-plus-python.readthedocs.io/en/stable/fields.html#oneofs-mutually-exclusive-fields

Attributes

NameDescription
name str
The resource name of the transfer config. Transfer config names have the form ``projects/{project_id}/locations/{region}/transferConfigs/{config_id}``. Where ``config_id`` is usually a uuid, even though it is not guaranteed or required. The name is ignored when creating a transfer config.
destination_dataset_id str
The BigQuery target dataset id. This field is a member of `oneof`_ ``destination``.
display_name str
User specified display name for the data transfer.
data_source_id str
Data source id. Cannot be changed once data transfer is created.
params google.protobuf.struct_pb2.Struct
Parameters specific to each data source. For more information see the bq tab in the 'Setting up a data transfer' section for each data source. For example the parameters for Cloud Storage transfers are listed here: https://cloud.google.com/bigquery-transfer/docs/cloud-storage-transfer#bq
schedule str
Data transfer schedule. If the data source does not support a custom schedule, this should be empty. If it is empty, the default value for the data source will be used. The specified times are in UTC. Examples of valid format: ``1st,3rd monday of month 15:30``, ``every wed,fri of jan,jun 13:15``, and ``first sunday of quarter 00:00``. See more explanation about the format here: https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format NOTE: The minimum interval time between recurring transfers depends on the data source; refer to the documentation for your data source.
schedule_options google.cloud.bigquery_datatransfer_v1.types.ScheduleOptions
Options customizing the data transfer schedule.
data_refresh_window_days int
The number of days to look back to automatically refresh the data. For example, if ``data_refresh_window_days = 10``, then every day BigQuery reingests data for [today-10, today-1], rather than ingesting data for just [today-1]. Only valid if the data source supports the feature. Set the value to 0 to use the default value.
disabled bool
Is this config disabled. When set to true, no runs are scheduled for a given transfer.
update_time google.protobuf.timestamp_pb2.Timestamp
Output only. Data transfer modification time. Ignored by server on input.
next_run_time google.protobuf.timestamp_pb2.Timestamp
Output only. Next time when data transfer will run.
state google.cloud.bigquery_datatransfer_v1.types.TransferState
Output only. State of the most recently updated transfer run.
user_id int
Deprecated. Unique ID of the user on whose behalf transfer is done.
dataset_region str
Output only. Region in which BigQuery dataset is located.
notification_pubsub_topic str
Pub/Sub topic where notifications will be sent after transfer runs associated with this transfer config finish. The format for specifying a pubsub topic is: ``projects/{project}/topics/{topic}``
email_preferences google.cloud.bigquery_datatransfer_v1.types.EmailPreferences
Email notifications will be sent according to these preferences to the email address of the user who owns this transfer config.
owner_info google.cloud.bigquery_datatransfer_v1.types.UserInfo
Output only. Information about the user whose credentials are used to transfer data. Populated only for ``transferConfigs.get`` requests. In case the user information is not available, this field will not be populated. This field is a member of `oneof`_ ``_owner_info``.

Inheritance

builtins.object > proto.message.Message > TransferConfig