Resource: DataSet
A collection of data sources sent for processing.
JSON representation |
---|
{ "name": string, "dataNames": [ string ], "dataSources": [ { object ( |
Fields | |
---|---|
name |
The dataset name, which will be used for querying, status and unload requests. This must be unique within a project. |
dataNames[] |
If left empty, all dimension names are included. This field works as a filter to avoid regenerating the data. |
dataSources[] |
Input data. |
state |
Dataset state in the system. |
status |
Dataset processing status. |
ttl |
Periodically we discard dataset A duration in seconds with up to nine fractional digits, ending with ' |
DataSource
A data source consists of multiple Event
objects stored on Cloud Storage. Each Event should be in JSON format, with one Event per line, also known as JSON Lines format.
JSON representation |
---|
{
"uri": string,
"bqMapping": {
object ( |
Fields | |
---|---|
uri |
Data source URI.
|
bqMapping |
For BigQuery inputs defines the columns that should be used for dimensions (including time and group ID). |
BigqueryMapping
Mapping of BigQuery columns to timestamp, groupId and dimensions.
JSON representation |
---|
{ "timestampColumn": string, "groupIdColumn": string, "dimensionColumn": [ string ] } |
Fields | |
---|---|
timestampColumn |
The column which should be used as the event timestamps. If not specified 'Timestamp' is used by default. The column may have TIMESTAMP or INT64 type (the latter is interpreted as microseconds since the Unix epoch). |
groupIdColumn |
The column which should be used as the group ID (grouping events into sessions). If not specified 'GroupId' is used by default, if the input table does not have such a column, random unique group IDs are generated automatically (different group ID per input row). |
dimensionColumn[] |
The list of columns that should be translated to dimensions. If empty, all columns are translated to dimensions. The timestamp and groupId columns should not be listed here again. Columns are expected to have primitive types (STRING, INT64, FLOAT64 or NUMERIC). |
State
DataSet state.
Enums | |
---|---|
STATE_UNSPECIFIED |
Unspecified / undefined state. |
UNKNOWN |
Dataset is unknown to the system; we have never seen this dataset before or we have seen this dataset but have fully GC-ed it. |
PENDING |
Dataset processing is pending. |
LOADING |
Dataset is loading. |
LOADED |
Dataset is loaded and can be queried. |
UNLOADING |
Dataset is unloading. |
UNLOADED |
Dataset is unloaded and is removed from the system. |
FAILED |
Dataset processing failed. |
Status
The Status
type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by gRPC. Each Status
message contains three pieces of data: error code, error message, and error details.
You can find out more about this error model and how to work with it in the API Design Guide.
JSON representation |
---|
{ "code": integer, "message": string, "details": [ { "@type": string, field1: ..., ... } ] } |
Fields | |
---|---|
code |
The status code, which should be an enum value of |
message |
A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the |
details[] |
A list of messages that carry the error details. There is a common set of message types for APIs to use. An object containing fields of an arbitrary type. An additional field |
Methods |
|
---|---|
|
Append events to a LOADED DataSet . |
|
Create a DataSet from data stored on Cloud Storage. |
|
Delete a DataSet from the system. |
|
Evaluate an explicit slice from a loaded DataSet . |
|
Lists DataSets under the project. |
|
Execute a Timeseries Insights query over a loaded DataSet . |