Required. Cloud Storage URIs to input files. Each URI can be up to 2000 characters long. URIs can match the full object path (for example, gs://bucket/directory/object.json) or a pattern matching one or more files, such as gs://bucket/directory/*.json.
A request can contain at most 100 files (or 100,000 files if dataSchema is content). Each file can be up to 2 GB (or 100 MB if dataSchema is content).
dataSchema
string
The schema to use when parsing the data from the source.
Supported values for document imports:
document (default): One JSON Document per line. Each document must have a valid Document.id.
content: Unstructured data (e.g. PDF, HTML). Each file matched by inputUris becomes a document, with the ID set to the first 128 bits of SHA256(URI) encoded as a hex string.
custom: One custom data JSON per row in arbitrary format that conforms to the defined Schema of the data store. This can only be used by the GENERIC Data Store vertical.
csv: A CSV file with header conforming to the defined Schema of the data store. Each entry after the header is imported as a Document. This can only be used by the GENERIC Data Store vertical.
Supported values for user event imports:
user_event (default): One JSON UserEvent per line.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-06-27 UTC."],[[["\u003cp\u003eInput content is sourced from Cloud Storage locations, defined by URIs in a JSON representation.\u003c/p\u003e\n"],["\u003cp\u003eThe \u003ccode\u003einputUris\u003c/code\u003e field requires an array of strings, each representing a Cloud Storage URI that can be up to 2000 characters long, and these URIs can target individual files or multiple files via patterns, but the request can have at most 100 or 100,000 files depending on the selected data schema.\u003c/p\u003e\n"],["\u003cp\u003eThe \u003ccode\u003edataSchema\u003c/code\u003e field specifies the structure of the input data, with options such as \u003ccode\u003edocument\u003c/code\u003e for JSON documents, \u003ccode\u003econtent\u003c/code\u003e for unstructured files, \u003ccode\u003ecustom\u003c/code\u003e for arbitrary JSON, or \u003ccode\u003ecsv\u003c/code\u003e for comma-separated value files.\u003c/p\u003e\n"],["\u003cp\u003eThe size of each file being imported can be up to 2GB, or up to 100MB if the \u003ccode\u003econtent\u003c/code\u003e data schema is used.\u003c/p\u003e\n"],["\u003cp\u003eThe \u003ccode\u003euser_event\u003c/code\u003e option exists as a value for the \u003ccode\u003edataSchema\u003c/code\u003e, used to import user event data, with one JSON \u003ccode\u003eUserEvent\u003c/code\u003e per line.\u003c/p\u003e\n"]]],[],null,[]]