Full name: projects.locations.datasets.fhirStores.import
Imports resources to the FHIR store by loading data from the specified sources. This method is optimized to load large quantities of data using import semantics that ignore some FHIR store configuration options and are not suitable for all use cases. It is primarily intended to load data into an empty FHIR store that is not being used by other clients. In cases where this method is not appropriate, consider using ExecuteBundle
to load data.
Every resource in the input must contain a client-supplied ID. Each resource is stored using the supplied ID regardless of the enableUpdateCreate
setting on the FHIR store. It is strongly advised not to include or encode any sensitive data such as patient identifiers in client-specified resource IDs. Those IDs are part of the FHIR resource path recorded in Cloud Audit Logs and Cloud Pub/Sub notifications. Those IDs can also be contained in reference fields within other resources.
The import process does not enforce referential integrity, regardless of the disableReferentialIntegrity
setting on the FHIR store. This allows the import of resources with arbitrary interdependencies without considering grouping or ordering, but if the input data contains invalid references or if some resources fail to be imported, the FHIR store might be left in a state that violates referential integrity.
The import process does not trigger Pub/Sub notification or BigQuery streaming update, regardless of how those are configured on the FHIR store.
If a resource with the specified ID already exists, the most recent version of the resource is overwritten without creating a new historical version, regardless of the disableResourceVersioning
setting on the FHIR store. If transient failures occur during the import, it's possible that successfully imported resources will be overwritten more than once.
The import operation is idempotent unless the input data contains multiple valid resources with the same ID but different contents. In that case, after the import completes, the store contains exactly one resource with that ID but there is no ordering guarantee on which version of the contents it will have. The operation result counters do not count duplicate IDs as an error and count one success for each resource in the input, which might result in a success count larger than the number of resources in the FHIR store. This often occurs when importing data organized in bundles produced by Patient-everything
where each bundle contains its own copy of a resource such as Practitioner that might be referred to by many patients.
If some resources fail to import, for example due to parsing errors, successfully imported resources are not rolled back.
The location and format of the input data is specified by the parameters in ImportResourcesRequest
. Note that if no format is specified, this method assumes the BUNDLE
format. When using the BUNDLE
format this method ignores the Bundle.type
field, except that history
bundles are rejected, and does not apply any of the bundle processing semantics for batch or transaction bundles. Unlike in ExecuteBundle
, transaction bundles are not executed as a single transaction and bundle-internal references are not rewritten. The bundle is treated as a collection of resources to be written as provided in Bundle.entry.resource
, ignoring Bundle.entry.request
. As an example, this allows the import of searchset
bundles produced by a FHIR search
or Patient-everything
operation.
This method returns an Operation
that can be used to track the status of the import by calling operations.get
.
Immediate fatal errors appear in the error
field, errors are also logged to Cloud Logging (see Viewing error logs in Cloud Logging). Otherwise, when the operation finishes, a detailed response of type ImportResourcesResponse
is returned in the response
field. The metadata
field type for this operation is OperationMetadata
.
HTTP request
POST https://healthcare.googleapis.com/v1/{name=projects/*/locations/*/datasets/*/fhirStores/*}:import
The URL uses gRPC Transcoding syntax.
Path parameters
Parameters | |
---|---|
name |
Required. The name of the FHIR store to import FHIR resources to, in the format of Authorization requires the following IAM permission on the specified resource
|
Request body
The request body contains data with the following structure:
JSON representation |
---|
{ "contentStructure": enum( |
Fields | |
---|---|
content |
The content structure in the source location. If not specified, the server treats the input source files as |
Union field To enable the Cloud Healthcare API to read from resources in your project such as Cloud Storage buckets, you must give the consumer Cloud Healthcare API service account the proper permissions. The service account is: |
|
gcs |
Cloud Storage source data location and import configuration. The Healthcare Service Agent account requires the Each Cloud Storage object should be a text file that contains the format specified in |
Response body
If successful, the response body contains an instance of Operation
.
Authorization scopes
Requires one of the following OAuth scopes:
https://www.googleapis.com/auth/cloud-healthcare
https://www.googleapis.com/auth/cloud-platform
For more information, see the Authentication Overview.
ContentStructure
The content structure of the input source files.
Enums | |
---|---|
CONTENT_STRUCTURE_UNSPECIFIED |
If the content structure is not specified, the default value BUNDLE is used. |
BUNDLE |
The source file contains one or more lines of newline-delimited JSON (ndjson). Each line is a bundle that contains one or more resources. |
RESOURCE |
The source file contains one or more lines of newline-delimited JSON (ndjson). Each line is a single resource. |
BUNDLE_PRETTY |
The entire file is one JSON bundle. The JSON can span multiple lines. |
RESOURCE_PRETTY |
The entire file is one JSON resource. The JSON can span multiple lines. |
GcsSource
Specifies the configuration for importing data from Cloud Storage.
JSON representation |
---|
{ "uri": string } |
Fields | |
---|---|
uri |
Points to a Cloud Storage URI containing file(s) to import. The URI must be in the following format:
Files matching the wildcard are expected to contain content only, no metadata. |