Output di testo in Cloud Storage
Mantieni tutto organizzato con le raccolte
Salva e classifica i contenuti in base alle tue preferenze.
Utilizza il connettore TextIO per scrivere file di testo in un bucket Cloud Storage.
Per saperne di più
Per la documentazione dettagliata che include questo esempio di codice, vedi quanto segue:
Esempio di codice
Salvo quando diversamente specificato, i contenuti di questa pagina sono concessi in base alla licenza Creative Commons Attribution 4.0, mentre gli esempi di codice sono concessi in base alla licenza Apache 2.0. Per ulteriori dettagli, consulta le norme del sito di Google Developers. Java è un marchio registrato di Oracle e/o delle sue consociate.
[[["Facile da capire","easyToUnderstand","thumb-up"],["Il problema è stato risolto","solvedMyProblem","thumb-up"],["Altra","otherUp","thumb-up"]],[["Difficile da capire","hardToUnderstand","thumb-down"],["Informazioni o codice di esempio errati","incorrectInformationOrSampleCode","thumb-down"],["Mancano le informazioni o gli esempi di cui ho bisogno","missingTheInformationSamplesINeed","thumb-down"],["Problema di traduzione","translationIssue","thumb-down"],["Altra","otherDown","thumb-down"]],[],[[["\u003cp\u003eThis page demonstrates how to use the TextIO connector to write text files to a Cloud Storage bucket.\u003c/p\u003e\n"],["\u003cp\u003eThe provided code samples show how to write text data to Cloud Storage using both Java and Python.\u003c/p\u003e\n"],["\u003cp\u003eBoth the Java and Python examples use Application Default Credentials for authentication with Dataflow.\u003c/p\u003e\n"],["\u003cp\u003eThe Java sample utilizes the \u003ccode\u003eTextIO.write()\u003c/code\u003e method to define the destination, file suffix, and compression settings, while the Python sample uses the \u003ccode\u003eWriteToText\u003c/code\u003e function.\u003c/p\u003e\n"]]],[],null,["# Output text to Cloud Storage\n\nUse the TextIO connector to write text files to a Cloud Storage bucket.\n\nExplore further\n---------------\n\n\nFor detailed documentation that includes this code sample, see the following:\n\n- [Write from Dataflow to Cloud Storage](/dataflow/docs/guides/write-to-cloud-storage)\n\nCode sample\n-----------\n\n### Java\n\n\nTo authenticate to Dataflow, set up Application Default Credentials.\nFor more information, see\n\n[Set up authentication for a local development environment](/docs/authentication/set-up-adc-local-dev-environment).\n\n import java.util.Arrays;\n import java.util.List;\n import org.apache.beam.sdk.Pipeline;\n import org.apache.beam.sdk.io.Compression;\n import org.apache.beam.sdk.io.TextIO;\n import org.apache.beam.sdk.options.Description;\n import org.apache.beam.sdk.options.PipelineOptions;\n import org.apache.beam.sdk.options.PipelineOptionsFactory;\n import org.apache.beam.sdk.transforms.Create;\n\n public class BatchWriteStorage {\n public interface Options extends PipelineOptions {\n @Description(\"The Cloud Storage bucket to write to\")\n String getBucketName();\n\n void setBucketName(String value);\n }\n\n // Write text data to Cloud Storage\n public static void main(String[] args) {\n final List\u003cString\u003e wordsList = Arrays.asList(\"1\", \"2\", \"3\", \"4\");\n\n var options = PipelineOptionsFactory.fromArgs(args).withValidation().as(Options.class);\n var pipeline = Pipeline.create(options);\n pipeline\n .apply(Create\n .of(wordsList))\n .apply(TextIO\n .write()\n .to(options.getBucketName())\n .withSuffix(\".txt\")\n .withCompression(Compression.GZIP)\n );\n pipeline.run().waitUntilFinish();\n }\n }\n\n### Python\n\n\nTo authenticate to Dataflow, set up Application Default Credentials.\nFor more information, see\n\n[Set up authentication for a local development environment](/docs/authentication/set-up-adc-local-dev-environment).\n\n import argparse\n from typing import List\n\n import apache_beam as beam\n from apache_beam.io.textio import WriteToText\n from apache_beam.options.pipeline_options import PipelineOptions\n\n from typing_extensions import Self\n\n\n def write_to_cloud_storage(argv: List[str] = None) -\u003e None:\n # Parse the pipeline options passed into the application.\n class MyOptions(PipelineOptions):\n @classmethod\n # Define a custom pipeline option that specfies the Cloud Storage bucket.\n def _add_argparse_args(cls: Self, parser: argparse.ArgumentParser) -\u003e None:\n parser.add_argument(\"--output\", required=True)\n\n wordsList = [\"1\", \"2\", \"3\", \"4\"]\n options = MyOptions()\n\n with beam.Pipeline(options=options.view_as(PipelineOptions)) as pipeline:\n (\n pipeline\n | \"Create elements\" \u003e\u003e beam.Create(wordsList)\n | \"Write Files\" \u003e\u003e WriteToText(options.output, file_name_suffix=\".txt\")\n )\n\nWhat's next\n-----------\n\n\nTo search and filter code samples for other Google Cloud products, see the\n[Google Cloud sample browser](/docs/samples?product=dataflow)."]]