Abuse monitoring

As part of providing Generative AI Services to customers, Google uses the following process to detect potential abuse and violations of its Acceptable Use Policy and Prohibited Use Policy.

  • Automated detection: Google uses automated safety classifiers to detect potential abuse and violations. For technical details on how safety classifiers work, see Configure safety filers.
  • Prompt logging: If automated safety classifiers detect suspicious activity that requires further investigation into whether a customer has violated our policies, then Google may log customer prompts solely for the purpose of examining whether a violation of the AUP or Prohibited Use Policy has occurred. This data will not be used to train or fine-tune any AI/ML models. This data is stored securely for up to 30 days in the same region or multi-region selected by the customer for their project and adheres to Google Cloud assurances ofsuch as Data Residency, Access Transparency and VPC Service Controls. Customers also have the option to request an opt-out from abuse logging (see below).
  • Action: Authorized Google employees may assess the flagged prompts and may reach out to the customer for clarification. Failure to address the behavior—or recurring or severe abuse—may result in suspension or termination of the customer's access to Vertex AI or Google Cloud services.
  • Services in scope: Vertex AI API, when used with Google's large language models.
  • Customers in scope: Only customers whose use of Google Cloud is governed by the Google Cloud Platform Terms of Service are subject to prompt logging for abuse monitoring.
  • Customer opt-out: Customers may request for an exception by filling out this form. If approved, Google will not store any prompts associated with the approved Google Cloud account.

What's next