- 0.61.0 (latest)
- 0.60.0
- 0.59.0
- 0.58.0
- 0.57.0
- 0.56.0
- 0.55.0
- 0.54.0
- 0.53.0
- 0.52.0
- 0.51.0
- 0.50.0
- 0.49.0
- 0.48.0
- 0.47.0
- 0.46.0
- 0.45.0
- 0.44.0
- 0.43.0
- 0.42.0
- 0.41.0
- 0.40.0
- 0.39.0
- 0.38.0
- 0.37.0
- 0.36.0
- 0.35.0
- 0.34.0
- 0.33.0
- 0.32.0
- 0.31.0
- 0.30.0
- 0.29.0
- 0.28.0
- 0.27.0
- 0.26.0
- 0.25.0
- 0.24.0
- 0.23.0
- 0.22.0
- 0.21.0
- 0.20.0
- 0.19.0
- 0.18.0
- 0.17.0
- 0.16.0
- 0.15.0
- 0.14.0
- 0.13.0
- 0.12.0
- 0.11.0
- 0.10.0
- 0.9.1
- 0.8.0
- 0.7.0
- 0.6.0
- 0.5.0
- 0.4.0
- 0.3.0
- 0.2.0
- 0.1.0
Reference documentation and code samples for the Vertex AI V1 API class Google::Cloud::AIPlatform::V1::LlmUtilityService::Client.
Client for the LlmUtilityService service.
Service for LLM related utility functions.
Inherits
- Object
Methods
.configure
def self.configure() { |config| ... } -> Client::Configuration
Configure the LlmUtilityService Client class.
See Configuration for a description of the configuration fields.
- (config) — Configure the Client client.
- config (Client::Configuration)
# Modify the configuration for all LlmUtilityService clients ::Google::Cloud::AIPlatform::V1::LlmUtilityService::Client.configure do |config| config.timeout = 10.0 end
#compute_tokens
def compute_tokens(request, options = nil) -> ::Google::Cloud::AIPlatform::V1::ComputeTokensResponse
def compute_tokens(endpoint: nil, instances: nil, model: nil, contents: nil) -> ::Google::Cloud::AIPlatform::V1::ComputeTokensResponse
Return a list of tokens based on the input text.
def compute_tokens(request, options = nil) -> ::Google::Cloud::AIPlatform::V1::ComputeTokensResponse
compute_tokens
via a request object, either of type
ComputeTokensRequest or an equivalent Hash.
- request (::Google::Cloud::AIPlatform::V1::ComputeTokensRequest, ::Hash) — A request object representing the call parameters. Required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash.
- options (::Gapic::CallOptions, ::Hash) — Overrides the default settings for this call, e.g, timeout, retries, etc. Optional.
def compute_tokens(endpoint: nil, instances: nil, model: nil, contents: nil) -> ::Google::Cloud::AIPlatform::V1::ComputeTokensResponse
compute_tokens
via keyword arguments. Note that at
least one keyword argument is required. To specify no parameters, or to keep all
the default parameter values, pass an empty Hash as a request object (see above).
- endpoint (::String) — Required. The name of the Endpoint requested to get lists of tokens and token ids.
- instances (::Array<::Google::Protobuf::Value, ::Hash>) — Optional. The instances that are the input to token computing API call. Schema is identical to the prediction schema of the text model, even for the non-text models, like chat models, or Codey models.
- model (::String) — Optional. The name of the publisher model requested to serve the prediction. Format: projects/{project}/locations/{location}/publishers//models/
- contents (::Array<::Google::Cloud::AIPlatform::V1::Content, ::Hash>) — Optional. Input content.
- (response, operation) — Access the result along with the RPC operation
- response (::Google::Cloud::AIPlatform::V1::ComputeTokensResponse)
- operation (::GRPC::ActiveCall::Operation)
- (::Google::Cloud::Error) — if the RPC is aborted.
Basic example
require "google/cloud/ai_platform/v1" # Create a client object. The client can be reused for multiple calls. client = Google::Cloud::AIPlatform::V1::LlmUtilityService::Client.new # Create a request. To set request fields, pass in keyword arguments. request = Google::Cloud::AIPlatform::V1::ComputeTokensRequest.new # Call the compute_tokens method. result = client.compute_tokens request # The returned object is of type Google::Cloud::AIPlatform::V1::ComputeTokensResponse. p result
#configure
def configure() { |config| ... } -> Client::Configuration
Configure the LlmUtilityService Client instance.
The configuration is set to the derived mode, meaning that values can be changed, but structural changes (adding new fields, etc.) are not allowed. Structural changes should be made on Client.configure.
See Configuration for a description of the configuration fields.
- (config) — Configure the Client client.
- config (Client::Configuration)
#count_tokens
def count_tokens(request, options = nil) -> ::Google::Cloud::AIPlatform::V1::CountTokensResponse
def count_tokens(endpoint: nil, model: nil, instances: nil, contents: nil) -> ::Google::Cloud::AIPlatform::V1::CountTokensResponse
Perform a token counting.
def count_tokens(request, options = nil) -> ::Google::Cloud::AIPlatform::V1::CountTokensResponse
count_tokens
via a request object, either of type
CountTokensRequest or an equivalent Hash.
- request (::Google::Cloud::AIPlatform::V1::CountTokensRequest, ::Hash) — A request object representing the call parameters. Required. To specify no parameters, or to keep all the default parameter values, pass an empty Hash.
- options (::Gapic::CallOptions, ::Hash) — Overrides the default settings for this call, e.g, timeout, retries, etc. Optional.
def count_tokens(endpoint: nil, model: nil, instances: nil, contents: nil) -> ::Google::Cloud::AIPlatform::V1::CountTokensResponse
count_tokens
via keyword arguments. Note that at
least one keyword argument is required. To specify no parameters, or to keep all
the default parameter values, pass an empty Hash as a request object (see above).
-
endpoint (::String) — Required. The name of the Endpoint requested to perform token counting.
Format:
projects/{project}/locations/{location}/endpoints/{endpoint}
-
model (::String) — Required. The name of the publisher model requested to serve the
prediction. Format:
projects/{project}/locations/{location}/publishers/*/models/*
- instances (::Array<::Google::Protobuf::Value, ::Hash>) — Required. The instances that are the input to token counting call. Schema is identical to the prediction schema of the underlying model.
- contents (::Array<::Google::Cloud::AIPlatform::V1::Content, ::Hash>) — Required. Input content.
- (response, operation) — Access the result along with the RPC operation
- response (::Google::Cloud::AIPlatform::V1::CountTokensResponse)
- operation (::GRPC::ActiveCall::Operation)
- (::Google::Cloud::Error) — if the RPC is aborted.
Basic example
require "google/cloud/ai_platform/v1" # Create a client object. The client can be reused for multiple calls. client = Google::Cloud::AIPlatform::V1::LlmUtilityService::Client.new # Create a request. To set request fields, pass in keyword arguments. request = Google::Cloud::AIPlatform::V1::CountTokensRequest.new # Call the count_tokens method. result = client.count_tokens request # The returned object is of type Google::Cloud::AIPlatform::V1::CountTokensResponse. p result
#iam_policy_client
def iam_policy_client() -> Google::Iam::V1::IAMPolicy::Client
Get the associated client for mix-in of the IAMPolicy.
- (Google::Iam::V1::IAMPolicy::Client)
#initialize
def initialize() { |config| ... } -> Client
Create a new LlmUtilityService client object.
- (config) — Configure the LlmUtilityService client.
- config (Client::Configuration)
- (Client) — a new instance of Client
# Create a client using the default configuration client = ::Google::Cloud::AIPlatform::V1::LlmUtilityService::Client.new # Create a client using a custom configuration client = ::Google::Cloud::AIPlatform::V1::LlmUtilityService::Client.new do |config| config.timeout = 10.0 end
#location_client
def location_client() -> Google::Cloud::Location::Locations::Client
Get the associated client for mix-in of the Locations.
- (Google::Cloud::Location::Locations::Client)
#universe_domain
def universe_domain() -> String
The effective universe domain
- (String)