public static final class PredictionServiceGrpc.PredictionServiceStub extends AbstractAsyncStub<PredictionServiceGrpc.PredictionServiceStub>
AutoML Prediction API.
On any input that is documented to expect a string parameter in
snake_case or dash-case, either of those cases is accepted.
Inheritance
java.lang.Object >
io.grpc.stub.AbstractStub >
io.grpc.stub.AbstractAsyncStub >
PredictionServiceGrpc.PredictionServiceStub
Inherited Members
io.grpc.stub.AbstractAsyncStub.<T>newStub(io.grpc.stub.AbstractStub.StubFactory<T>,io.grpc.Channel)
io.grpc.stub.AbstractAsyncStub.<T>newStub(io.grpc.stub.AbstractStub.StubFactory<T>,io.grpc.Channel,io.grpc.CallOptions)
io.grpc.stub.AbstractStub.<T>withOption(io.grpc.CallOptions.Key<T>,T)
io.grpc.stub.AbstractStub.build(io.grpc.Channel,io.grpc.CallOptions)
io.grpc.stub.AbstractStub.getCallOptions()
io.grpc.stub.AbstractStub.getChannel()
io.grpc.stub.AbstractStub.withCallCredentials(io.grpc.CallCredentials)
io.grpc.stub.AbstractStub.withChannel(io.grpc.Channel)
io.grpc.stub.AbstractStub.withCompression(java.lang.String)
io.grpc.stub.AbstractStub.withDeadline(io.grpc.Deadline)
io.grpc.stub.AbstractStub.withDeadlineAfter(long,java.util.concurrent.TimeUnit)
io.grpc.stub.AbstractStub.withExecutor(java.util.concurrent.Executor)
io.grpc.stub.AbstractStub.withInterceptors(io.grpc.ClientInterceptor...)
io.grpc.stub.AbstractStub.withMaxInboundMessageSize(int)
io.grpc.stub.AbstractStub.withMaxOutboundMessageSize(int)
io.grpc.stub.AbstractStub.withWaitForReady()
Methods
batchPredict(BatchPredictRequest request, StreamObserver<Operation> responseObserver)
public void batchPredict(BatchPredictRequest request, StreamObserver<Operation> responseObserver)
Perform a batch prediction. Unlike the online Predict, batch
prediction result won't be immediately available in the response. Instead,
a long running operation object is returned. User can poll the operation
result via GetOperation
method. Once the operation is done, BatchPredictResult is returned in
the response field.
Available for following ML scenarios:
- AutoML Vision Classification
- AutoML Vision Object Detection
- AutoML Video Intelligence Classification
- AutoML Video Intelligence Object Tracking * AutoML Natural Language Classification
- AutoML Natural Language Entity Extraction
- AutoML Natural Language Sentiment Analysis
- AutoML Tables
Parameters
build(Channel channel, CallOptions callOptions)
protected PredictionServiceGrpc.PredictionServiceStub build(Channel channel, CallOptions callOptions)
Parameters
Name | Description |
channel | io.grpc.Channel
|
callOptions | io.grpc.CallOptions
|
Returns
Overrides
io.grpc.stub.AbstractStub.build(io.grpc.Channel,io.grpc.CallOptions)
predict(PredictRequest request, StreamObserver<PredictResponse> responseObserver)
public void predict(PredictRequest request, StreamObserver<PredictResponse> responseObserver)
Perform an online prediction. The prediction result is directly
returned in the response.
Available for following ML scenarios, and their expected request payloads:
AutoML Vision Classification
- An image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
AutoML Vision Object Detection
- An image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
AutoML Natural Language Classification
- A TextSnippet up to 60,000 characters, UTF-8 encoded or a document in
.PDF, .TIF or .TIFF format with size upto 2MB.
AutoML Natural Language Entity Extraction
- A TextSnippet up to 10,000 characters, UTF-8 NFC encoded or a document
in .PDF, .TIF or .TIFF format with size upto 20MB.
AutoML Natural Language Sentiment Analysis
- A TextSnippet up to 60,000 characters, UTF-8 encoded or a document in
.PDF, .TIF or .TIFF format with size upto 2MB.
AutoML Translation
- A TextSnippet up to 25,000 characters, UTF-8 encoded.
AutoML Tables
- A row with column values matching
the columns of the model, up to 5MB. Not available for FORECASTING
prediction_type
.
Parameters