- 2.55.0 (latest)
- 2.54.0
- 2.53.0
- 2.52.0
- 2.51.0
- 2.49.0
- 2.48.0
- 2.47.0
- 2.46.0
- 2.45.0
- 2.44.0
- 2.43.0
- 2.42.0
- 2.41.0
- 2.40.0
- 2.39.0
- 2.37.0
- 2.36.0
- 2.35.0
- 2.34.0
- 2.33.0
- 2.32.0
- 2.31.0
- 2.30.0
- 2.29.0
- 2.28.0
- 2.27.0
- 2.24.0
- 2.23.0
- 2.22.0
- 2.21.0
- 2.20.0
- 2.19.0
- 2.18.0
- 2.17.0
- 2.16.0
- 2.15.0
- 2.14.0
- 2.13.0
- 2.12.0
- 2.11.0
- 2.10.0
- 2.9.0
- 2.8.0
- 2.7.0
- 2.6.0
- 2.5.0
- 2.4.0
- 2.3.18
- 2.2.3
- 2.1.23
public static final class PredictionServiceGrpc.PredictionServiceStub extends AbstractAsyncStub<PredictionServiceGrpc.PredictionServiceStub>
A stub to allow clients to do asynchronous rpc calls to service PredictionService.
AutoML Prediction API. On any input that is documented to expect a string parameter in snake_case or kebab-case, either of those cases is accepted.
Inheritance
java.lang.Object > io.grpc.stub.AbstractStub > io.grpc.stub.AbstractAsyncStub > PredictionServiceGrpc.PredictionServiceStubMethods
batchPredict(BatchPredictRequest request, StreamObserver<Operation> responseObserver)
public void batchPredict(BatchPredictRequest request, StreamObserver<Operation> responseObserver)
Perform a batch prediction. Unlike the online Predict, batch prediction result won't be immediately available in the response. Instead, a long running operation object is returned. User can poll the operation result via GetOperation method. Once the operation is done, BatchPredictResult is returned in the response field. Available for following ML problems:
- Image Classification
- Image Object Detection
- Video Classification
- Video Object Tracking * Text Extraction
- Tables
Parameters | |
---|---|
Name | Description |
request |
BatchPredictRequest |
responseObserver |
io.grpc.stub.StreamObserver<Operation> |
build(Channel channel, CallOptions callOptions)
protected PredictionServiceGrpc.PredictionServiceStub build(Channel channel, CallOptions callOptions)
Parameters | |
---|---|
Name | Description |
channel |
io.grpc.Channel |
callOptions |
io.grpc.CallOptions |
Returns | |
---|---|
Type | Description |
PredictionServiceGrpc.PredictionServiceStub |
predict(PredictRequest request, StreamObserver<PredictResponse> responseObserver)
public void predict(PredictRequest request, StreamObserver<PredictResponse> responseObserver)
Perform an online prediction. The prediction result will be directly returned in the response. Available for following ML problems, and their expected request payloads:
- Image Classification - Image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
- Image Object Detection - Image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
- Text Classification - TextSnippet, content up to 60,000 characters, UTF-8 encoded.
- Text Extraction - TextSnippet, content up to 30,000 characters, UTF-8 NFC encoded.
- Translation - TextSnippet, content up to 25,000 characters, UTF-8 encoded.
- Tables - Row, with column values matching the columns of the model, up to 5MB. Not available for FORECASTING prediction_type.
- Text Sentiment - TextSnippet, content up 500 characters, UTF-8 encoded.
Parameters | |
---|---|
Name | Description |
request |
PredictRequest |
responseObserver |
io.grpc.stub.StreamObserver<PredictResponse> |