- 3.13.0 (latest)
- 3.12.0
- 3.11.0
- 3.10.0
- 3.9.0
- 3.8.0
- 3.7.0
- 3.6.0
- 3.5.0
- 3.4.0
- 3.3.0
- 3.2.0
- 3.1.0
- 3.0.0
- 2.28.0
- 2.27.0
- 2.26.0
- 2.25.0
- 2.24.0
- 2.23.0
- 2.22.0
- 2.21.0
- 2.20.0
- 2.19.0
- 2.18.0
- 2.17.0
- 2.16.0
- 2.15.0
- 2.14.0
- 2.13.0
- 2.12.0
- 2.11.0
- 2.10.0
- 2.9.0
- 2.8.0
- 2.7.0
- 2.6.0
- 2.5.0
- 2.4.0
- 2.3.0
- 2.2.0
- 2.1.0
- 2.0.0
- 1.8.0
- 1.7.0
- 1.6.0
- 1.5.0
- 1.4.0
- 1.3.0
- 1.2.0
- 1.1.0
- 1.0.0
public sealed class Explanation : IMessage<Explanation>, IEquatable<Explanation>, IDeepCloneable<Explanation>, IBufferMessage, IMessage
Reference documentation and code samples for the Cloud AI Platform v1 API class Explanation.
Explanation of a prediction (provided in [PredictResponse.predictions][google.cloud.aiplatform.v1.PredictResponse.predictions]) produced by the Model on a given [instance][google.cloud.aiplatform.v1.ExplainRequest.instances].
Implements
IMessageExplanation, IEquatableExplanation, IDeepCloneableExplanation, IBufferMessage, IMessageNamespace
Google.Cloud.AIPlatform.V1Assembly
Google.Cloud.AIPlatform.V1.dll
Constructors
Explanation()
public Explanation()
Explanation(Explanation)
public Explanation(Explanation other)
Parameter | |
---|---|
Name | Description |
other |
Explanation |
Properties
Attributions
public RepeatedField<Attribution> Attributions { get; }
Output only. Feature attributions grouped by predicted outputs.
For Models that predict only one output, such as regression Models that predict only one score, there is only one attibution that explains the predicted output. For Models that predict multiple outputs, such as multiclass Models that predict multiple classes, each element explains one specific item. [Attribution.output_index][google.cloud.aiplatform.v1.Attribution.output_index] can be used to identify which output this attribution is explaining.
By default, we provide Shapley values for the predicted class. However,
you can configure the explanation request to generate Shapley values for
any other classes too. For example, if a model predicts a probability of
0.4
for approving a loan application, the model's decision is to reject
the application since p(reject) = 0.6 > p(approve) = 0.4
, and the default
Shapley values would be computed for rejection decision and not approval,
even though the latter might be the positive class.
If users set [ExplanationParameters.top_k][google.cloud.aiplatform.v1.ExplanationParameters.top_k], the attributions are sorted by [instance_output_value][google.cloud.aiplatform.v1.Attribution.instance_output_value] in descending order. If [ExplanationParameters.output_indices][google.cloud.aiplatform.v1.ExplanationParameters.output_indices] is specified, the attributions are stored by [Attribution.output_index][google.cloud.aiplatform.v1.Attribution.output_index] in the same order as they appear in the output_indices.
Property Value | |
---|---|
Type | Description |
RepeatedFieldAttribution |
Neighbors
public RepeatedField<Neighbor> Neighbors { get; }
Output only. List of the nearest neighbors for example-based explanations.
For models deployed with the examples explanations feature enabled, the attributions field is empty and instead the neighbors field is populated.
Property Value | |
---|---|
Type | Description |
RepeatedFieldNeighbor |