You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -2326,6 +2332,10 @@ class LLMBasedMetricSpec(_common.BaseModel):
2326
2332
default=None,
2327
2333
description="""Dynamically generate rubrics using this specification.""",
2328
2334
)
2335
+
metric_resource_name: Optional[str] = Field(
2336
+
default=None,
2337
+
description="""The resource name of the metric definition. Example: projects/{project}/locations/{location}/evaluationMetrics/{evaluation_metric_id}""",
2338
+
)
2329
2339
2330
2340
2331
2341
class LLMBasedMetricSpecDict(TypedDict, total=False):
@@ -2350,6 +2360,9 @@ class LLMBasedMetricSpecDict(TypedDict, total=False):
@@ -2482,6 +2495,10 @@ class EvaluationRunMetric(_common.BaseModel):
2482
2495
metric_config: Optional[UnifiedMetric] = Field(
2483
2496
default=None, description="""The unified metric used for evaluation run."""
2484
2497
)
2498
+
metric_resource_name: Optional[str] = Field(
2499
+
default=None,
2500
+
description="""The resource name of the metric definition. Example: projects/{project}/locations/{location}/evaluationMetrics/{evaluation_metric_id}""",
2501
+
)
2485
2502
2486
2503
2487
2504
class EvaluationRunMetricDict(TypedDict, total=False):
@@ -2493,6 +2510,9 @@ class EvaluationRunMetricDict(TypedDict, total=False):
2493
2510
metric_config: Optional[UnifiedMetricDict]
2494
2511
"""The unified metric used for evaluation run."""
2495
2512
2513
+
metric_resource_name: Optional[str]
2514
+
"""The resource name of the metric definition. Example: projects/{project}/locations/{location}/evaluationMetrics/{evaluation_metric_id}"""
@@ -4439,6 +4459,10 @@ class Metric(_common.BaseModel):
4439
4459
default=None,
4440
4460
description="""Optional steering instruction parameters for the automated predefined metric.""",
4441
4461
)
4462
+
metric_resource_name: Optional[str] = Field(
4463
+
default=None,
4464
+
description="""The resource name of the metric definition. Example: projects/{project}/locations/{location}/evaluationMetrics/{evaluation_metric_id}""",
4465
+
)
4442
4466
4443
4467
# Allow extra fields to support metric-specific config fields.
4444
4468
model_config = ConfigDict(extra="allow")
@@ -4643,6 +4667,9 @@ class MetricDict(TypedDict, total=False):
4643
4667
metric_spec_parameters: Optional[dict[str, Any]]
4644
4668
"""Optional steering instruction parameters for the automated predefined metric."""
4645
4669
4670
+
metric_resource_name: Optional[str]
4671
+
"""The resource name of the metric definition. Example: projects/{project}/locations/{location}/evaluationMetrics/{evaluation_metric_id}"""
4672
+
4646
4673
4647
4674
MetricOrDict = Union[Metric, MetricDict]
4648
4675
@@ -5354,6 +5381,10 @@ class _GenerateInstanceRubricsRequest(_common.BaseModel):
5354
5381
default=None,
5355
5382
description="""Specification for how the rubrics should be generated.""",
5356
5383
)
5384
+
metric_resource_name: Optional[str] = Field(
5385
+
default=None,
5386
+
description="""The resource name of the metric definition to use for rubric generation.""",
0 commit comments