|
50 | 50 | ) |
51 | 51 | from vertexai.preview.evaluation.metrics import pairwise_metric |
52 | 52 | from vertexai.preview.evaluation.metrics import pointwise_metric |
53 | | -from vertexai.preview.evaluation.metrics import ( |
54 | | - rubric_based_metric, |
55 | | -) |
56 | 53 | from google.protobuf import json_format |
57 | 54 |
|
58 | 55 |
|
@@ -260,14 +257,23 @@ def build_request( |
260 | 257 | _default_templates.PAIRWISE_MULTIMODAL_UNDERSTANDING_RUBRIC_CRITIQUE_TEMPLATE, |
261 | 258 | _default_templates.PAIRWISE_TEXT_QUALITY_RUBRIC_CRITIQUE_TEMPLATE, |
262 | 259 | ): |
263 | | - model_based_metric_instance_input["rubrics"] = _format_rubrics( |
264 | | - model_based_metric_instance_input["rubrics"] |
| 260 | + model_based_metric_instance_input[ |
| 261 | + constants.Dataset.RUBRICS_COLUMN |
| 262 | + ] = _format_rubrics( |
| 263 | + model_based_metric_instance_input[constants.Dataset.RUBRICS_COLUMN] |
| 264 | + ) |
| 265 | + if ( |
| 266 | + constants.Dataset.RUBRICS_COLUMN in model_based_metric_instance_input |
| 267 | + and isinstance( |
| 268 | + model_based_metric_instance_input[constants.Dataset.RUBRICS_COLUMN], |
| 269 | + List, |
| 270 | + ) |
| 271 | + ): |
| 272 | + model_based_metric_instance_input[ |
| 273 | + constants.Dataset.RUBRICS_COLUMN |
| 274 | + ] = "\n".join( |
| 275 | + model_based_metric_instance_input[constants.Dataset.RUBRICS_COLUMN] |
265 | 276 | ) |
266 | | - if isinstance(metric, rubric_based_metric.RubricBasedMetric): |
267 | | - if isinstance(model_based_metric_instance_input["rubrics"], List): |
268 | | - model_based_metric_instance_input["rubrics"] = "\n".join( |
269 | | - model_based_metric_instance_input["rubrics"] |
270 | | - ) |
271 | 277 |
|
272 | 278 | if metric_name == constants.Metric.EXACT_MATCH: |
273 | 279 | instance = gapic_eval_service_types.ExactMatchInput( |
|
0 commit comments