Skip to content

Commit f5c13a3

Browse files
committed
fix metric name
1 parent ca40723 commit f5c13a3

1 file changed

Lines changed: 2 additions & 2 deletions

File tree

.config/metrics-genai-v0.1.0.json

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -288,7 +288,7 @@
288288
}
289289
},
290290
{
291-
"id": "genai_evaluation_protectedMaterial",
291+
"id": "genai_evaluation_protected_material",
292292
"displayName": "Protected material score",
293293
"description": "Protected material score by Azure AI content safety API. The Protected material detection APIs scan the output of large language models to identify and flag known protected material. See (https://learn.microsoft.com/en-us/azure/ai-services/content-safety/concepts/protected-material?tabs=text) for more details.",
294294
"lifecycle": "Active",
@@ -305,7 +305,7 @@
305305
}
306306
},
307307
{
308-
"id": "genai_evaluation_hateUnfairness",
308+
"id": "genai_evaluation_hate_unfairness",
309309
"displayName": "Hate unfairness score",
310310
"description": "Hate and unfair score given by Azure AI evaluation. Hateful and unfair content refers to any language pertaining to hate toward or unfair representations of individuals and social groups along factors including but not limited to race, ethnicity, nationality, gender, sexual orientation, religion, immigration status, ability, personal appearance, and body size. Unfairness occurs when AI systems treat or represent social groups inequitably, creating or contributing to societal inequities. Safety evaluations annotate self-harm-related content using a 0-7 scale. See (https://learn.microsoft.com/en-us/python/api/azure-ai-evaluation/azure.ai.evaluation.hateunfairnessevaluator?view=azure-python) for more details.",
311311
"lifecycle": "Active",

0 commit comments

Comments
 (0)