tasks = [
{
"task_type": "auto_evaluation.guided_decoding",
"alias": "multi_response_option_judge",
"configuration": {
"model": "openai/gpt-4o",
"prompt": "Evaluate this response...",
"response_format": {
"type": "object",
"properties": {
"is_helpful": {
"type": "boolean",
"description": "Whether the response is helpful"
},
"quality_score": {
"type": "integer",
"minimum": 1,
"maximum": 5,
"description": "Quality score from 1 to 5"
},
"accuracy_score": {
"type": "number",
"minimum": 0.0,
"maximum": 1.0,
"description": "Accuracy as a decimal"
},
"category": {
"type": "string",
"enum": ["excellent", "good", "fair", "poor"],
"description": "Quality category"
},
"reasoning": {
"type": "string",
"description": "Explanation of the evaluation"
}
},
"required": ["is_helpful", "quality_score", "accuracy_score", "category", "reasoning"]
}
}
}
]