(no commit message)

This commit is contained in:
2026-02-13 12:46:16 -08:00
parent d033cc9a9d
commit 2df0d63981
2 changed files with 34 additions and 18 deletions

View File

@@ -1,7 +1,7 @@
{ {
"model": null, "model": null,
"signature": { "signature": {
"description": "Answer the question. Be sure to verbally express your uncertainty in your thought process.", "description": "Determine if the predicted answer from the LLM matches the ground truth answer.",
"properties": { "properties": {
"question": { "question": {
"__dspy_field_type": "input", "__dspy_field_type": "input",
@@ -10,25 +10,37 @@
"title": "Question", "title": "Question",
"type": "string" "type": "string"
}, },
"reasoning": { "predicted_answer": {
"__dspy_field_type": "output", "__dspy_field_type": "input",
"desc": "Your step by step reasoning about the target audience. Verbalize uncertainty.", "desc": "The response to the question from the LLM",
"prefix": "Reasoning:", "prefix": "Predicted Answer:",
"title": "Reasoning", "title": "Predicted Answer",
"type": "string" "type": "string"
}, },
"answer": { "ground_truth_answer": {
"__dspy_field_type": "input",
"desc": "The ground truth answer to the question",
"prefix": "Ground Truth Answer:",
"title": "Ground Truth Answer",
"type": "string"
},
"label": {
"__dspy_field_type": "output", "__dspy_field_type": "output",
"desc": "The response to the question", "desc": "Whether the predicted answer is correct or incorrect",
"prefix": "Answer:", "enum": [
"title": "Answer", "correct",
"incorrect"
],
"prefix": "Label:",
"title": "Label",
"type": "string" "type": "string"
} }
}, },
"required": [ "required": [
"question", "question",
"reasoning", "predicted_answer",
"answer" "ground_truth_answer",
"label"
], ],
"title": "AudienceSig", "title": "AudienceSig",
"type": "object" "type": "object"

View File

@@ -3,24 +3,28 @@
"train": [], "train": [],
"demos": [], "demos": [],
"signature": { "signature": {
"instructions": "Answer the question. Be sure to verbally express your uncertainty in your thought process.", "instructions": "Determine if the predicted answer from the LLM matches the ground truth answer.",
"fields": [ "fields": [
{ {
"prefix": "Question:", "prefix": "Question:",
"description": "The question to answer" "description": "The question to answer"
}, },
{ {
"prefix": "Reasoning:", "prefix": "Predicted Answer:",
"description": "Your step by step reasoning about the target audience. Verbalize uncertainty." "description": "The response to the question from the LLM"
}, },
{ {
"prefix": "Answer:", "prefix": "Ground Truth Answer:",
"description": "The response to the question" "description": "The ground truth answer to the question"
},
{
"prefix": "Label:",
"description": "Whether the predicted answer is correct or incorrect"
} }
] ]
}, },
"lm": { "lm": {
"model": "together_ai/Qwen/Qwen3-VL-32B-Instruct", "model": "together_ai/Qwen/Qwen2.5-7B-Instruct-Turbo",
"model_type": "chat", "model_type": "chat",
"cache": true, "cache": true,
"num_retries": 3, "num_retries": 3,