(no commit message)
This commit is contained in:
40
config.json
40
config.json
@@ -1,7 +1,7 @@
|
||||
{
|
||||
"model": null,
|
||||
"signature": {
|
||||
"description": "Determine if the predicted answer from the LLM matches the ground truth answer.",
|
||||
"description": "Answer the question. Be sure to verbally express your uncertainty in your thought process.",
|
||||
"properties": {
|
||||
"question": {
|
||||
"__dspy_field_type": "input",
|
||||
@@ -10,37 +10,25 @@
|
||||
"title": "Question",
|
||||
"type": "string"
|
||||
},
|
||||
"predicted_answer": {
|
||||
"__dspy_field_type": "input",
|
||||
"desc": "The response to the question from the LLM",
|
||||
"prefix": "Predicted Answer:",
|
||||
"title": "Predicted Answer",
|
||||
"type": "string"
|
||||
},
|
||||
"ground_truth_answer": {
|
||||
"__dspy_field_type": "input",
|
||||
"desc": "The ground truth answer to the question",
|
||||
"prefix": "Ground Truth Answer:",
|
||||
"title": "Ground Truth Answer",
|
||||
"type": "string"
|
||||
},
|
||||
"label": {
|
||||
"reasoning": {
|
||||
"__dspy_field_type": "output",
|
||||
"desc": "Whether the predicted answer is correct or incorrect",
|
||||
"enum": [
|
||||
"correct",
|
||||
"incorrect"
|
||||
],
|
||||
"prefix": "Label:",
|
||||
"title": "Label",
|
||||
"desc": "Your step by step reasoning about the target audience. Verbalize uncertainty.",
|
||||
"prefix": "Reasoning:",
|
||||
"title": "Reasoning",
|
||||
"type": "string"
|
||||
},
|
||||
"answer": {
|
||||
"__dspy_field_type": "output",
|
||||
"desc": "The response to the question",
|
||||
"prefix": "Answer:",
|
||||
"title": "Answer",
|
||||
"type": "string"
|
||||
}
|
||||
},
|
||||
"required": [
|
||||
"question",
|
||||
"predicted_answer",
|
||||
"ground_truth_answer",
|
||||
"label"
|
||||
"reasoning",
|
||||
"answer"
|
||||
],
|
||||
"title": "AudienceSig",
|
||||
"type": "object"
|
||||
|
||||
16
program.json
16
program.json
@@ -3,28 +3,24 @@
|
||||
"train": [],
|
||||
"demos": [],
|
||||
"signature": {
|
||||
"instructions": "Determine if the predicted answer from the LLM matches the ground truth answer.",
|
||||
"instructions": "Answer the question. Be sure to verbally express your uncertainty in your thought process.",
|
||||
"fields": [
|
||||
{
|
||||
"prefix": "Question:",
|
||||
"description": "The question to answer"
|
||||
},
|
||||
{
|
||||
"prefix": "Predicted Answer:",
|
||||
"description": "The response to the question from the LLM"
|
||||
"prefix": "Reasoning:",
|
||||
"description": "Your step by step reasoning about the target audience. Verbalize uncertainty."
|
||||
},
|
||||
{
|
||||
"prefix": "Ground Truth Answer:",
|
||||
"description": "The ground truth answer to the question"
|
||||
},
|
||||
{
|
||||
"prefix": "Label:",
|
||||
"description": "Whether the predicted answer is correct or incorrect"
|
||||
"prefix": "Answer:",
|
||||
"description": "The response to the question"
|
||||
}
|
||||
]
|
||||
},
|
||||
"lm": {
|
||||
"model": "together_ai/Qwen/Qwen2.5-7B-Instruct-Turbo",
|
||||
"model": "together_ai/Qwen/Qwen3-VL-32B-Instruct",
|
||||
"model_type": "chat",
|
||||
"cache": true,
|
||||
"num_retries": 3,
|
||||
|
||||
Reference in New Issue
Block a user