Files
HadTavern/.agentui/vars/p_prompt_combine_claude.json

105 lines
2.5 KiB
JSON

{
"snapshot": {
"incoming": {
"method": "POST",
"url": "http://localhost/test",
"path": "/test",
"query": "",
"headers": {
"x": "X-HEADER"
},
"json": {
"messages": [
{
"role": "system",
"content": "Системный-тест CLAUDE"
},
{
"role": "user",
"content": "Прив"
},
{
"role": "assistant",
"content": "Привет!"
}
]
}
},
"params": {
"temperature": 0.25
},
"model": "gpt-x",
"vendor_format": "openai",
"system": "",
"OUT": {
"n1": {
"result": {
"echo": {
"url": "http://mock.local/v1/messages",
"headers": {
"Content-Type": "application/json"
},
"payload": {
"model": "gpt-x",
"system": [
{
"type": "text",
"text": "Ты — Narrator-chan."
},
{
"type": "text",
"text": "Системный-тест CLAUDE"
}
],
"messages": [
{
"role": "user",
"content": [
{
"type": "text",
"text": "Системный-тест CLAUDE"
}
]
},
{
"role": "user",
"content": [
{
"type": "text",
"text": "Прив"
}
]
},
{
"role": "assistant",
"content": [
{
"type": "text",
"text": "Привет!"
}
]
},
{
"role": "user",
"content": [
{
"type": "text",
"text": "как лела"
}
]
}
]
}
}
},
"response_text": "http://mock.local/v1/messages"
}
},
"OUT_TEXT": {
"n1": "http://mock.local/v1/messages"
},
"LAST_NODE": "n1",
"OUT1": "http://mock.local/v1/messages",
"EXEC_TRACE": "n1(ProviderCall)"
}
}