Files
openclaw-ollama-toolcall-proxy/test/fixtures/vllm-xml-response.json

22 lines
438 B
JSON
Executable File

{
"id": "chatcmpl-123",
"object": "chat.completion",
"created": 1715012345,
"model": "Qwen3.5-27B",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "<function=read>\n<parameter=path>\n/tmp/test.txt\n</parameter>\n</function>"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 56,
"completion_tokens": 31,
"total_tokens": 87
}
}