aboutsummaryrefslogtreecommitdiffstats
path: root/src/plugins/lua/gpt.lua
diff options
context:
space:
mode:
Diffstat (limited to 'src/plugins/lua/gpt.lua')
-rw-r--r--src/plugins/lua/gpt.lua16
1 files changed, 13 insertions, 3 deletions
diff --git a/src/plugins/lua/gpt.lua b/src/plugins/lua/gpt.lua
index feccae73f..e4a77c6dd 100644
--- a/src/plugins/lua/gpt.lua
+++ b/src/plugins/lua/gpt.lua
@@ -48,6 +48,8 @@ gpt {
allow_passthrough = false;
# Check messages that are apparent ham (no action and negative score)
allow_ham = false;
+ # default send response_format field { type = "json_object" }
+ include_response_format = true,
}
]])
return
@@ -393,7 +395,6 @@ local function default_llm_check(task)
model = settings.model,
max_tokens = settings.max_tokens,
temperature = settings.temperature,
- response_format = { type = "json_object" },
messages = {
{
role = 'system',
@@ -418,6 +419,11 @@ local function default_llm_check(task)
}
}
+ -- Conditionally add response_format
+ if settings.include_response_format then
+ body.response_format = { type = "json_object" }
+ end
+
upstream = settings.upstreams:get_upstream_round_robin()
local http_params = {
url = settings.url,
@@ -498,7 +504,6 @@ local function ollama_check(task)
model = settings.model,
max_tokens = settings.max_tokens,
temperature = settings.temperature,
- response_format = { type = "json_object" },
messages = {
{
role = 'system',
@@ -523,6 +528,11 @@ local function ollama_check(task)
}
}
+ -- Conditionally add response_format
+ if settings.include_response_format then
+ body.response_format = { type = "json_object" }
+ end
+
upstream = settings.upstreams:get_upstream_round_robin()
local http_params = {
url = settings.url,
@@ -618,4 +628,4 @@ if opts then
parent = id,
score = -2.0,
})
-end \ No newline at end of file
+end