From 10b3cefd4e8dd99c73cf11c838d80ad6238a0b57 Mon Sep 17 00:00:00 2001 From: Vsevolod Stakhov Date: Mon, 22 Jul 2024 13:34:37 +0100 Subject: [PATCH] [Minor] Use gpt-4o-mini by default --- conf/modules.d/gpt.conf | 2 +- src/plugins/lua/gpt.lua | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/conf/modules.d/gpt.conf b/conf/modules.d/gpt.conf index eac9952e5..c76a08c92 100644 --- a/conf/modules.d/gpt.conf +++ b/conf/modules.d/gpt.conf @@ -18,7 +18,7 @@ gpt { # Your key to access the API (add this to enable this plugin) #api_key = "xxx"; # Model name - model = "gpt-3.5-turbo"; + model = "gpt-4o-mini"; # Maximum tokens to generate max_tokens = 1000; # Temperature for sampling diff --git a/src/plugins/lua/gpt.lua b/src/plugins/lua/gpt.lua index 5e19d2998..61217cbf8 100644 --- a/src/plugins/lua/gpt.lua +++ b/src/plugins/lua/gpt.lua @@ -27,7 +27,7 @@ gpt { # Your key to access the API api_key = "xxx"; # Model name - model = "gpt-3.5-turbo"; + model = "gpt-4o-mini"; # Maximum tokens to generate max_tokens = 1000; # Temperature for sampling @@ -69,7 +69,7 @@ local default_symbols_to_except = { local settings = { type = 'openai', api_key = nil, - model = 'gpt-3.5-turbo', + model = 'gpt-4o-mini', max_tokens = 1000, temperature = 0.0, timeout = 10, -- 2.39.5