From 681ab587509d6a6512e3e6b4f5b9dee77f3c430d Mon Sep 17 00:00:00 2001 From: "David J. Bianco" Date: Fri, 25 Apr 2025 09:28:08 -0400 Subject: [PATCH] Changed default LLM to 'gpt-4o-mini' Only used if the model isn't specified in the config or on the command line --- SSH/ssh_server.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/SSH/ssh_server.py b/SSH/ssh_server.py index b6a00b9..10aa2b7 100755 --- a/SSH/ssh_server.py +++ b/SSH/ssh_server.py @@ -308,7 +308,7 @@ def get_user_accounts() -> dict: def choose_llm(llm_provider: Optional[str] = None, model_name: Optional[str] = None): llm_provider_name = llm_provider or config['llm'].get("llm_provider", "openai") llm_provider_name = llm_provider_name.lower() - model_name = model_name or config['llm'].get("model_name", "gpt-3.5-turbo") + model_name = model_name or config['llm'].get("model_name", "gpt-4o-mini") # Get temperature parameter from config, default to 0.2 if not specified temperature = config['llm'].getfloat("temperature", 0.2)