mirror of https://github.com/kortix-ai/suna.git
small fix
This commit is contained in:
parent
f67c65d4ae
commit
f0656b4370
|
@ -75,15 +75,15 @@ async def run_agent(
|
||||||
thread_manager.add_tool(DataProvidersTool)
|
thread_manager.add_tool(DataProvidersTool)
|
||||||
|
|
||||||
|
|
||||||
# Only include sample response if the model name does not contain "anthropic"
|
if "gemini-2.5-flash" in model_name.lower():
|
||||||
if "anthropic" not in model_name.lower():
|
system_message = { "role": "system", "content": get_gemini_system_prompt() } # example included
|
||||||
|
elif "anthropic" not in model_name.lower():
|
||||||
|
# Only include sample response if the model name does not contain "anthropic"
|
||||||
sample_response_path = os.path.join(os.path.dirname(__file__), 'sample_responses/1.txt')
|
sample_response_path = os.path.join(os.path.dirname(__file__), 'sample_responses/1.txt')
|
||||||
with open(sample_response_path, 'r') as file:
|
with open(sample_response_path, 'r') as file:
|
||||||
sample_response = file.read()
|
sample_response = file.read()
|
||||||
|
|
||||||
system_message = { "role": "system", "content": get_system_prompt() + "\n\n <sample_assistant_response>" + sample_response + "</sample_assistant_response>" }
|
system_message = { "role": "system", "content": get_system_prompt() + "\n\n <sample_assistant_response>" + sample_response + "</sample_assistant_response>" }
|
||||||
elif "gemini-2.5-flash" in model_name.lower():
|
|
||||||
system_message = { "role": "system", "content": get_gemini_system_prompt() } # example included
|
|
||||||
else:
|
else:
|
||||||
system_message = { "role": "system", "content": get_system_prompt() }
|
system_message = { "role": "system", "content": get_system_prompt() }
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue