Fixing tests

This commit is contained in:
Brandon Hancock
2025-01-02 15:51:23 -05:00

View File

@@ -1474,7 +1474,7 @@ def test_llm_call_with_ollama_llama3():
max_tokens=30, max_tokens=30,
) )
messages = [ messages = [
{"role": "user", "content": "Respond in 20 words. Which model are you??"} {"role": "user", "content": "Respond in 20 words. Which model are you?"}
] ]
response = llm.call(messages) response = llm.call(messages)