mirror of
https://github.com/crewAIInc/crewAI.git
synced 2026-01-19 21:08:13 +00:00
Compare commits
9 Commits
devin/1768
...
devin/1751
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
d86259b0b9 | ||
|
|
1f106015ea | ||
|
|
f388890971 | ||
|
|
3220575d29 | ||
|
|
6be376f804 | ||
|
|
5b548d618d | ||
|
|
2ffed3ccf0 | ||
|
|
1ea3fc44fa | ||
|
|
6e91a26785 |
@@ -41,6 +41,7 @@ from crewai.agents.parser import (
|
|||||||
)
|
)
|
||||||
from crewai.flow.flow_trackable import FlowTrackable
|
from crewai.flow.flow_trackable import FlowTrackable
|
||||||
from crewai.llm import LLM
|
from crewai.llm import LLM
|
||||||
|
from crewai.llms.base_llm import BaseLLM
|
||||||
from crewai.tools.base_tool import BaseTool
|
from crewai.tools.base_tool import BaseTool
|
||||||
from crewai.tools.structured_tool import CrewStructuredTool
|
from crewai.tools.structured_tool import CrewStructuredTool
|
||||||
from crewai.utilities import I18N
|
from crewai.utilities import I18N
|
||||||
@@ -209,8 +210,8 @@ class LiteAgent(FlowTrackable, BaseModel):
|
|||||||
def setup_llm(self):
|
def setup_llm(self):
|
||||||
"""Set up the LLM and other components after initialization."""
|
"""Set up the LLM and other components after initialization."""
|
||||||
self.llm = create_llm(self.llm)
|
self.llm = create_llm(self.llm)
|
||||||
if not isinstance(self.llm, LLM):
|
if not isinstance(self.llm, BaseLLM):
|
||||||
raise ValueError("Unable to create LLM instance")
|
raise ValueError(f"Expected LLM instance of type BaseLLM, got {type(self.llm).__name__}")
|
||||||
|
|
||||||
# Initialize callbacks
|
# Initialize callbacks
|
||||||
token_callback = TokenCalcHandler(token_cost_process=self._token_process)
|
token_callback = TokenCalcHandler(token_cost_process=self._token_process)
|
||||||
@@ -232,7 +233,8 @@ class LiteAgent(FlowTrackable, BaseModel):
|
|||||||
elif isinstance(self.guardrail, str):
|
elif isinstance(self.guardrail, str):
|
||||||
from crewai.tasks.llm_guardrail import LLMGuardrail
|
from crewai.tasks.llm_guardrail import LLMGuardrail
|
||||||
|
|
||||||
assert isinstance(self.llm, LLM)
|
if not isinstance(self.llm, BaseLLM):
|
||||||
|
raise TypeError(f"Guardrail requires LLM instance of type BaseLLM, got {type(self.llm).__name__}")
|
||||||
|
|
||||||
self._guardrail = LLMGuardrail(description=self.guardrail, llm=self.llm)
|
self._guardrail = LLMGuardrail(description=self.guardrail, llm=self.llm)
|
||||||
|
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ Classes:
|
|||||||
|
|
||||||
from typing import Any, Optional, Tuple
|
from typing import Any, Optional, Tuple
|
||||||
|
|
||||||
from crewai.llm import LLM
|
from crewai.llms.base_llm import BaseLLM
|
||||||
from crewai.tasks.task_output import TaskOutput
|
from crewai.tasks.task_output import TaskOutput
|
||||||
from crewai.utilities.logger import Logger
|
from crewai.utilities.logger import Logger
|
||||||
|
|
||||||
@@ -47,7 +47,7 @@ class HallucinationGuardrail:
|
|||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
context: str,
|
context: str,
|
||||||
llm: LLM,
|
llm: BaseLLM,
|
||||||
threshold: Optional[float] = None,
|
threshold: Optional[float] = None,
|
||||||
tool_response: str = "",
|
tool_response: str = "",
|
||||||
):
|
):
|
||||||
@@ -60,7 +60,7 @@ class HallucinationGuardrail:
|
|||||||
tool_response: Optional tool response information that would be used in evaluation.
|
tool_response: Optional tool response information that would be used in evaluation.
|
||||||
"""
|
"""
|
||||||
self.context = context
|
self.context = context
|
||||||
self.llm: LLM = llm
|
self.llm: BaseLLM = llm
|
||||||
self.threshold = threshold
|
self.threshold = threshold
|
||||||
self.tool_response = tool_response
|
self.tool_response = tool_response
|
||||||
self._logger = Logger(verbose=True)
|
self._logger = Logger(verbose=True)
|
||||||
|
|||||||
@@ -1,10 +1,9 @@
|
|||||||
from typing import Any, Optional, Tuple
|
from typing import Any, Tuple
|
||||||
|
|
||||||
from pydantic import BaseModel, Field
|
from pydantic import BaseModel, Field
|
||||||
|
|
||||||
from crewai.agent import Agent, LiteAgentOutput
|
from crewai.agent import Agent, LiteAgentOutput
|
||||||
from crewai.llm import LLM
|
from crewai.llms.base_llm import BaseLLM
|
||||||
from crewai.task import Task
|
|
||||||
from crewai.tasks.task_output import TaskOutput
|
from crewai.tasks.task_output import TaskOutput
|
||||||
|
|
||||||
|
|
||||||
@@ -32,11 +31,11 @@ class LLMGuardrail:
|
|||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
description: str,
|
description: str,
|
||||||
llm: LLM,
|
llm: BaseLLM,
|
||||||
):
|
):
|
||||||
self.description = description
|
self.description = description
|
||||||
|
|
||||||
self.llm: LLM = llm
|
self.llm: BaseLLM = llm
|
||||||
|
|
||||||
def _validate_output(self, task_output: TaskOutput) -> LiteAgentOutput:
|
def _validate_output(self, task_output: TaskOutput) -> LiteAgentOutput:
|
||||||
agent = Agent(
|
agent = Agent(
|
||||||
|
|||||||
@@ -146,12 +146,12 @@ def test_lite_agent_with_tools():
|
|||||||
"What is the population of Tokyo and how many people would that be per square kilometer if Tokyo's area is 2,194 square kilometers?"
|
"What is the population of Tokyo and how many people would that be per square kilometer if Tokyo's area is 2,194 square kilometers?"
|
||||||
)
|
)
|
||||||
|
|
||||||
assert (
|
assert "21 million" in result.raw or "37 million" in result.raw, (
|
||||||
"21 million" in result.raw or "37 million" in result.raw
|
"Agent should find Tokyo's population"
|
||||||
), "Agent should find Tokyo's population"
|
)
|
||||||
assert (
|
assert "per square kilometer" in result.raw, (
|
||||||
"per square kilometer" in result.raw
|
"Agent should calculate population density"
|
||||||
), "Agent should calculate population density"
|
)
|
||||||
|
|
||||||
received_events = []
|
received_events = []
|
||||||
|
|
||||||
@@ -316,11 +316,17 @@ def test_sets_parent_flow_when_inside_flow():
|
|||||||
flow.kickoff()
|
flow.kickoff()
|
||||||
assert captured_agent.parent_flow is flow
|
assert captured_agent.parent_flow is flow
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.vcr(filter_headers=["authorization"])
|
@pytest.mark.vcr(filter_headers=["authorization"])
|
||||||
def test_guardrail_is_called_using_string():
|
def test_guardrail_is_called_using_string():
|
||||||
guardrail_events = defaultdict(list)
|
guardrail_events = defaultdict(list)
|
||||||
from crewai.utilities.events import LLMGuardrailCompletedEvent, LLMGuardrailStartedEvent
|
from crewai.utilities.events import (
|
||||||
|
LLMGuardrailCompletedEvent,
|
||||||
|
LLMGuardrailStartedEvent,
|
||||||
|
)
|
||||||
|
|
||||||
with crewai_event_bus.scoped_handlers():
|
with crewai_event_bus.scoped_handlers():
|
||||||
|
|
||||||
@crewai_event_bus.on(LLMGuardrailStartedEvent)
|
@crewai_event_bus.on(LLMGuardrailStartedEvent)
|
||||||
def capture_guardrail_started(source, event):
|
def capture_guardrail_started(source, event):
|
||||||
guardrail_events["started"].append(event)
|
guardrail_events["started"].append(event)
|
||||||
@@ -338,17 +344,26 @@ def test_guardrail_is_called_using_string():
|
|||||||
|
|
||||||
result = agent.kickoff(messages="Top 10 best players in the world?")
|
result = agent.kickoff(messages="Top 10 best players in the world?")
|
||||||
|
|
||||||
assert len(guardrail_events['started']) == 2
|
assert len(guardrail_events["started"]) == 2
|
||||||
assert len(guardrail_events['completed']) == 2
|
assert len(guardrail_events["completed"]) == 2
|
||||||
assert not guardrail_events['completed'][0].success
|
assert not guardrail_events["completed"][0].success
|
||||||
assert guardrail_events['completed'][1].success
|
assert guardrail_events["completed"][1].success
|
||||||
assert "Here are the top 10 best soccer players in the world, focusing exclusively on Brazilian players" in result.raw
|
assert (
|
||||||
|
"Here are the top 10 best soccer players in the world, focusing exclusively on Brazilian players"
|
||||||
|
in result.raw
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.vcr(filter_headers=["authorization"])
|
@pytest.mark.vcr(filter_headers=["authorization"])
|
||||||
def test_guardrail_is_called_using_callable():
|
def test_guardrail_is_called_using_callable():
|
||||||
guardrail_events = defaultdict(list)
|
guardrail_events = defaultdict(list)
|
||||||
from crewai.utilities.events import LLMGuardrailCompletedEvent, LLMGuardrailStartedEvent
|
from crewai.utilities.events import (
|
||||||
|
LLMGuardrailCompletedEvent,
|
||||||
|
LLMGuardrailStartedEvent,
|
||||||
|
)
|
||||||
|
|
||||||
with crewai_event_bus.scoped_handlers():
|
with crewai_event_bus.scoped_handlers():
|
||||||
|
|
||||||
@crewai_event_bus.on(LLMGuardrailStartedEvent)
|
@crewai_event_bus.on(LLMGuardrailStartedEvent)
|
||||||
def capture_guardrail_started(source, event):
|
def capture_guardrail_started(source, event):
|
||||||
guardrail_events["started"].append(event)
|
guardrail_events["started"].append(event)
|
||||||
@@ -366,16 +381,22 @@ def test_guardrail_is_called_using_callable():
|
|||||||
|
|
||||||
result = agent.kickoff(messages="Top 1 best players in the world?")
|
result = agent.kickoff(messages="Top 1 best players in the world?")
|
||||||
|
|
||||||
assert len(guardrail_events['started']) == 1
|
assert len(guardrail_events["started"]) == 1
|
||||||
assert len(guardrail_events['completed']) == 1
|
assert len(guardrail_events["completed"]) == 1
|
||||||
assert guardrail_events['completed'][0].success
|
assert guardrail_events["completed"][0].success
|
||||||
assert "Pelé - Santos, 1958" in result.raw
|
assert "Pelé - Santos, 1958" in result.raw
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.vcr(filter_headers=["authorization"])
|
@pytest.mark.vcr(filter_headers=["authorization"])
|
||||||
def test_guardrail_reached_attempt_limit():
|
def test_guardrail_reached_attempt_limit():
|
||||||
guardrail_events = defaultdict(list)
|
guardrail_events = defaultdict(list)
|
||||||
from crewai.utilities.events import LLMGuardrailCompletedEvent, LLMGuardrailStartedEvent
|
from crewai.utilities.events import (
|
||||||
|
LLMGuardrailCompletedEvent,
|
||||||
|
LLMGuardrailStartedEvent,
|
||||||
|
)
|
||||||
|
|
||||||
with crewai_event_bus.scoped_handlers():
|
with crewai_event_bus.scoped_handlers():
|
||||||
|
|
||||||
@crewai_event_bus.on(LLMGuardrailStartedEvent)
|
@crewai_event_bus.on(LLMGuardrailStartedEvent)
|
||||||
def capture_guardrail_started(source, event):
|
def capture_guardrail_started(source, event):
|
||||||
guardrail_events["started"].append(event)
|
guardrail_events["started"].append(event)
|
||||||
@@ -388,18 +409,23 @@ def test_guardrail_reached_attempt_limit():
|
|||||||
role="Sports Analyst",
|
role="Sports Analyst",
|
||||||
goal="Gather information about the best soccer players",
|
goal="Gather information about the best soccer players",
|
||||||
backstory="""You are an expert at gathering and organizing information. You carefully collect details and present them in a structured way.""",
|
backstory="""You are an expert at gathering and organizing information. You carefully collect details and present them in a structured way.""",
|
||||||
guardrail=lambda output: (False, "You are not allowed to include Brazilian players"),
|
guardrail=lambda output: (
|
||||||
|
False,
|
||||||
|
"You are not allowed to include Brazilian players",
|
||||||
|
),
|
||||||
guardrail_max_retries=2,
|
guardrail_max_retries=2,
|
||||||
)
|
)
|
||||||
|
|
||||||
with pytest.raises(Exception, match="Agent's guardrail failed validation after 2 retries"):
|
with pytest.raises(
|
||||||
|
Exception, match="Agent's guardrail failed validation after 2 retries"
|
||||||
|
):
|
||||||
agent.kickoff(messages="Top 10 best players in the world?")
|
agent.kickoff(messages="Top 10 best players in the world?")
|
||||||
|
|
||||||
assert len(guardrail_events['started']) == 3 # 2 retries + 1 initial call
|
assert len(guardrail_events["started"]) == 3 # 2 retries + 1 initial call
|
||||||
assert len(guardrail_events['completed']) == 3 # 2 retries + 1 initial call
|
assert len(guardrail_events["completed"]) == 3 # 2 retries + 1 initial call
|
||||||
assert not guardrail_events['completed'][0].success
|
assert not guardrail_events["completed"][0].success
|
||||||
assert not guardrail_events['completed'][1].success
|
assert not guardrail_events["completed"][1].success
|
||||||
assert not guardrail_events['completed'][2].success
|
assert not guardrail_events["completed"][2].success
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.vcr(filter_headers=["authorization"])
|
@pytest.mark.vcr(filter_headers=["authorization"])
|
||||||
@@ -412,9 +438,100 @@ def test_agent_output_when_guardrail_returns_base_model():
|
|||||||
role="Sports Analyst",
|
role="Sports Analyst",
|
||||||
goal="Gather information about the best soccer players",
|
goal="Gather information about the best soccer players",
|
||||||
backstory="""You are an expert at gathering and organizing information. You carefully collect details and present them in a structured way.""",
|
backstory="""You are an expert at gathering and organizing information. You carefully collect details and present them in a structured way.""",
|
||||||
guardrail=lambda output: (True, Player(name="Lionel Messi", country="Argentina")),
|
guardrail=lambda output: (
|
||||||
|
True,
|
||||||
|
Player(name="Lionel Messi", country="Argentina"),
|
||||||
|
),
|
||||||
)
|
)
|
||||||
|
|
||||||
result = agent.kickoff(messages="Top 10 best players in the world?")
|
result = agent.kickoff(messages="Top 10 best players in the world?")
|
||||||
|
|
||||||
assert result.pydantic == Player(name="Lionel Messi", country="Argentina")
|
assert result.pydantic == Player(name="Lionel Messi", country="Argentina")
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.vcr(filter_headers=["authorization"])
|
||||||
|
def test_lite_agent_with_custom_llm_and_guardrails():
|
||||||
|
"""Test that CustomLLM (inheriting from BaseLLM) works with guardrails."""
|
||||||
|
from crewai.llms.base_llm import BaseLLM
|
||||||
|
|
||||||
|
class CustomLLM(BaseLLM):
|
||||||
|
def __init__(self, response: str = "Custom response"):
|
||||||
|
super().__init__(model="custom-model")
|
||||||
|
self.response = response
|
||||||
|
self.call_count = 0
|
||||||
|
|
||||||
|
def call(
|
||||||
|
self,
|
||||||
|
messages,
|
||||||
|
tools=None,
|
||||||
|
callbacks=None,
|
||||||
|
available_functions=None,
|
||||||
|
from_task=None,
|
||||||
|
from_agent=None,
|
||||||
|
) -> str:
|
||||||
|
self.call_count += 1
|
||||||
|
|
||||||
|
if "valid" in str(messages) and "feedback" in str(messages):
|
||||||
|
return '{"valid": true, "feedback": null}'
|
||||||
|
|
||||||
|
if "Thought:" in str(messages):
|
||||||
|
return f"Thought: I will analyze soccer players\nFinal Answer: {self.response}"
|
||||||
|
|
||||||
|
return self.response
|
||||||
|
|
||||||
|
def supports_function_calling(self) -> bool:
|
||||||
|
return False
|
||||||
|
|
||||||
|
def supports_stop_words(self) -> bool:
|
||||||
|
return False
|
||||||
|
|
||||||
|
def get_context_window_size(self) -> int:
|
||||||
|
return 4096
|
||||||
|
|
||||||
|
custom_llm = CustomLLM(response="Brazilian soccer players are the best!")
|
||||||
|
|
||||||
|
agent = Agent(
|
||||||
|
role="Sports Analyst",
|
||||||
|
goal="Analyze soccer players",
|
||||||
|
backstory="You analyze soccer players and their performance.",
|
||||||
|
llm=custom_llm,
|
||||||
|
guardrail="Only include Brazilian players",
|
||||||
|
)
|
||||||
|
|
||||||
|
result = agent.kickoff("Tell me about the best soccer players")
|
||||||
|
|
||||||
|
assert custom_llm.call_count > 0
|
||||||
|
assert "Brazilian" in result.raw
|
||||||
|
|
||||||
|
custom_llm2 = CustomLLM(response="Original response")
|
||||||
|
|
||||||
|
def test_guardrail(output):
|
||||||
|
return (True, "Modified by guardrail")
|
||||||
|
|
||||||
|
agent2 = Agent(
|
||||||
|
role="Test Agent",
|
||||||
|
goal="Test goal",
|
||||||
|
backstory="Test backstory",
|
||||||
|
llm=custom_llm2,
|
||||||
|
guardrail=test_guardrail,
|
||||||
|
)
|
||||||
|
|
||||||
|
result2 = agent2.kickoff("Test message")
|
||||||
|
assert result2.raw == "Modified by guardrail"
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.vcr(filter_headers=["authorization"])
|
||||||
|
def test_lite_agent_with_invalid_llm():
|
||||||
|
"""Test that LiteAgent raises proper error when create_llm returns None."""
|
||||||
|
from unittest.mock import patch
|
||||||
|
|
||||||
|
with patch("crewai.lite_agent.create_llm", return_value=None):
|
||||||
|
agent = Agent(
|
||||||
|
role="Test Agent",
|
||||||
|
goal="Test goal",
|
||||||
|
backstory="Test backstory",
|
||||||
|
llm="invalid-model",
|
||||||
|
)
|
||||||
|
with pytest.raises(ValueError) as exc_info:
|
||||||
|
agent.kickoff("Test message")
|
||||||
|
assert "Expected LLM instance of type BaseLLM" in str(exc_info.value)
|
||||||
|
|||||||
Reference in New Issue
Block a user