Files
crewAI/tests/procedure/test_procedure.py

202 lines
6.6 KiB
Python

from unittest.mock import Mock, PropertyMock
import pytest
from crewai.crew import Crew
from crewai.crews.crew_output import CrewOutput
from crewai.procedure.procedure import Procedure
from crewai.tasks.task_output import TaskOutput
@pytest.fixture
def mock_crew():
crew = Mock(spec=Crew)
task_output = TaskOutput(
description="Test task", raw="Task output", agent="Test Agent"
)
crew_output = CrewOutput(
raw="Test output",
tasks_output=[task_output],
token_usage={"total_tokens": 100, "prompt_tokens": 50, "completion_tokens": 50},
json_dict={"key": "value"}, # Add this line
)
async def async_kickoff(inputs=None):
return crew_output
crew.kickoff.return_value = crew_output
crew.kickoff_async.side_effect = async_kickoff
return crew
def test_procedure_initialization():
"""
Test that a Procedure is correctly initialized with the given crews.
"""
crew1 = Mock(spec=Crew)
crew2 = Mock(spec=Crew)
# Add properties required by validators
type(crew1).verbose = PropertyMock(return_value=True)
type(crew2).verbose = PropertyMock(return_value=True)
type(crew1).output_log_file = PropertyMock(return_value=False)
type(crew2).output_log_file = PropertyMock(return_value=False)
procedure = Procedure(crews=[crew1, crew2])
assert len(procedure.crews) == 2
assert procedure.crews[0] == crew1
assert procedure.crews[1] == crew2
@pytest.mark.asyncio
async def test_procedure_kickoff_single_input(mock_crew):
"""
Test that Procedure.kickoff() correctly processes a single input
and returns the expected CrewOutput.
"""
procedure = Procedure(crews=[mock_crew])
input_data = {"key": "value"}
result = await procedure.kickoff([input_data])
mock_crew.kickoff_async.assert_called_once_with(inputs=input_data)
assert len(result) == 1
assert isinstance(result[0], CrewOutput)
assert result[0].raw == "Test output"
assert len(result[0].tasks_output) == 1
assert result[0].tasks_output[0].raw == "Task output"
assert result[0].token_usage == {
"total_tokens": 100,
"prompt_tokens": 50,
"completion_tokens": 50,
}
@pytest.mark.asyncio
async def test_procedure_kickoff_multiple_inputs(mock_crew):
"""
Test that Procedure.kickoff() correctly processes multiple inputs
and returns the expected CrewOutputs.
"""
procedure = Procedure(crews=[mock_crew, mock_crew])
input_data = [{"key1": "value1"}, {"key2": "value2"}]
result = await procedure.kickoff(input_data)
expected_call_count = 4 # 2 crews x 2 inputs = 4
assert mock_crew.kickoff_async.call_count == expected_call_count
assert len(result) == 2
assert all(isinstance(r, CrewOutput) for r in result)
assert all(len(r.tasks_output) == 1 for r in result)
assert all(
r.token_usage
== {"total_tokens": 100, "prompt_tokens": 50, "completion_tokens": 50}
for r in result
)
@pytest.mark.asyncio
async def test_procedure_chaining():
"""
Test that Procedure correctly chains multiple crews, passing the output
of one crew as input to the next crew in the sequence.
This test verifies:
1. The first crew receives the initial input.
2. The second crew receives the output from the first crew as its input.
3. The final output contains the result from the last crew in the chain.
4. Task outputs and token usage are correctly propagated through the chain.
"""
crew1, crew2 = Mock(spec=Crew), Mock(spec=Crew)
task_output1 = TaskOutput(description="Task 1", raw="Output 1", agent="Agent 1")
task_output2 = TaskOutput(description="Task 2", raw="Final output", agent="Agent 2")
crew_output1 = CrewOutput(
raw="Output 1",
tasks_output=[task_output1],
token_usage={"total_tokens": 100, "prompt_tokens": 50, "completion_tokens": 50},
json_dict={"key1": "value1"},
)
crew_output2 = CrewOutput(
raw="Final output",
tasks_output=[task_output2],
token_usage={"total_tokens": 150, "prompt_tokens": 75, "completion_tokens": 75},
json_dict={"key2": "value2"},
)
async def async_kickoff1(inputs=None):
return crew_output1
async def async_kickoff2(inputs=None):
return crew_output2
crew1.kickoff_async.side_effect = async_kickoff1
crew2.kickoff_async.side_effect = async_kickoff2
procedure = Procedure(crews=[crew1, crew2])
input_data = [{"initial": "data"}]
result = await procedure.kickoff(input_data)
# Check that the first crew received the initial input
crew1.kickoff_async.assert_called_once_with(inputs={"initial": "data"})
# Check that the second crew received the output from the first crew as its input
crew2.kickoff_async.assert_called_once_with(inputs=crew_output1.to_dict())
# Check the final output
assert len(result) == 1
assert isinstance(result[0], CrewOutput)
assert result[0].raw == "Final output"
assert len(result[0].tasks_output) == 1
assert result[0].tasks_output[0].raw == "Final output"
assert result[0].token_usage == {
"total_tokens": 150,
"prompt_tokens": 75,
"completion_tokens": 75,
}
assert result[0].json_dict == {"key2": "value2"}
@pytest.mark.asyncio
async def test_procedure_invalid_input_type():
"""
Test that Procedure.kickoff() raises a TypeError when given an invalid input type.
"""
procedure = Procedure(crews=[Mock(spec=Crew)])
with pytest.raises(TypeError):
await procedure.kickoff("invalid input")
@pytest.mark.asyncio
async def test_procedure_token_usage_aggregation():
"""
Test that Procedure correctly aggregates token usage across multiple crews.
"""
crew1, crew2 = Mock(spec=Crew), Mock(spec=Crew)
crew1.kickoff.return_value = CrewOutput(
raw="Output 1",
tasks_output=[
TaskOutput(description="Task 1", raw="Output 1", agent="Agent 1")
],
token_usage={"total_tokens": 100, "prompt_tokens": 50, "completion_tokens": 50},
)
crew2.kickoff.return_value = CrewOutput(
raw="Output 2",
tasks_output=[
TaskOutput(description="Task 2", raw="Output 2", agent="Agent 2")
],
token_usage={"total_tokens": 150, "prompt_tokens": 75, "completion_tokens": 75},
)
procedure = Procedure([crew1, crew2])
result = await procedure.kickoff([{"initial": "data"}])
assert result[0].token_usage == {
"total_tokens": 250,
"prompt_tokens": 125,
"completion_tokens": 125,
}
assert result[0].token_usage == {
"total_tokens": 250,
"prompt_tokens": 125,
"completion_tokens": 125,
}