mirror of
https://github.com/crewAIInc/crewAI.git
synced 2026-01-04 05:38:33 +00:00
* Add in missing triple quote and execution time to resume agent functionality * Fixing broken kwargs and other issues causing our tests to fail
412 lines
16 KiB
Python
412 lines
16 KiB
Python
import os
|
|
from inspect import signature
|
|
from typing import Any, List, Optional, Tuple
|
|
|
|
from langchain.agents.agent import RunnableAgent
|
|
from langchain.agents.tools import BaseTool
|
|
from langchain.agents.tools import tool as LangChainTool
|
|
from langchain_core.agents import AgentAction
|
|
from langchain_core.callbacks import BaseCallbackHandler
|
|
from langchain_openai import ChatOpenAI
|
|
from pydantic import Field, InstanceOf, PrivateAttr, model_validator
|
|
|
|
from crewai.agents import CacheHandler, CrewAgentExecutor, CrewAgentParser
|
|
from crewai.agents.agent_builder.base_agent import BaseAgent
|
|
from crewai.memory.contextual.contextual_memory import ContextualMemory
|
|
from crewai.tools.agent_tools import AgentTools
|
|
from crewai.utilities import Converter, Prompts
|
|
from crewai.utilities.constants import TRAINED_AGENTS_DATA_FILE, TRAINING_DATA_FILE
|
|
from crewai.utilities.token_counter_callback import TokenCalcHandler
|
|
from crewai.utilities.training_handler import CrewTrainingHandler
|
|
|
|
agentops = None
|
|
try:
|
|
import agentops # type: ignore # Name "agentops" already defined on line 21
|
|
from agentops import track_agent
|
|
except ImportError:
|
|
|
|
def track_agent():
|
|
def noop(f):
|
|
return f
|
|
|
|
return noop
|
|
|
|
|
|
@track_agent()
|
|
class Agent(BaseAgent):
|
|
"""Represents an agent in a system.
|
|
|
|
Each agent has a role, a goal, a backstory, and an optional language model (llm).
|
|
The agent can also have memory, can operate in verbose mode, and can delegate tasks to other agents.
|
|
|
|
Attributes:
|
|
agent_executor: An instance of the CrewAgentExecutor class.
|
|
role: The role of the agent.
|
|
goal: The objective of the agent.
|
|
backstory: The backstory of the agent.
|
|
config: Dict representation of agent configuration.
|
|
llm: The language model that will run the agent.
|
|
function_calling_llm: The language model that will handle the tool calling for this agent, it overrides the crew function_calling_llm.
|
|
max_iter: Maximum number of iterations for an agent to execute a task.
|
|
memory: Whether the agent should have memory or not.
|
|
max_rpm: Maximum number of requests per minute for the agent execution to be respected.
|
|
verbose: Whether the agent execution should be in verbose mode.
|
|
allow_delegation: Whether the agent is allowed to delegate tasks to other agents.
|
|
tools: Tools at agents disposal
|
|
step_callback: Callback to be executed after each step of the agent execution.
|
|
callbacks: A list of callback functions from the langchain library that are triggered during the agent's execution process
|
|
"""
|
|
|
|
_times_executed: int = PrivateAttr(default=0)
|
|
max_execution_time: Optional[int] = Field(
|
|
default=None,
|
|
description="Maximum execution time for an agent to execute a task",
|
|
)
|
|
agent_ops_agent_name: str = None # type: ignore # Incompatible types in assignment (expression has type "None", variable has type "str")
|
|
agent_ops_agent_id: str = None # type: ignore # Incompatible types in assignment (expression has type "None", variable has type "str")
|
|
cache_handler: InstanceOf[CacheHandler] = Field(
|
|
default=None, description="An instance of the CacheHandler class."
|
|
)
|
|
step_callback: Optional[Any] = Field(
|
|
default=None,
|
|
description="Callback to be executed after each step of the agent execution.",
|
|
)
|
|
llm: Any = Field(
|
|
default_factory=lambda: ChatOpenAI(
|
|
model=os.environ.get("OPENAI_MODEL_NAME", "gpt-4o")
|
|
),
|
|
description="Language model that will run the agent.",
|
|
)
|
|
function_calling_llm: Optional[Any] = Field(
|
|
description="Language model that will run the agent.", default=None
|
|
)
|
|
callbacks: Optional[List[InstanceOf[BaseCallbackHandler]]] = Field(
|
|
default=None, description="Callback to be executed"
|
|
)
|
|
system_template: Optional[str] = Field(
|
|
default=None, description="System format for the agent."
|
|
)
|
|
prompt_template: Optional[str] = Field(
|
|
default=None, description="Prompt format for the agent."
|
|
)
|
|
response_template: Optional[str] = Field(
|
|
default=None, description="Response format for the agent."
|
|
)
|
|
tools_results: Optional[List[Any]] = Field(
|
|
default=[], description="Results of the tools used by the agent."
|
|
)
|
|
allow_code_execution: Optional[bool] = Field(
|
|
default=False, description="Enable code execution for the agent."
|
|
)
|
|
max_retry_limit: int = Field(
|
|
default=2,
|
|
description="Maximum number of retries for an agent to execute a task when an error occurs.",
|
|
)
|
|
|
|
def __init__(__pydantic_self__, **data):
|
|
config = data.pop("config", {})
|
|
super().__init__(**config, **data)
|
|
__pydantic_self__.agent_ops_agent_name = __pydantic_self__.role
|
|
|
|
@model_validator(mode="after")
|
|
def set_agent_executor(self) -> "Agent":
|
|
"""Ensure agent executor and token process are set."""
|
|
if hasattr(self.llm, "model_name"):
|
|
token_handler = TokenCalcHandler(self.llm.model_name, self._token_process)
|
|
|
|
# Ensure self.llm.callbacks is a list
|
|
if not isinstance(self.llm.callbacks, list):
|
|
self.llm.callbacks = []
|
|
|
|
# Check if an instance of TokenCalcHandler already exists in the list
|
|
if not any(
|
|
isinstance(handler, TokenCalcHandler) for handler in self.llm.callbacks
|
|
):
|
|
self.llm.callbacks.append(token_handler)
|
|
|
|
if agentops and not any(
|
|
isinstance(handler, agentops.LangchainCallbackHandler)
|
|
for handler in self.llm.callbacks
|
|
):
|
|
agentops.stop_instrumenting()
|
|
self.llm.callbacks.append(agentops.LangchainCallbackHandler())
|
|
|
|
if not self.agent_executor:
|
|
if not self.cache_handler:
|
|
self.cache_handler = CacheHandler()
|
|
self.set_cache_handler(self.cache_handler)
|
|
return self
|
|
|
|
def execute_task(
|
|
self,
|
|
task: Any,
|
|
context: Optional[str] = None,
|
|
tools: Optional[List[Any]] = None,
|
|
) -> str:
|
|
"""Execute a task with the agent.
|
|
|
|
Args:
|
|
task: Task to execute.
|
|
context: Context to execute the task in.
|
|
tools: Tools to use for the task.
|
|
|
|
Returns:
|
|
Output of the agent
|
|
"""
|
|
if self.tools_handler:
|
|
self.tools_handler.last_used_tool = {} # type: ignore # Incompatible types in assignment (expression has type "dict[Never, Never]", variable has type "ToolCalling")
|
|
|
|
task_prompt = task.prompt()
|
|
|
|
if context:
|
|
task_prompt = self.i18n.slice("task_with_context").format(
|
|
task=task_prompt, context=context
|
|
)
|
|
|
|
if self.crew and self.crew.memory:
|
|
contextual_memory = ContextualMemory(
|
|
self.crew._short_term_memory,
|
|
self.crew._long_term_memory,
|
|
self.crew._entity_memory,
|
|
)
|
|
memory = contextual_memory.build_context_for_task(task, context)
|
|
if memory.strip() != "":
|
|
task_prompt += self.i18n.slice("memory").format(memory=memory)
|
|
|
|
tools = tools or self.tools or []
|
|
parsed_tools = self._parse_tools(tools)
|
|
self.create_agent_executor(tools=tools)
|
|
self.agent_executor.tools = parsed_tools
|
|
self.agent_executor.task = task
|
|
|
|
self.agent_executor.tools_description = self._render_text_description_and_args(
|
|
parsed_tools
|
|
)
|
|
self.agent_executor.tools_names = self.__tools_names(parsed_tools)
|
|
|
|
if self.crew and self.crew._train:
|
|
task_prompt = self._training_handler(task_prompt=task_prompt)
|
|
else:
|
|
task_prompt = self._use_trained_data(task_prompt=task_prompt)
|
|
|
|
try:
|
|
result = self.agent_executor.invoke(
|
|
{
|
|
"input": task_prompt,
|
|
"tool_names": self.agent_executor.tools_names,
|
|
"tools": self.agent_executor.tools_description,
|
|
}
|
|
)["output"]
|
|
except Exception as e:
|
|
self._times_executed += 1
|
|
if self._times_executed > self.max_retry_limit:
|
|
raise e
|
|
result = self.execute_task(task, context, tools)
|
|
|
|
if self.max_rpm:
|
|
self._rpm_controller.stop_rpm_counter()
|
|
|
|
# If there was any tool in self.tools_results that had result_as_answer
|
|
# set to True, return the results of the last tool that had
|
|
# result_as_answer set to True
|
|
for tool_result in self.tools_results: # type: ignore # Item "None" of "list[Any] | None" has no attribute "__iter__" (not iterable)
|
|
if tool_result.get("result_as_answer", False):
|
|
result = tool_result["result"]
|
|
|
|
return result
|
|
|
|
def format_log_to_str(
|
|
self,
|
|
intermediate_steps: List[Tuple[AgentAction, str]],
|
|
observation_prefix: str = "Observation: ",
|
|
llm_prefix: str = "",
|
|
) -> str:
|
|
"""Construct the scratchpad that lets the agent continue its thought process."""
|
|
thoughts = ""
|
|
for action, observation in intermediate_steps:
|
|
thoughts += action.log
|
|
thoughts += f"\n{observation_prefix}{observation}\n{llm_prefix}"
|
|
return thoughts
|
|
|
|
def create_agent_executor(self, tools=None) -> None:
|
|
"""Create an agent executor for the agent.
|
|
|
|
Returns:
|
|
An instance of the CrewAgentExecutor class.
|
|
"""
|
|
tools = tools or self.tools or []
|
|
|
|
agent_args = {
|
|
"input": lambda x: x["input"],
|
|
"tools": lambda x: x["tools"],
|
|
"tool_names": lambda x: x["tool_names"],
|
|
"agent_scratchpad": lambda x: self.format_log_to_str(
|
|
x["intermediate_steps"]
|
|
),
|
|
}
|
|
|
|
executor_args = {
|
|
"llm": self.llm,
|
|
"i18n": self.i18n,
|
|
"crew": self.crew,
|
|
"crew_agent": self,
|
|
"tools": self._parse_tools(tools),
|
|
"verbose": self.verbose,
|
|
"original_tools": tools,
|
|
"handle_parsing_errors": True,
|
|
"max_iterations": self.max_iter,
|
|
"max_execution_time": self.max_execution_time,
|
|
"step_callback": self.step_callback,
|
|
"tools_handler": self.tools_handler,
|
|
"function_calling_llm": self.function_calling_llm,
|
|
"callbacks": self.callbacks,
|
|
"max_tokens": self.max_tokens,
|
|
}
|
|
|
|
if self._rpm_controller:
|
|
executor_args["request_within_rpm_limit"] = (
|
|
self._rpm_controller.check_or_wait
|
|
)
|
|
|
|
prompt = Prompts(
|
|
i18n=self.i18n,
|
|
tools=tools,
|
|
system_template=self.system_template,
|
|
prompt_template=self.prompt_template,
|
|
response_template=self.response_template,
|
|
).task_execution()
|
|
|
|
execution_prompt = prompt.partial(
|
|
goal=self.goal,
|
|
role=self.role,
|
|
backstory=self.backstory,
|
|
)
|
|
|
|
stop_words = [self.i18n.slice("observation")]
|
|
|
|
if self.response_template:
|
|
stop_words.append(
|
|
self.response_template.split("{{ .Response }}")[1].strip()
|
|
)
|
|
|
|
bind = self.llm.bind(stop=stop_words)
|
|
|
|
inner_agent = agent_args | execution_prompt | bind | CrewAgentParser(agent=self)
|
|
self.agent_executor = CrewAgentExecutor(
|
|
agent=RunnableAgent(runnable=inner_agent), **executor_args
|
|
)
|
|
|
|
def get_delegation_tools(self, agents: List[BaseAgent]):
|
|
agent_tools = AgentTools(agents=agents)
|
|
tools = agent_tools.tools()
|
|
return tools
|
|
|
|
def get_code_execution_tools(self):
|
|
try:
|
|
from crewai_tools import CodeInterpreterTool
|
|
|
|
return [CodeInterpreterTool()]
|
|
except ModuleNotFoundError:
|
|
self._logger.log(
|
|
"info", "Coding tools not available. Install crewai_tools. "
|
|
)
|
|
|
|
def get_output_converter(self, llm, text, model, instructions):
|
|
return Converter(llm=llm, text=text, model=model, instructions=instructions)
|
|
|
|
def _parse_tools(self, tools: List[Any]) -> List[LangChainTool]: # type: ignore # Function "langchain_core.tools.tool" is not valid as a type
|
|
"""Parse tools to be used for the task."""
|
|
tools_list = []
|
|
try:
|
|
# tentatively try to import from crewai_tools import BaseTool as CrewAITool
|
|
from crewai_tools import BaseTool as CrewAITool
|
|
|
|
for tool in tools:
|
|
if isinstance(tool, CrewAITool):
|
|
tools_list.append(tool.to_langchain())
|
|
else:
|
|
tools_list.append(tool)
|
|
except ModuleNotFoundError:
|
|
tools_list = []
|
|
for tool in tools:
|
|
tools_list.append(tool)
|
|
|
|
return tools_list
|
|
|
|
def _training_handler(self, task_prompt: str) -> str:
|
|
"""Handle training data for the agent task prompt to improve output on Training."""
|
|
if data := CrewTrainingHandler(TRAINING_DATA_FILE).load():
|
|
agent_id = str(self.id)
|
|
|
|
if data.get(agent_id):
|
|
human_feedbacks = [
|
|
i["human_feedback"] for i in data.get(agent_id, {}).values()
|
|
]
|
|
task_prompt += "You MUST follow these feedbacks: \n " + "\n - ".join(
|
|
human_feedbacks
|
|
)
|
|
|
|
return task_prompt
|
|
|
|
def _use_trained_data(self, task_prompt: str) -> str:
|
|
"""Use trained data for the agent task prompt to improve output."""
|
|
if data := CrewTrainingHandler(TRAINED_AGENTS_DATA_FILE).load():
|
|
if trained_data_output := data.get(self.role):
|
|
task_prompt += "You MUST follow these feedbacks: \n " + "\n - ".join(
|
|
trained_data_output["suggestions"]
|
|
)
|
|
return task_prompt
|
|
|
|
def _render_text_description(self, tools: List[BaseTool]) -> str:
|
|
"""Render the tool name and description in plain text.
|
|
|
|
Output will be in the format of:
|
|
|
|
.. code-block:: markdown
|
|
|
|
search: This tool is used for search
|
|
calculator: This tool is used for math
|
|
"""
|
|
description = "\n".join(
|
|
[
|
|
f"Tool name: {tool.name}\nTool description:\n{tool.description}"
|
|
for tool in tools
|
|
]
|
|
)
|
|
|
|
return description
|
|
|
|
def _render_text_description_and_args(self, tools: List[BaseTool]) -> str:
|
|
"""Render the tool name, description, and args in plain text.
|
|
|
|
Output will be in the format of:
|
|
|
|
.. code-block:: markdown
|
|
|
|
search: This tool is used for search, args: {"query": {"type": "string"}}
|
|
calculator: This tool is used for math, \
|
|
args: {"expression": {"type": "string"}}
|
|
"""
|
|
tool_strings = []
|
|
for tool in tools:
|
|
args_schema = str(tool.args)
|
|
if hasattr(tool, "func") and tool.func:
|
|
sig = signature(tool.func)
|
|
description = (
|
|
f"Tool Name: {tool.name}{sig}\nTool Description: {tool.description}"
|
|
)
|
|
else:
|
|
description = (
|
|
f"Tool Name: {tool.name}\nTool Description: {tool.description}"
|
|
)
|
|
tool_strings.append(f"{description}\nTool Arguments: {args_schema}")
|
|
|
|
return "\n".join(tool_strings)
|
|
|
|
@staticmethod
|
|
def __tools_names(tools) -> str:
|
|
return ", ".join([t.name for t in tools])
|
|
|
|
def __repr__(self):
|
|
return f"Agent(role={self.role}, goal={self.goal}, backstory={self.backstory})"
|