mirror of
https://github.com/crewAIInc/crewAI.git
synced 2026-01-08 15:48:29 +00:00
wip
This commit is contained in:
@@ -21,9 +21,7 @@ from typing_extensions import Self
|
||||
from crewai.a2a.config import A2AConfig
|
||||
from crewai.agents.agent_builder.base_agent import BaseAgent
|
||||
from crewai.agents.cache.cache_handler import CacheHandler
|
||||
from crewai.agents.crew_agent_executor import CrewAgentExecutor
|
||||
|
||||
# from crewai.agents.crew_agent_executor_flow import CrewAgentExecutorFlow
|
||||
from crewai.agents.crew_agent_executor_flow import CrewAgentExecutorFlow
|
||||
from crewai.events.event_bus import crewai_event_bus
|
||||
from crewai.events.types.knowledge_events import (
|
||||
KnowledgeQueryCompletedEvent,
|
||||
@@ -99,7 +97,7 @@ class Agent(BaseAgent):
|
||||
The agent can also have memory, can operate in verbose mode, and can delegate tasks to other agents.
|
||||
|
||||
Attributes:
|
||||
agent_executor: An instance of the CrewAgentExecutor class.
|
||||
agent_executor: An instance of the CrewAgentExecutor or CrewAgentExecutorFlow class.
|
||||
role: The role of the agent.
|
||||
goal: The objective of the agent.
|
||||
backstory: The backstory of the agent.
|
||||
@@ -184,10 +182,6 @@ class Agent(BaseAgent):
|
||||
default=None,
|
||||
description="Maximum number of reasoning attempts before executing the task. If None, will try until ready.",
|
||||
)
|
||||
# use_flow_executor: bool = Field(
|
||||
# default=False,
|
||||
# description="Use Flow-based executor instead of traditional while-loop executor.",
|
||||
# )
|
||||
embedder: EmbedderConfig | None = Field(
|
||||
default=None,
|
||||
description="Embedder configuration for the agent.",
|
||||
@@ -654,7 +648,7 @@ class Agent(BaseAgent):
|
||||
rpm_limit_fn=rpm_limit_fn,
|
||||
)
|
||||
else:
|
||||
self.agent_executor = CrewAgentExecutor(
|
||||
self.agent_executor = CrewAgentExecutorFlow(
|
||||
llm=self.llm,
|
||||
task=task, # type: ignore[arg-type]
|
||||
agent=self,
|
||||
|
||||
@@ -289,10 +289,8 @@ class CrewAgentExecutorFlow(Flow[AgentReActState], CrewAgentExecutorMixin):
|
||||
color="blue",
|
||||
)
|
||||
try:
|
||||
# RPM enforcement (line 227)
|
||||
enforce_rpm_limit(self.request_within_rpm_limit)
|
||||
|
||||
# LLM call with hooks (lines 229-238)
|
||||
# Note: Hooks are already integrated in get_llm_response utility
|
||||
answer = get_llm_response(
|
||||
llm=self.llm,
|
||||
@@ -304,7 +302,6 @@ class CrewAgentExecutorFlow(Flow[AgentReActState], CrewAgentExecutorMixin):
|
||||
response_model=self.response_model,
|
||||
executor_context=self,
|
||||
)
|
||||
print(f"answer for iteration: {self.state.iterations} is {answer}")
|
||||
|
||||
# Parse response (line 239)
|
||||
formatted_answer = process_llm_response(answer, self.use_stop_words)
|
||||
|
||||
Reference in New Issue
Block a user