fix: Resolve all remaining lint issues (S101, RUF005, N806)

- Replace remaining assert statements with conditional checks
- Fix list concatenation to use iterable unpacking
- Change variable names from UPPER_CASE to lower_case
- All lint checks now pass locally

Co-Authored-By: João <joao@crewai.com>
This commit is contained in:
Devin AI
2025-09-18 20:46:07 +00:00
parent d0e26f37e5
commit 9a0b3e881d

View File

@@ -600,7 +600,7 @@ class LLM(BaseLLM):
full_response += chunk_content full_response += chunk_content
# Emit the chunk event # Emit the chunk event
assert hasattr(crewai_event_bus, "emit") if hasattr(crewai_event_bus, "emit"):
crewai_event_bus.emit( crewai_event_bus.emit(
self, self,
event=LLMStreamChunkEvent( event=LLMStreamChunkEvent(
@@ -755,7 +755,7 @@ class LLM(BaseLLM):
return full_response return full_response
# Emit failed event and re-raise the exception # Emit failed event and re-raise the exception
assert hasattr(crewai_event_bus, "emit") if hasattr(crewai_event_bus, "emit"):
crewai_event_bus.emit( crewai_event_bus.emit(
self, self,
event=LLMCallFailedEvent( event=LLMCallFailedEvent(
@@ -975,7 +975,7 @@ class LLM(BaseLLM):
fn = available_functions[function_name] fn = available_functions[function_name]
# --- 3.2) Execute function # --- 3.2) Execute function
assert hasattr(crewai_event_bus, "emit") if hasattr(crewai_event_bus, "emit"):
started_at = datetime.now() started_at = datetime.now()
crewai_event_bus.emit( crewai_event_bus.emit(
self, self,
@@ -1015,7 +1015,7 @@ class LLM(BaseLLM):
function_name, lambda: None function_name, lambda: None
) # Ensure fn is always a callable ) # Ensure fn is always a callable
logging.error(f"Error executing function '{function_name}': {e}") logging.error(f"Error executing function '{function_name}': {e}")
assert hasattr(crewai_event_bus, "emit") if hasattr(crewai_event_bus, "emit"):
crewai_event_bus.emit( crewai_event_bus.emit(
self, self,
event=LLMCallFailedEvent(error=f"Tool execution error: {e!s}"), event=LLMCallFailedEvent(error=f"Tool execution error: {e!s}"),
@@ -1067,7 +1067,7 @@ class LLM(BaseLLM):
LLMContextLengthExceededException: If input exceeds model's context limit LLMContextLengthExceededException: If input exceeds model's context limit
""" """
# --- 1) Emit call started event # --- 1) Emit call started event
assert hasattr(crewai_event_bus, "emit") if hasattr(crewai_event_bus, "emit"):
crewai_event_bus.emit( crewai_event_bus.emit(
self, self,
event=LLMCallStartedEvent( event=LLMCallStartedEvent(
@@ -1140,7 +1140,7 @@ class LLM(BaseLLM):
from_agent=from_agent, from_agent=from_agent,
) )
assert hasattr(crewai_event_bus, "emit") if hasattr(crewai_event_bus, "emit"):
crewai_event_bus.emit( crewai_event_bus.emit(
self, self,
event=LLMCallFailedEvent( event=LLMCallFailedEvent(
@@ -1166,7 +1166,7 @@ class LLM(BaseLLM):
from_agent: Optional agent object from_agent: Optional agent object
messages: Optional messages object messages: Optional messages object
""" """
assert hasattr(crewai_event_bus, "emit") if hasattr(crewai_event_bus, "emit"):
crewai_event_bus.emit( crewai_event_bus.emit(
self, self,
event=LLMCallCompletedEvent( event=LLMCallCompletedEvent(
@@ -1222,7 +1222,7 @@ class LLM(BaseLLM):
if "mistral" in self.model.lower(): if "mistral" in self.model.lower():
# Check if the last message has a role of 'assistant' # Check if the last message has a role of 'assistant'
if messages and messages[-1]["role"] == "assistant": if messages and messages[-1]["role"] == "assistant":
return messages + [{"role": "user", "content": "Please continue."}] return [*messages, {"role": "user", "content": "Please continue."}]
return messages return messages
# TODO: Remove this code after merging PR https://github.com/BerriAI/litellm/pull/10917 # TODO: Remove this code after merging PR https://github.com/BerriAI/litellm/pull/10917
@@ -1232,7 +1232,7 @@ class LLM(BaseLLM):
and messages and messages
and messages[-1]["role"] == "assistant" and messages[-1]["role"] == "assistant"
): ):
return messages + [{"role": "user", "content": ""}] return [*messages, {"role": "user", "content": ""}]
# Handle Anthropic models # Handle Anthropic models
if self.is_anthropic: if self.is_anthropic:
@@ -1306,14 +1306,14 @@ class LLM(BaseLLM):
if self.context_window_size != 0: if self.context_window_size != 0:
return self.context_window_size return self.context_window_size
MIN_CONTEXT = 1024 min_context = 1024
MAX_CONTEXT = 2097152 # Current max from gemini-1.5-pro max_context = 2097152 # Current max from gemini-1.5-pro
# Validate all context window sizes # Validate all context window sizes
for key, value in LLM_CONTEXT_WINDOW_SIZES.items(): for key, value in LLM_CONTEXT_WINDOW_SIZES.items():
if value < MIN_CONTEXT or value > MAX_CONTEXT: if value < min_context or value > max_context:
raise ValueError( raise ValueError(
f"Context window for {key} must be between {MIN_CONTEXT} and {MAX_CONTEXT}" f"Context window for {key} must be between {min_context} and {max_context}"
) )
self.context_window_size = int( self.context_window_size = int(