fixing type check

This commit is contained in:
Brandon Hancock
2025-03-05 12:42:25 -05:00
parent 2e9945c406
commit aebb4141b6

View File

@@ -343,8 +343,6 @@ class LLM:
chunk_content = delta chunk_content = delta
if chunk_content: if chunk_content:
full_response += chunk_content
print(f"Chunk content: {chunk_content}")
crewai_event_bus.emit( crewai_event_bus.emit(
self, self,
event=LLMStreamChunkEvent(chunk=chunk_content), event=LLMStreamChunkEvent(chunk=chunk_content),
@@ -384,16 +382,7 @@ class LLM:
f"Extracted text from last chunk: {full_response}" f"Extracted text from last chunk: {full_response}"
) )
# --- 6) If still empty, use a default response # --- 6) Check for tool calls in the final response
if not full_response.strip():
logging.warning("Using default response as fallback")
full_response = "I apologize, but I couldn't generate a proper response. Please try again or rephrase your request."
print("LAST CHUNK:", last_chunk)
if hasattr(last_chunk, "usage"):
print("LAST CHUNK USAGE:", last_chunk.usage)
# --- 7) Check for tool calls in the final response
if isinstance(last_chunk, ModelResponse): if isinstance(last_chunk, ModelResponse):
usage_info = getattr(last_chunk, "usage", None) usage_info = getattr(last_chunk, "usage", None)
choices = getattr(last_chunk, "choices", []) choices = getattr(last_chunk, "choices", [])
@@ -408,8 +397,7 @@ class LLM:
if tool_result is not None: if tool_result is not None:
return tool_result return tool_result
# --- 8) Log token usage if available in streaming mode # --- 7) Log token usage if available in streaming mode
# Safely handle callbacks with usage info # Safely handle callbacks with usage info
if callbacks and len(callbacks) > 0: if callbacks and len(callbacks) > 0:
for callback in callbacks: for callback in callbacks:
@@ -417,7 +405,6 @@ class LLM:
usage_info = ( usage_info = (
getattr(last_chunk, "usage", None) if last_chunk else None getattr(last_chunk, "usage", None) if last_chunk else None
) )
print("USAGE INFO", usage_info)
if usage_info: if usage_info:
callback.log_success_event( callback.log_success_event(
kwargs=params, kwargs=params,
@@ -426,7 +413,7 @@ class LLM:
end_time=0, end_time=0,
) )
# --- 9) Emit completion event and return response # --- 8) Emit completion event and return response
self._handle_emit_call_events(full_response, LLMCallType.LLM_CALL) self._handle_emit_call_events(full_response, LLMCallType.LLM_CALL)
return full_response return full_response