* fix: attempt to make embedchain optional
* fix: drop pydantic_settings dependency
* fix: ensure the package is importable without any extra dependency
After making embedchain option many packages were unstalled which caused errors in some tools due to failing import directives
This commit updates tool prompts to explicitly highlight that some tools
can accept both local file paths and remote URLs.
The improved prompts ensure LLMs understand they may pass remote
resources.
* Create tool for generating automations in Studio
This commit creates a tool to use CrewAI Enterprise API to generate
crews using CrewAI Studio.
* Replace CREWAI_BASE_URL with CREWAI_PLUS_URL
* Add missing /crewai_plus in URL
* Add contextual AI tools with async support
* Fix package version issues and update README
* Rename contextual tools to contextualai and update contents
* Update tools init for contextualai tools
* feat: Resolved no module found error for nest_asyncio
* Updated nest_asyncio import
---------
Co-authored-by: QJ <qj@QJs-MacBook-Pro.local>
Co-authored-by: Qile-Jiang <qile.jiang@contextual.ai>
* feat: add InvokeCrewAIAutomationTool for external crew API integration
* feat: add InvokeCrewAIAutomationTool class for executing CrewAI tasks programmatically
* feat: initialize rag
* refactor: using cosine distance metric for chromadb
* feat: use RecursiveCharacterTextSplitter as chunker strategy
* feat: support chucker and loader per data_type
* feat: adding JSON loader
* feat: adding CSVLoader
* feat: adding loader for DOCX files
* feat: add loader for MDX files
* feat: add loader for XML files
* feat: add loader for parser Webpage
* feat: support to load files from an entire directory
* feat: support to auto-load the loaders for additional DataType
* feat: add chuckers for some specific data type
- Each chunker uses separators specific to its content type
* feat: prevent document duplication and centralize content management
- Implement document deduplication logic in RAG
* Check for existing documents by source reference
* Compare doc IDs to detect content changes
* Automatically replace outdated content while preventing duplicates
- Centralize common functionality for better maintainability
* Create SourceContent class to handle URLs, files, and text uniformly
* Extract shared utilities (compute_sha256) to misc.py
* Standardize doc ID generation across all loaders
- Improve RAG system architecture
* All loaders now inherit consistent patterns from centralized BaseLoader
* Better separation of concerns with dedicated content management classes
* Standardized LoaderResult structure across all loader implementations
* chore: split text loaders file
* test: adding missing tests about RAG loaders
* refactor: QOL
* fix: add missing uv syntax on DOCXLoader
* Stagehand tool improvements
This commit significantly improves the StagehandTool reliability and usability when working with CrewAI agents by addressing several critical
issues:
## Key Improvements
### 1. Atomic Action Support
- Added _extract_steps() method to break complex instructions into individual steps
- Added _simplify_instruction() method for intelligent error recovery
- Sequential execution of micro-actions with proper DOM settling between steps
- Prevents token limit issues on complex pages by encouraging scoped actions
### 2. Enhanced Schema Design
- Made instruction field optional to handle navigation-only commands
- Added smart defaults for missing instructions based on command_type
- Improved field descriptions to guide agents toward atomic actions with location context
- Prevents "instruction Field required" validation errors
### 3. Intelligent API Key Management
- Added _get_model_api_key() method with automatic detection based on model type
- Support for OpenAI (GPT), Anthropic (Claude), and Google (Gemini) API keys
- Removes need for manual model API key configuration
### 4. Robust Error Recovery
- Step-by-step execution with individual error handling per atomic action
- Automatic retry with simplified instructions when complex actions fail
- Comprehensive error logging and reporting for debugging
- Graceful degradation instead of complete failure
### 5. Token Management & Performance
- Tool descriptions encourage atomic, scoped actions (e.g., "click search box in header")
- Prevents "prompt too long" errors on complex pages like Wikipedia
- Location-aware instruction patterns for better DOM targeting
- Reduced observe-act cycles through better instruction decomposition
### 6. Enhanced Testing Support
- Comprehensive async mock objects for testing mode
- Proper async/sync compatibility for different execution contexts
- Enhanced resource cleanup and session management
* Update stagehand_tool.py
removeing FixedStagehandTool in favour of StagehandTool
* removed comment
* Cleanup
Revoved unused class
Improved tool description
* feat: add SerperScrapeWebsiteTool for extracting clean content from URLs
* feat: add required SERPER_API_KEY env var validation to SerperScrapeWebsiteTool
* Enhance EnterpriseActionTool with improved schema processing and error handling
- Added methods for sanitizing names and processing schema types, including support for nested models and nullable types.
- Improved error handling during schema creation and processing, with warnings for failures.
- Updated parameter handling in the `_run` method to clean up `kwargs` before sending requests.
- Introduced a detailed description generation for nested schema structures to enhance tool documentation.
* Add tests for EnterpriseActionTool schema conversion and validation
- Introduced a new test class for validating complex nested schemas in EnterpriseActionTool.
- Added tests for schema conversion, optional fields, enum validation, and required nested fields.
- Implemented execution tests to ensure the tool can handle complex validated input correctly.
- Verified model naming conventions and added tests for simpler schemas with basic enum validation.
- Enhanced overall test coverage for the EnterpriseActionTool functionality.
* Update chromadb dependency version in pyproject.toml and uv.lock
- Changed chromadb version from >=0.4.22 to ==0.5.23 in both pyproject.toml and uv.lock to ensure compatibility and stability.
* Update test workflow configuration
- Changed EMBEDCHAIN_DB_URI to point to a temporary test database location.
- Added CHROMA_PERSIST_PATH for specifying the path to the Chroma test database.
- Cleaned up the test run command in the workflow file.
* reverted
* INTPYTHON-580 Design and Implement MongoDBVectorSearchTool
* add implementation
* wip
* wip
* finish tests
* add todo
* refactor to wrap langchain-mongodb
* cleanup
* address review
* Fix usage of EnvVar class
* inline code
* lint
* lint
* fix usage of SearchIndexModel
* Refactor: Update EnvVar import path and remove unused tests.utils module
- Changed import of EnvVar from tests.utils to crewai.tools in multiple files.
- Updated README.md for MongoDB vector search tool with additional context.
- Modified subprocess command in vector_search.py for package installation.
- Cleaned up test_generate_tool_specs.py to improve mock patching syntax.
- Deleted unused tests/utils.py file.
* update the crewai dep and the lockfile
* chore: update package versions and dependencies in uv.lock
- Removed `auth0-python` package.
- Updated `crewai` version to 0.140.0 and adjusted its dependencies.
- Changed `json-repair` version to 0.25.2.
- Updated `litellm` version to 1.72.6.
- Modified dependency markers for several packages to improve compatibility with Python versions.
* refactor: improve MongoDB vector search tool with enhanced error handling and new dimensions field
- Added logging for error handling in the _run method and during client cleanup.
- Introduced a new 'dimensions' field in the MongoDBVectorSearchConfig for embedding vector size.
- Refactored the _run method to return JSON formatted results and handle exceptions gracefully.
- Cleaned up import statements and improved code readability.
* address review
* update tests
* debug
* fix test
* fix test
* fix test
* support azure openai
---------
Co-authored-by: lorenzejay <lorenzejaytech@gmail.com>
* - Added CouchbaseFTSVectorStore as a CrewAI tool.
- Wrote a README to setup the tool.
- Wrote test cases.
- Added Couchbase as an optional dependency in the project.
* Fixed naming in some places. Added docstrings. Added instructions on how to create a vector search index.
* Fixed pyproject.toml
* error handling and response format
- Removed unnecessary ImportError for missing 'couchbase' package.
- Changed response format from a concatenated string to a JSON array for search results.
- Updated error handling to return error messages instead of raising exceptions in certain cases.
- Adjusted tests to reflect changes in response format and error handling.
* Update dependencies in pyproject.toml and uv.lock
- Changed pydantic version from 2.6.1 to 2.10.6 in both pyproject.toml and uv.lock.
- Updated crewai-tools version from 0.42.2 to 0.42.3 in uv.lock.
- Adjusted pydantic-core version from 2.33.1 to 2.27.2 in uv.lock, reflecting the new pydantic version.
* Removed restrictive pydantic version and updated uv.lock
* synced lockfile
* regenerated lockfile
* updated lockfile
* regenerated lockfile
* Update tool specifications for
* Fix test cases
---------
Co-authored-by: AayushTyagi1 <tyagiaayush5@gmail.com>
Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
* refactor: enhance schema handling in EnterpriseActionTool
- Extracted schema property and required field extraction into separate methods for better readability and maintainability.
- Introduced methods to analyze field types and create Pydantic field definitions based on nullability and requirement status.
- Updated the _run method to handle required nullable fields, ensuring they are set to None if not provided in kwargs.
* refactor: streamline nullable field handling in EnterpriseActionTool
- Removed commented-out code related to handling required nullable fields for clarity.
- Simplified the logic in the _run method to focus on processing parameters without unnecessary comments.
- Added TYPE_CHECKING imports for FirecrawlApp to enhance type safety.
- Updated configuration keys in FirecrawlCrawlWebsiteTool and FirecrawlScrapeWebsiteTool to camelCase for consistency.
- Introduced error handling in the _run methods of both tools to ensure FirecrawlApp is properly initialized before usage.
- Adjusted parameters passed to crawl_url and scrape_url methods to use 'params' instead of unpacking the config dictionary directly.
* feat: add support for parsing actions list from environment variables
This commit introduces a new function, _parse_actions_list, to handle the parsing of a string representation of a list of tool names from environment variables. The CrewaiEnterpriseTools now utilizes this function to filter tools based on the parsed actions list, enhancing flexibility in tool selection. Additionally, a new test case is added to verify the correct usage of the environment actions list.
* test: simplify environment actions list test setup
This commit refactors the test for CrewaiEnterpriseTools to streamline the setup of environment variables. The environment token and actions list are now set in a single patch.dict call, improving readability and reducing redundancy in the test code.
* refactor: remove token validation from EnterpriseActionKitToolAdapter and CrewaiEnterpriseTools
This commit simplifies the initialization of the EnterpriseActionKitToolAdapter and CrewaiEnterpriseTools by removing the explicit validation for the enterprise action token. The token can now be set to None without raising an error, allowing for more flexible usage.
* added loggers for monitoring
* fixed typo
* fix: enhance token handling in EnterpriseActionKitToolAdapter and CrewaiEnterpriseTools
This commit improves the handling of the enterprise action token by allowing it to be fetched from environment variables if not provided. It adds checks to ensure the token is set before making API requests, enhancing robustness and flexibility.
* removed redundancy
* test: add new test for environment token fallback in CrewaiEnterpriseTools
This update introduces a new test case to verify that the environment token is used when no token is provided during the initialization of CrewaiEnterpriseTools. Additionally, minor formatting adjustments were made to existing assertions for consistency.
* test: update environment token test to clear environment variables
This change modifies the test for CrewaiEnterpriseTools to ensure that the environment variables are cleared before setting the test token. This ensures a clean test environment and prevents potential interference from other tests.
* drop redundancy
* feat: support to complex filter on ToolCollection
* refactor: use proper tool collection methot to filter tool in CrewAiEnterpriseTools
* feat: allow to filter available MCP tools
This change allows accessing tools by name (tools["tool_name"]) in addition to
index (tools[0]), making it more intuitive and convenient to work with multiple
tools without needing to remember their position in the list
* feat: add explictly package_dependencies in the Tools
* feat: collect package_dependencies from Tool to add in tool.specs.json
* feat: add default value in run_params Tool' specs
* fix: support get boolean values
This commit also refactor test to make easier define newest attributes into a Tool
* refactor: remove token validation from EnterpriseActionKitToolAdapter and CrewaiEnterpriseTools
This commit simplifies the initialization of the EnterpriseActionKitToolAdapter and CrewaiEnterpriseTools by removing the explicit validation for the enterprise action token. The token can now be set to None without raising an error, allowing for more flexible usage.
* added loggers for monitoring
* fixed typo