* Performed spell check across the entire documentation
Thank you once again!
* Performed spell check across the most of code base
Folders been checked:
- agents
- cli
- memory
- project
- tasks
- telemetry
- tools
- translations
* Trying to add a max_token for the agents, so they limited by number of tokens.
* Performed spell check across the rest of code base, and enahnced the yaml paraser code a little
* Small change in the main agent doc
* Improve _save_file method to handle both dict and str inputs
- Add check for dict type input
- Use json.dump for dict serialization
- Convert non-dict inputs to string
- Remove type ignore comments
---------
Co-authored-by: João Moura <joaomdmoura@gmail.com>
* feat: add ability to set LLM for AgentPLanner on Crew
* feat: fixes issue on instantiating the ChatOpenAI on the crew
* docs: add docs for the planning_llm new parameter
* docs: change message to ChatOpenAI llm
* feat: add tests
* feat: add crew Testing/evalauting feature
* feat: add docs and add unit test
* feat: improve testing output table
* feat: add tests
* feat: fix type checking issue
* feat: add raise ValueError when testing if output is not the expected
* docs: add docs for Testing
* feat: improve tests and fix some issue
* feat: back to sync
* feat: change opdeai model
* feat: fix test
To solve :
I encountered an error while trying to use the tool. This was the error: DuckDuckGoSearchRun._run() got an unexpected keyword argument 'q'.
Tool duckduckgo_search accepts these inputs: A wrapper around DuckDuckGo Search. Useful for when you need to answer questions about current events. Input should be a search query.
refer : https://github.com/joaomdmoura/crewAI/issues/316
The memory documentation left me with a lot of questions. After I went through the code to find an answer. I added this paragraph to explain what I found. Hope this is helpful.
* feat: add planning feature to crew
* feat: add test to planning handler and change to execute_async method
* docs: add planning parameter to the Core documentation
* docs: add planning docs
* fix: fix type checking issue
* fix: test and logic
* Cleaned up task execution to now have separate paths for async and sync execution. Updating all kickoff functions to return CrewOutput. WIP. Waiting for Joao feedback on async task execution with task_output
* Consistently storing async and sync output for context
* outline tests I need to create going forward
* Major rehaul of TaskOutput and CrewOutput. Updated all tests to work with new change. Need to add in a few final tricky async tests and add a few more to verify output types on TaskOutput and CrewOutput.
* Encountering issues with callback. Need to test on main. WIP
* working on tests. WIP
* WIP. Figuring out disconnect issue.
* Cleaned up logs now that I've isolated the issue to the LLM
* more wip.
* WIP. It looks like usage metrics has always been broken for async
* Update parent crew who is managing for_each loop
* Merge in main to bugfix/kickoff-for-each-usage-metrics
* Clean up code for review
* Add new tests
* Final cleanup. Ready for review.
* Moving copy functionality from Agent to BaseAgent
* Fix renaming issue
* Fix linting errors
* use BaseAgent instead of Agent where applicable
* Fixing missing function. Working on tests.
* WIP. Needing team to review change
* Fixing issues brought about by merge
* WIP: need to fix json encoder
* WIP need to fix encoder
* WIP
* WIP: replay working with async. need to add tests
* Implement major fixes from yesterdays group conversation. Now working on tests.
* The majority of tasks are working now. Need to fix converter class
* Fix final failing test
* Fix linting and type-checker issues
* Add more tests to fully test CrewOutput and TaskOutput changes
* Add in validation for async cannot depend on other async tasks.
* WIP: working replay feat fixing inputs, need tests
* WIP: core logic of seq and heir for executing tasks added into one
* Update validators and tests
* better logic for seq and hier
* replay working for both seq and hier just need tests
* fixed context
* added cli command + code cleanup TODO: need better refactoring
* refactoring for cleaner code
* added better tests
* removed todo comments and fixed some tests
* fix logging now all tests should pass
* cleaner code
* ensure replay is delcared when replaying specific tasks
* ensure hierarchical works
* better typing for stored_outputs and separated task_output_handler
* added better tests
* added replay feature to crew docs
* easier cli command name
* fixing changes
* using sqllite instead of .json file for logging previous task_outputs
* tools fix
* added to docs and fixed tests
* fixed .db
* fixed docs and removed unneeded comments
* separating ltm and replay db
* fixed printing colors
* added how to doc
---------
Co-authored-by: Brandon Hancock <brandon@brandonhancock.io>
- Added detailed steps for training the crew programmatically.
- Clarified the distinction between using the CLI and programmatic approaches.
This update makes it easier for users to understand how to train their crew both through the CLI and programmatically, whether using a UI or API endpoints.
Again Thank you to the author for the great project and the excellent foundation provided!
- Added a "Parameters" column to attribute tables. Improved overall document formatting for enhanced readability and ease of use.
Thank you to the author for the great project and the excellent foundation provided!
* better spacing
* works with llama index
* works on langchain custom just need delegation to work
* cleanup for custom_agent class
* works with different argument expectations for agent_executor
* cleanup for hierarchial process, better agent_executor args handler and added to the crew agent doc page
* removed code examples for langchain + llama index, added to docs instead
* added key output if return is not a str for and added some tests
* added hinting for CustomAgent class
* removed pass as it was not needed
* closer just need to figuire ou agentTools
* running agents - llamaindex and langchain with base agent
* some cleanup on baseAgent
* minimum for agent to run for base class and ensure it works with hierarchical process
* cleanup for original agent to take on BaseAgent class
* Agent takes on langchainagent and cleanup across
* token handling working for usage_metrics to continue working
* installed llama-index, updated docs and added better name
* fixed some type errors
* base agent holds token_process
* heirarchail process uses proper tools and no longer relies on hasattr for token_processes
* removal of test_custom_agent_executions
* this fixes copying agents
* leveraging an executor class for trigger llamaindex agent
* llama index now has ask_human
* executor mixins added
* added output converter base class
* type listed
* cleanup for output conversions and tokenprocess eliminated redundancy
* properly handling tokens
* simplified token calc handling
* original agent with base agent builder structure setup
* better docs
* no more llama-index dep
* cleaner docs
* test fixes
* poetry reverts and better docs
* base_agent_tools set for third party agents
* updated task and test fix
I saw a rendedered whitespace inconsistency in the Tasks docs here:
ed31860071/docs/core-concepts/Tasks.md (L173)
So I set out to patch that up to make it easier to read. I then noticed
there were a few whitespace inconsistencies:
- 2 spaces
- 4 whitespaces
- tabs
It appears that the 4 whitespaces is the prevalent whitesapce usage, so
I overwrote other whitespace usages with that in this commit.
Co-authored-by: Rueben Ramirez <rramirez@ruebens-mbp.tail7c016.ts.net>
Co-authored-by: João Moura <joaomdmoura@gmail.com>
* log_file
feature: added a new parameter for crew that creates a txt file to log agent execution
* unit tests and documentation
unit test if file is created but not what is inside the file
* feature: human input per task
* Update executor.py
* Update executor.py
* Update executor.py
* Update executor.py
* Update executor.py
* feat: change human input for unit testing
added documentation and unit test
* Create test_agent_human_input.yaml
add yaml for test
---------
Co-authored-by: João Moura <joaomdmoura@gmail.com>