From 6aed0e911ec6611dfc1bd343ecf80970774afd3f Mon Sep 17 00:00:00 2001 From: Bobby Lindsey Date: Tue, 21 Jan 2025 16:37:57 -0700 Subject: [PATCH] Removed unnecessary constants; updated docs to align with bootstrap naming convention --- docs/concepts/llms.mdx | 4 ++-- src/crewai/cli/constants.py | 16 ---------------- 2 files changed, 2 insertions(+), 18 deletions(-) diff --git a/docs/concepts/llms.mdx b/docs/concepts/llms.mdx index cf564f2e7..261a1fdd8 100644 --- a/docs/concepts/llms.mdx +++ b/docs/concepts/llms.mdx @@ -244,7 +244,7 @@ There are three ways to configure LLMs in CrewAI. Choose the method that best fi # llm: bedrock/meta.llama2-70b-chat-v1 # Amazon SageMaker Models - Enterprise-grade - # llm: sagemaker/ + # llm: sagemaker/ # Mistral Models - Open source alternative # llm: mistral/mistral-large-latest @@ -520,7 +520,7 @@ Learn how to get the most out of your LLM configuration: Example usage: ```python Code llm = LLM( - model="sagemaker/" + model="sagemaker/" ) ``` diff --git a/src/crewai/cli/constants.py b/src/crewai/cli/constants.py index 1d70d0921..b97b4f208 100644 --- a/src/crewai/cli/constants.py +++ b/src/crewai/cli/constants.py @@ -63,20 +63,6 @@ ENV_VARS = { "key_name": "AWS_REGION_NAME", }, ], - "sagemaker": [ - { - "prompt": "Enter your AWS Access Key ID (press Enter to skip)", - "key_name": "AWS_ACCESS_KEY_ID", - }, - { - "prompt": "Enter your AWS Secret Access Key (press Enter to skip)", - "key_name": "AWS_SECRET_ACCESS_KEY", - }, - { - "prompt": "Enter your AWS Region Name (press Enter to skip)", - "key_name": "AWS_REGION_NAME", - }, - ], "azure": [ { "prompt": "Enter your Azure deployment name (must start with 'azure/')", @@ -123,7 +109,6 @@ PROVIDERS = [ "ollama", "watson", "bedrock", - "sagemaker", "azure", "cerebras", "sambanova", @@ -254,7 +239,6 @@ MODELS = { "bedrock/mistral.mistral-7b-instruct-v0:2", "bedrock/mistral.mixtral-8x7b-instruct-v0:1", ], - "sagemaker": ["sagemaker/"], "sambanova": [ "sambanova/Meta-Llama-3.3-70B-Instruct", "sambanova/QwQ-32B-Preview",