Removed unnecessary constants; updated docs to align with bootstrap naming convention

This commit is contained in:
Bobby Lindsey
2025-01-21 16:37:57 -07:00
parent 9b3004911b
commit 6aed0e911e
2 changed files with 2 additions and 18 deletions

View File

@@ -244,7 +244,7 @@ There are three ways to configure LLMs in CrewAI. Choose the method that best fi
# llm: bedrock/meta.llama2-70b-chat-v1
# Amazon SageMaker Models - Enterprise-grade
# llm: sagemaker/<your-sagemaker-endpoint>
# llm: sagemaker/<my-endpoint>
# Mistral Models - Open source alternative
# llm: mistral/mistral-large-latest
@@ -520,7 +520,7 @@ Learn how to get the most out of your LLM configuration:
Example usage:
```python Code
llm = LLM(
model="sagemaker/<your-sagemaker-endpoint>"
model="sagemaker/<my-endpoint>"
)
```
</Accordion>

View File

@@ -63,20 +63,6 @@ ENV_VARS = {
"key_name": "AWS_REGION_NAME",
},
],
"sagemaker": [
{
"prompt": "Enter your AWS Access Key ID (press Enter to skip)",
"key_name": "AWS_ACCESS_KEY_ID",
},
{
"prompt": "Enter your AWS Secret Access Key (press Enter to skip)",
"key_name": "AWS_SECRET_ACCESS_KEY",
},
{
"prompt": "Enter your AWS Region Name (press Enter to skip)",
"key_name": "AWS_REGION_NAME",
},
],
"azure": [
{
"prompt": "Enter your Azure deployment name (must start with 'azure/')",
@@ -123,7 +109,6 @@ PROVIDERS = [
"ollama",
"watson",
"bedrock",
"sagemaker",
"azure",
"cerebras",
"sambanova",
@@ -254,7 +239,6 @@ MODELS = {
"bedrock/mistral.mistral-7b-instruct-v0:2",
"bedrock/mistral.mixtral-8x7b-instruct-v0:1",
],
"sagemaker": ["sagemaker/<your-sagemaker-endpoint>"],
"sambanova": [
"sambanova/Meta-Llama-3.3-70B-Instruct",
"sambanova/QwQ-32B-Preview",