diff --git a/docs/concepts/cli.mdx b/docs/concepts/cli.mdx index ab21232ad..ad64ee938 100644 --- a/docs/concepts/cli.mdx +++ b/docs/concepts/cli.mdx @@ -161,6 +161,7 @@ The CLI will initially prompt for API keys for the following services: * Groq * Anthropic * Google Gemini +* SambaNova When you select a provider, the CLI will prompt you to enter your API key. diff --git a/docs/concepts/llms.mdx b/docs/concepts/llms.mdx index 4e269d71a..851e93085 100644 --- a/docs/concepts/llms.mdx +++ b/docs/concepts/llms.mdx @@ -146,6 +146,19 @@ Here's a detailed breakdown of supported models and their capabilities, you can Groq is known for its fast inference speeds, making it suitable for real-time applications. + + | Model | Context Window | Best For | + |-------|---------------|-----------| + | Llama 3.1 70B/8B | Up to 131,072 tokens | High-performance, large context tasks | + | Llama 3.1 405B | 8,192 tokens | High-performance and output quality | + | Llama 3.2 Series | 8,192 tokens | General-purpose tasks, multimodal | + | Llama 3.3 70B | Up to 131,072 tokens | High-performance and output quality| + | Qwen2 familly | 8,192 tokens | High-performance and output quality | + + + [SambaNova](https://cloud.sambanova.ai/) has several models with fast inference speed at full precision. + + | Provider | Context Window | Key Features | |----------|---------------|--------------| diff --git a/docs/how-to/llm-connections.mdx b/docs/how-to/llm-connections.mdx index a69aa4c86..25509c299 100644 --- a/docs/how-to/llm-connections.mdx +++ b/docs/how-to/llm-connections.mdx @@ -32,6 +32,7 @@ LiteLLM supports a wide range of providers, including but not limited to: - Cloudflare Workers AI - DeepInfra - Groq +- SambaNova - [NVIDIA NIMs](https://docs.api.nvidia.com/nim/reference/models-1) - And many more! diff --git a/src/crewai/cli/constants.py b/src/crewai/cli/constants.py index 2d3c5bdad..d878c6f34 100644 --- a/src/crewai/cli/constants.py +++ b/src/crewai/cli/constants.py @@ -85,6 +85,12 @@ ENV_VARS = { "key_name": "CEREBRAS_API_KEY", }, ], + "sambanova": [ + { + "prompt": "Enter your SambaNovaCloud API key (press Enter to skip)", + "key_name": "SAMBANOVA_API_KEY", + } + ], } @@ -98,6 +104,7 @@ PROVIDERS = [ "bedrock", "azure", "cerebras", + "sambanova", ] MODELS = { @@ -156,6 +163,19 @@ MODELS = { "bedrock/mistral.mistral-7b-instruct-v0:2", "bedrock/mistral.mixtral-8x7b-instruct-v0:1", ], + "sambanova": [ + "sambanova/Meta-Llama-3.3-70B-Instruct", + "sambanova/QwQ-32B-Preview", + "sambanova/Qwen2.5-72B-Instruct", + "sambanova/Qwen2.5-Coder-32B-Instruct", + "sambanova/Meta-Llama-3.1-405B-Instruct", + "sambanova/Meta-Llama-3.1-70B-Instruct", + "sambanova/Meta-Llama-3.1-8B-Instruct", + "sambanova/Llama-3.2-90B-Vision-Instruct", + "sambanova/Llama-3.2-11B-Vision-Instruct", + "sambanova/Meta-Llama-3.2-3B-Instruct", + "sambanova/Meta-Llama-3.2-1B-Instruct", + ], } DEFAULT_LLM_MODEL = "gpt-4o-mini" diff --git a/src/crewai/llm.py b/src/crewai/llm.py index 085a0abdc..74a6bc2c3 100644 --- a/src/crewai/llm.py +++ b/src/crewai/llm.py @@ -76,6 +76,18 @@ LLM_CONTEXT_WINDOW_SIZES = { "mixtral-8x7b-32768": 32768, "llama-3.3-70b-versatile": 128000, "llama-3.3-70b-instruct": 128000, + #sambanova + "Meta-Llama-3.3-70B-Instruct": 131072, + "QwQ-32B-Preview": 8192, + "Qwen2.5-72B-Instruct": 8192, + "Qwen2.5-Coder-32B-Instruct": 8192, + "Meta-Llama-3.1-405B-Instruct": 8192, + "Meta-Llama-3.1-70B-Instruct": 131072, + "Meta-Llama-3.1-8B-Instruct": 131072, + "Llama-3.2-90B-Vision-Instruct": 16384, + "Llama-3.2-11B-Vision-Instruct": 16384, + "Meta-Llama-3.2-3B-Instruct": 4096, + "Meta-Llama-3.2-1B-Instruct": 16384, } DEFAULT_CONTEXT_WINDOW_SIZE = 8192