diff --git a/.changeset/beige-penguins-greet.md b/.changeset/beige-penguins-greet.md new file mode 100644 index 000000000000..22e6868c97d1 --- /dev/null +++ b/.changeset/beige-penguins-greet.md @@ -0,0 +1,5 @@ +--- +'@ai-sdk/deepinfra': patch +--- + +feat (providers/deepinfra): add llama 4 models diff --git a/content/providers/01-ai-sdk-providers/11-deepinfra.mdx b/content/providers/01-ai-sdk-providers/11-deepinfra.mdx index b451445fd84c..23296f0e2b2e 100644 --- a/content/providers/01-ai-sdk-providers/11-deepinfra.mdx +++ b/content/providers/01-ai-sdk-providers/11-deepinfra.mdx @@ -82,30 +82,32 @@ DeepInfra language models can also be used in the `streamText` function (see [AI ## Model Capabilities -| Model | Image Input | Object Generation | Tool Usage | Tool Streaming | -| ---------------------------------------------- | ------------------- | ------------------- | ------------------- | ------------------- | -| `meta-llama/Llama-3.3-70B-Instruct-Turbo` | | | | | -| `meta-llama/Llama-3.3-70B-Instruct` | | | | | -| `meta-llama/Meta-Llama-3.1-405B-Instruct` | | | | | -| `meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo` | | | | | -| `meta-llama/Meta-Llama-3.1-70B-Instruct` | | | | | -| `meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo` | | | | | -| `meta-llama/Meta-Llama-3.1-8B-Instruct` | | | | | -| `meta-llama/Llama-3.2-11B-Vision-Instruct` | | | | | -| `meta-llama/Llama-3.2-90B-Vision-Instruct` | | | | | -| `mistralai/Mixtral-8x7B-Instruct-v0.1` | | | | | -| `deepseek-ai/DeepSeek-V3` | | | | | -| `deepseek-ai/DeepSeek-R1` | | | | | -| `deepseek-ai/DeepSeek-R1-Distill-Llama-70B` | | | | | -| `deepseek-ai/DeepSeek-R1-Turbo` | | | | | -| `nvidia/Llama-3.1-Nemotron-70B-Instruct` | | | | | -| `Qwen/Qwen2-7B-Instruct` | | | | | -| `Qwen/Qwen2.5-72B-Instruct` | | | | | -| `Qwen/Qwen2.5-Coder-32B-Instruct` | | | | | -| `Qwen/QwQ-32B-Preview` | | | | | -| `google/codegemma-7b-it` | | | | | -| `google/gemma-2-9b-it` | | | | | -| `microsoft/WizardLM-2-8x22B` | | | | | +| Model | Image Input | Object Generation | Tool Usage | Tool Streaming | +| --------------------------------------------------- | ------------------- | ------------------- | ------------------- | ------------------- | +| `meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8` | | | | | +| `meta-llama/Llama-4-Scout-17B-16E-Instruct` | | | | | +| `meta-llama/Llama-3.3-70B-Instruct-Turbo` | | | | | +| `meta-llama/Llama-3.3-70B-Instruct` | | | | | +| `meta-llama/Meta-Llama-3.1-405B-Instruct` | | | | | +| `meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo` | | | | | +| `meta-llama/Meta-Llama-3.1-70B-Instruct` | | | | | +| `meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo` | | | | | +| `meta-llama/Meta-Llama-3.1-8B-Instruct` | | | | | +| `meta-llama/Llama-3.2-11B-Vision-Instruct` | | | | | +| `meta-llama/Llama-3.2-90B-Vision-Instruct` | | | | | +| `mistralai/Mixtral-8x7B-Instruct-v0.1` | | | | | +| `deepseek-ai/DeepSeek-V3` | | | | | +| `deepseek-ai/DeepSeek-R1` | | | | | +| `deepseek-ai/DeepSeek-R1-Distill-Llama-70B` | | | | | +| `deepseek-ai/DeepSeek-R1-Turbo` | | | | | +| `nvidia/Llama-3.1-Nemotron-70B-Instruct` | | | | | +| `Qwen/Qwen2-7B-Instruct` | | | | | +| `Qwen/Qwen2.5-72B-Instruct` | | | | | +| `Qwen/Qwen2.5-Coder-32B-Instruct` | | | | | +| `Qwen/QwQ-32B-Preview` | | | | | +| `google/codegemma-7b-it` | | | | | +| `google/gemma-2-9b-it` | | | | | +| `microsoft/WizardLM-2-8x22B` | | | | | The table above lists popular models. Please see the [DeepInfra diff --git a/content/providers/01-ai-sdk-providers/index.mdx b/content/providers/01-ai-sdk-providers/index.mdx index 503a3119e3de..dd0b183414b0 100644 --- a/content/providers/01-ai-sdk-providers/index.mdx +++ b/content/providers/01-ai-sdk-providers/index.mdx @@ -17,51 +17,53 @@ There are also [community providers](./community-providers) that have been creat Not all providers support all AI SDK features. Here's a quick comparison of the capabilities of popular models: -| Provider | Model | Image Input | Object Generation | Tool Usage | Tool Streaming | -| ------------------------------------------------------------------------ | --------------------------------------------- | ------------------- | ------------------- | ------------------- | ------------------- | -| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-2-1212` | | | | | -| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-2-vision-1212` | | | | | -| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-beta` | | | | | -| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-vision-beta` | | | | | -| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4o` | | | | | -| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4o-mini` | | | | | -| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4-turbo` | | | | | -| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4` | | | | | -| [OpenAI](/providers/ai-sdk-providers/openai) | `o1` | | | | | -| [OpenAI](/providers/ai-sdk-providers/openai) | `o1-mini` | | | | | -| [OpenAI](/providers/ai-sdk-providers/openai) | `o1-preview` | | | | | -| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-7-sonnet-20250219` | | | | | -| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-5-sonnet-20241022` | | | | | -| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-5-sonnet-20240620` | | | | | -| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-5-haiku-20241022` | | | | | -| [Groq](/providers/ai-sdk-providers/groq) | `meta-llama/llama-4-scout-17b-16e-instruct` | | | | | -| [Groq](/providers/ai-sdk-providers/groq) | `deepseek-r1-distill-llama-70b` | | | | | -| [Groq](/providers/ai-sdk-providers/groq) | `llama-3.3-70b-versatile` | | | | | -| [Groq](/providers/ai-sdk-providers/groq) | `llama-3.1-8b-instant` | | | | | -| [Groq](/providers/ai-sdk-providers/groq) | `mistral-saba-24b` | | | | | -| [Groq](/providers/ai-sdk-providers/groq) | `qwen-qwq-32b` | | | | | -| [Groq](/providers/ai-sdk-providers/groq) | `mixtral-8x7b-32768` | | | | | -| [Groq](/providers/ai-sdk-providers/groq) | `gemma2-9b-it` | | | | | -| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo` | | | | | -| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `meta-llama/Llama-3.3-70B-Instruct` | | | | | -| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-V3` | | | | | -| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-R1` | | | | | -| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-R1-Distill-Llama-70B` | | | | | -| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-R1-Turbo` | | | | | -| [Mistral](/providers/ai-sdk-providers/mistral) | `pixtral-large-latest` | | | | | -| [Mistral](/providers/ai-sdk-providers/mistral) | `mistral-large-latest` | | | | | -| [Mistral](/providers/ai-sdk-providers/mistral) | `mistral-small-latest` | | | | | -| [Mistral](/providers/ai-sdk-providers/mistral) | `pixtral-12b-2409` | | | | | -| [Google Generative AI](/providers/ai-sdk-providers/google-generative-ai) | `gemini-2.0-flash-exp` | | | | | -| [Google Generative AI](/providers/ai-sdk-providers/google-generative-ai) | `gemini-1.5-flash` | | | | | -| [Google Generative AI](/providers/ai-sdk-providers/google-generative-ai) | `gemini-1.5-pro` | | | | | -| [Google Vertex](/providers/ai-sdk-providers/google-vertex) | `gemini-2.0-flash-exp` | | | | | -| [Google Vertex](/providers/ai-sdk-providers/google-vertex) | `gemini-1.5-flash` | | | | | -| [Google Vertex](/providers/ai-sdk-providers/google-vertex) | `gemini-1.5-pro` | | | | | -| [DeepSeek](/providers/ai-sdk-providers/deepseek) | `deepseek-chat` | | | | | -| [DeepSeek](/providers/ai-sdk-providers/deepseek) | `deepseek-reasoner` | | | | | -| [Cerebras](/providers/ai-sdk-providers/cerebras) | `llama3.1-8b` | | | | | -| [Cerebras](/providers/ai-sdk-providers/cerebras) | `llama3.3-70b` | | | | | +| Provider | Model | Image Input | Object Generation | Tool Usage | Tool Streaming | +| ------------------------------------------------------------------------ | --------------------------------------------------- | ------------------- | ------------------- | ------------------- | ------------------- | +| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-2-1212` | | | | | +| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-2-vision-1212` | | | | | +| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-beta` | | | | | +| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-vision-beta` | | | | | +| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4o` | | | | | +| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4o-mini` | | | | | +| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4-turbo` | | | | | +| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4` | | | | | +| [OpenAI](/providers/ai-sdk-providers/openai) | `o1` | | | | | +| [OpenAI](/providers/ai-sdk-providers/openai) | `o1-mini` | | | | | +| [OpenAI](/providers/ai-sdk-providers/openai) | `o1-preview` | | | | | +| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-7-sonnet-20250219` | | | | | +| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-5-sonnet-20241022` | | | | | +| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-5-sonnet-20240620` | | | | | +| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-5-haiku-20241022` | | | | | +| [Groq](/providers/ai-sdk-providers/groq) | `meta-llama/llama-4-scout-17b-16e-instruct` | | | | | +| [Groq](/providers/ai-sdk-providers/groq) | `deepseek-r1-distill-llama-70b` | | | | | +| [Groq](/providers/ai-sdk-providers/groq) | `llama-3.3-70b-versatile` | | | | | +| [Groq](/providers/ai-sdk-providers/groq) | `llama-3.1-8b-instant` | | | | | +| [Groq](/providers/ai-sdk-providers/groq) | `mistral-saba-24b` | | | | | +| [Groq](/providers/ai-sdk-providers/groq) | `qwen-qwq-32b` | | | | | +| [Groq](/providers/ai-sdk-providers/groq) | `mixtral-8x7b-32768` | | | | | +| [Groq](/providers/ai-sdk-providers/groq) | `gemma2-9b-it` | | | | | +| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8` | | | | | +| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `meta-llama/Llama-4-Scout-17B-16E-Instruct` | | | | | +| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo` | | | | | +| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `meta-llama/Llama-3.3-70B-Instruct` | | | | | +| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-V3` | | | | | +| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-R1` | | | | | +| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-R1-Distill-Llama-70B` | | | | | +| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-R1-Turbo` | | | | | +| [Mistral](/providers/ai-sdk-providers/mistral) | `pixtral-large-latest` | | | | | +| [Mistral](/providers/ai-sdk-providers/mistral) | `mistral-large-latest` | | | | | +| [Mistral](/providers/ai-sdk-providers/mistral) | `mistral-small-latest` | | | | | +| [Mistral](/providers/ai-sdk-providers/mistral) | `pixtral-12b-2409` | | | | | +| [Google Generative AI](/providers/ai-sdk-providers/google-generative-ai) | `gemini-2.0-flash-exp` | | | | | +| [Google Generative AI](/providers/ai-sdk-providers/google-generative-ai) | `gemini-1.5-flash` | | | | | +| [Google Generative AI](/providers/ai-sdk-providers/google-generative-ai) | `gemini-1.5-pro` | | | | | +| [Google Vertex](/providers/ai-sdk-providers/google-vertex) | `gemini-2.0-flash-exp` | | | | | +| [Google Vertex](/providers/ai-sdk-providers/google-vertex) | `gemini-1.5-flash` | | | | | +| [Google Vertex](/providers/ai-sdk-providers/google-vertex) | `gemini-1.5-pro` | | | | | +| [DeepSeek](/providers/ai-sdk-providers/deepseek) | `deepseek-chat` | | | | | +| [DeepSeek](/providers/ai-sdk-providers/deepseek) | `deepseek-reasoner` | | | | | +| [Cerebras](/providers/ai-sdk-providers/cerebras) | `llama3.1-8b` | | | | | +| [Cerebras](/providers/ai-sdk-providers/cerebras) | `llama3.3-70b` | | | | | This table is not exhaustive. Additional models can be found in the provider diff --git a/examples/ai-core/src/e2e/deepinfra.test.ts b/examples/ai-core/src/e2e/deepinfra.test.ts index fc549841dcc3..bc1943a7d61e 100644 --- a/examples/ai-core/src/e2e/deepinfra.test.ts +++ b/examples/ai-core/src/e2e/deepinfra.test.ts @@ -13,6 +13,8 @@ createFeatureTestSuite({ name: 'DeepInfra', models: { languageModels: [ + createChatModel('meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8'), + createChatModel('meta-llama/Llama-4-Scout-17B-16E-Instruct'), createChatModel('deepseek-ai/DeepSeek-V3'), createChatModel('deepseek-ai/DeepSeek-R1'), createChatModel('deepseek-ai/DeepSeek-R1-Distill-Llama-70B'), diff --git a/packages/deepinfra/src/deepinfra-chat-settings.ts b/packages/deepinfra/src/deepinfra-chat-settings.ts index 8694d195e1e3..6c45c9f7dd0d 100644 --- a/packages/deepinfra/src/deepinfra-chat-settings.ts +++ b/packages/deepinfra/src/deepinfra-chat-settings.ts @@ -2,6 +2,8 @@ import { OpenAICompatibleChatSettings } from '@ai-sdk/openai-compatible'; // https://deepinfra.com/models/text-generation export type DeepInfraChatModelId = + | 'meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8' + | 'meta-llama/Llama-4-Scout-17B-16E-Instruct' | 'meta-llama/Llama-3.3-70B-Instruct' | 'meta-llama/Llama-3.3-70B-Instruct-Turbo' | 'meta-llama/Meta-Llama-3.1-70B-Instruct'