diff --git a/.changeset/beige-penguins-greet.md b/.changeset/beige-penguins-greet.md
new file mode 100644
index 000000000000..22e6868c97d1
--- /dev/null
+++ b/.changeset/beige-penguins-greet.md
@@ -0,0 +1,5 @@
+---
+'@ai-sdk/deepinfra': patch
+---
+
+feat (providers/deepinfra): add llama 4 models
diff --git a/content/providers/01-ai-sdk-providers/11-deepinfra.mdx b/content/providers/01-ai-sdk-providers/11-deepinfra.mdx
index b451445fd84c..23296f0e2b2e 100644
--- a/content/providers/01-ai-sdk-providers/11-deepinfra.mdx
+++ b/content/providers/01-ai-sdk-providers/11-deepinfra.mdx
@@ -82,30 +82,32 @@ DeepInfra language models can also be used in the `streamText` function (see [AI
## Model Capabilities
-| Model | Image Input | Object Generation | Tool Usage | Tool Streaming |
-| ---------------------------------------------- | ------------------- | ------------------- | ------------------- | ------------------- |
-| `meta-llama/Llama-3.3-70B-Instruct-Turbo` | | | | |
-| `meta-llama/Llama-3.3-70B-Instruct` | | | | |
-| `meta-llama/Meta-Llama-3.1-405B-Instruct` | | | | |
-| `meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo` | | | | |
-| `meta-llama/Meta-Llama-3.1-70B-Instruct` | | | | |
-| `meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo` | | | | |
-| `meta-llama/Meta-Llama-3.1-8B-Instruct` | | | | |
-| `meta-llama/Llama-3.2-11B-Vision-Instruct` | | | | |
-| `meta-llama/Llama-3.2-90B-Vision-Instruct` | | | | |
-| `mistralai/Mixtral-8x7B-Instruct-v0.1` | | | | |
-| `deepseek-ai/DeepSeek-V3` | | | | |
-| `deepseek-ai/DeepSeek-R1` | | | | |
-| `deepseek-ai/DeepSeek-R1-Distill-Llama-70B` | | | | |
-| `deepseek-ai/DeepSeek-R1-Turbo` | | | | |
-| `nvidia/Llama-3.1-Nemotron-70B-Instruct` | | | | |
-| `Qwen/Qwen2-7B-Instruct` | | | | |
-| `Qwen/Qwen2.5-72B-Instruct` | | | | |
-| `Qwen/Qwen2.5-Coder-32B-Instruct` | | | | |
-| `Qwen/QwQ-32B-Preview` | | | | |
-| `google/codegemma-7b-it` | | | | |
-| `google/gemma-2-9b-it` | | | | |
-| `microsoft/WizardLM-2-8x22B` | | | | |
+| Model | Image Input | Object Generation | Tool Usage | Tool Streaming |
+| --------------------------------------------------- | ------------------- | ------------------- | ------------------- | ------------------- |
+| `meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8` | | | | |
+| `meta-llama/Llama-4-Scout-17B-16E-Instruct` | | | | |
+| `meta-llama/Llama-3.3-70B-Instruct-Turbo` | | | | |
+| `meta-llama/Llama-3.3-70B-Instruct` | | | | |
+| `meta-llama/Meta-Llama-3.1-405B-Instruct` | | | | |
+| `meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo` | | | | |
+| `meta-llama/Meta-Llama-3.1-70B-Instruct` | | | | |
+| `meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo` | | | | |
+| `meta-llama/Meta-Llama-3.1-8B-Instruct` | | | | |
+| `meta-llama/Llama-3.2-11B-Vision-Instruct` | | | | |
+| `meta-llama/Llama-3.2-90B-Vision-Instruct` | | | | |
+| `mistralai/Mixtral-8x7B-Instruct-v0.1` | | | | |
+| `deepseek-ai/DeepSeek-V3` | | | | |
+| `deepseek-ai/DeepSeek-R1` | | | | |
+| `deepseek-ai/DeepSeek-R1-Distill-Llama-70B` | | | | |
+| `deepseek-ai/DeepSeek-R1-Turbo` | | | | |
+| `nvidia/Llama-3.1-Nemotron-70B-Instruct` | | | | |
+| `Qwen/Qwen2-7B-Instruct` | | | | |
+| `Qwen/Qwen2.5-72B-Instruct` | | | | |
+| `Qwen/Qwen2.5-Coder-32B-Instruct` | | | | |
+| `Qwen/QwQ-32B-Preview` | | | | |
+| `google/codegemma-7b-it` | | | | |
+| `google/gemma-2-9b-it` | | | | |
+| `microsoft/WizardLM-2-8x22B` | | | | |
The table above lists popular models. Please see the [DeepInfra
diff --git a/content/providers/01-ai-sdk-providers/index.mdx b/content/providers/01-ai-sdk-providers/index.mdx
index 503a3119e3de..dd0b183414b0 100644
--- a/content/providers/01-ai-sdk-providers/index.mdx
+++ b/content/providers/01-ai-sdk-providers/index.mdx
@@ -17,51 +17,53 @@ There are also [community providers](./community-providers) that have been creat
Not all providers support all AI SDK features. Here's a quick comparison of the capabilities of popular models:
-| Provider | Model | Image Input | Object Generation | Tool Usage | Tool Streaming |
-| ------------------------------------------------------------------------ | --------------------------------------------- | ------------------- | ------------------- | ------------------- | ------------------- |
-| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-2-1212` | | | | |
-| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-2-vision-1212` | | | | |
-| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-beta` | | | | |
-| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-vision-beta` | | | | |
-| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4o` | | | | |
-| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4o-mini` | | | | |
-| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4-turbo` | | | | |
-| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4` | | | | |
-| [OpenAI](/providers/ai-sdk-providers/openai) | `o1` | | | | |
-| [OpenAI](/providers/ai-sdk-providers/openai) | `o1-mini` | | | | |
-| [OpenAI](/providers/ai-sdk-providers/openai) | `o1-preview` | | | | |
-| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-7-sonnet-20250219` | | | | |
-| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-5-sonnet-20241022` | | | | |
-| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-5-sonnet-20240620` | | | | |
-| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-5-haiku-20241022` | | | | |
-| [Groq](/providers/ai-sdk-providers/groq) | `meta-llama/llama-4-scout-17b-16e-instruct` | | | | |
-| [Groq](/providers/ai-sdk-providers/groq) | `deepseek-r1-distill-llama-70b` | | | | |
-| [Groq](/providers/ai-sdk-providers/groq) | `llama-3.3-70b-versatile` | | | | |
-| [Groq](/providers/ai-sdk-providers/groq) | `llama-3.1-8b-instant` | | | | |
-| [Groq](/providers/ai-sdk-providers/groq) | `mistral-saba-24b` | | | | |
-| [Groq](/providers/ai-sdk-providers/groq) | `qwen-qwq-32b` | | | | |
-| [Groq](/providers/ai-sdk-providers/groq) | `mixtral-8x7b-32768` | | | | |
-| [Groq](/providers/ai-sdk-providers/groq) | `gemma2-9b-it` | | | | |
-| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo` | | | | |
-| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `meta-llama/Llama-3.3-70B-Instruct` | | | | |
-| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-V3` | | | | |
-| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-R1` | | | | |
-| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-R1-Distill-Llama-70B` | | | | |
-| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-R1-Turbo` | | | | |
-| [Mistral](/providers/ai-sdk-providers/mistral) | `pixtral-large-latest` | | | | |
-| [Mistral](/providers/ai-sdk-providers/mistral) | `mistral-large-latest` | | | | |
-| [Mistral](/providers/ai-sdk-providers/mistral) | `mistral-small-latest` | | | | |
-| [Mistral](/providers/ai-sdk-providers/mistral) | `pixtral-12b-2409` | | | | |
-| [Google Generative AI](/providers/ai-sdk-providers/google-generative-ai) | `gemini-2.0-flash-exp` | | | | |
-| [Google Generative AI](/providers/ai-sdk-providers/google-generative-ai) | `gemini-1.5-flash` | | | | |
-| [Google Generative AI](/providers/ai-sdk-providers/google-generative-ai) | `gemini-1.5-pro` | | | | |
-| [Google Vertex](/providers/ai-sdk-providers/google-vertex) | `gemini-2.0-flash-exp` | | | | |
-| [Google Vertex](/providers/ai-sdk-providers/google-vertex) | `gemini-1.5-flash` | | | | |
-| [Google Vertex](/providers/ai-sdk-providers/google-vertex) | `gemini-1.5-pro` | | | | |
-| [DeepSeek](/providers/ai-sdk-providers/deepseek) | `deepseek-chat` | | | | |
-| [DeepSeek](/providers/ai-sdk-providers/deepseek) | `deepseek-reasoner` | | | | |
-| [Cerebras](/providers/ai-sdk-providers/cerebras) | `llama3.1-8b` | | | | |
-| [Cerebras](/providers/ai-sdk-providers/cerebras) | `llama3.3-70b` | | | | |
+| Provider | Model | Image Input | Object Generation | Tool Usage | Tool Streaming |
+| ------------------------------------------------------------------------ | --------------------------------------------------- | ------------------- | ------------------- | ------------------- | ------------------- |
+| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-2-1212` | | | | |
+| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-2-vision-1212` | | | | |
+| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-beta` | | | | |
+| [xAI Grok](/providers/ai-sdk-providers/xai) | `grok-vision-beta` | | | | |
+| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4o` | | | | |
+| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4o-mini` | | | | |
+| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4-turbo` | | | | |
+| [OpenAI](/providers/ai-sdk-providers/openai) | `gpt-4` | | | | |
+| [OpenAI](/providers/ai-sdk-providers/openai) | `o1` | | | | |
+| [OpenAI](/providers/ai-sdk-providers/openai) | `o1-mini` | | | | |
+| [OpenAI](/providers/ai-sdk-providers/openai) | `o1-preview` | | | | |
+| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-7-sonnet-20250219` | | | | |
+| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-5-sonnet-20241022` | | | | |
+| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-5-sonnet-20240620` | | | | |
+| [Anthropic](/providers/ai-sdk-providers/anthropic) | `claude-3-5-haiku-20241022` | | | | |
+| [Groq](/providers/ai-sdk-providers/groq) | `meta-llama/llama-4-scout-17b-16e-instruct` | | | | |
+| [Groq](/providers/ai-sdk-providers/groq) | `deepseek-r1-distill-llama-70b` | | | | |
+| [Groq](/providers/ai-sdk-providers/groq) | `llama-3.3-70b-versatile` | | | | |
+| [Groq](/providers/ai-sdk-providers/groq) | `llama-3.1-8b-instant` | | | | |
+| [Groq](/providers/ai-sdk-providers/groq) | `mistral-saba-24b` | | | | |
+| [Groq](/providers/ai-sdk-providers/groq) | `qwen-qwq-32b` | | | | |
+| [Groq](/providers/ai-sdk-providers/groq) | `mixtral-8x7b-32768` | | | | |
+| [Groq](/providers/ai-sdk-providers/groq) | `gemma2-9b-it` | | | | |
+| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8` | | | | |
+| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `meta-llama/Llama-4-Scout-17B-16E-Instruct` | | | | |
+| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo` | | | | |
+| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `meta-llama/Llama-3.3-70B-Instruct` | | | | |
+| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-V3` | | | | |
+| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-R1` | | | | |
+| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-R1-Distill-Llama-70B` | | | | |
+| [DeepInfra](/providers/ai-sdk-providers/deepinfra) | `deepseek-ai/DeepSeek-R1-Turbo` | | | | |
+| [Mistral](/providers/ai-sdk-providers/mistral) | `pixtral-large-latest` | | | | |
+| [Mistral](/providers/ai-sdk-providers/mistral) | `mistral-large-latest` | | | | |
+| [Mistral](/providers/ai-sdk-providers/mistral) | `mistral-small-latest` | | | | |
+| [Mistral](/providers/ai-sdk-providers/mistral) | `pixtral-12b-2409` | | | | |
+| [Google Generative AI](/providers/ai-sdk-providers/google-generative-ai) | `gemini-2.0-flash-exp` | | | | |
+| [Google Generative AI](/providers/ai-sdk-providers/google-generative-ai) | `gemini-1.5-flash` | | | | |
+| [Google Generative AI](/providers/ai-sdk-providers/google-generative-ai) | `gemini-1.5-pro` | | | | |
+| [Google Vertex](/providers/ai-sdk-providers/google-vertex) | `gemini-2.0-flash-exp` | | | | |
+| [Google Vertex](/providers/ai-sdk-providers/google-vertex) | `gemini-1.5-flash` | | | | |
+| [Google Vertex](/providers/ai-sdk-providers/google-vertex) | `gemini-1.5-pro` | | | | |
+| [DeepSeek](/providers/ai-sdk-providers/deepseek) | `deepseek-chat` | | | | |
+| [DeepSeek](/providers/ai-sdk-providers/deepseek) | `deepseek-reasoner` | | | | |
+| [Cerebras](/providers/ai-sdk-providers/cerebras) | `llama3.1-8b` | | | | |
+| [Cerebras](/providers/ai-sdk-providers/cerebras) | `llama3.3-70b` | | | | |
This table is not exhaustive. Additional models can be found in the provider
diff --git a/examples/ai-core/src/e2e/deepinfra.test.ts b/examples/ai-core/src/e2e/deepinfra.test.ts
index fc549841dcc3..bc1943a7d61e 100644
--- a/examples/ai-core/src/e2e/deepinfra.test.ts
+++ b/examples/ai-core/src/e2e/deepinfra.test.ts
@@ -13,6 +13,8 @@ createFeatureTestSuite({
name: 'DeepInfra',
models: {
languageModels: [
+ createChatModel('meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8'),
+ createChatModel('meta-llama/Llama-4-Scout-17B-16E-Instruct'),
createChatModel('deepseek-ai/DeepSeek-V3'),
createChatModel('deepseek-ai/DeepSeek-R1'),
createChatModel('deepseek-ai/DeepSeek-R1-Distill-Llama-70B'),
diff --git a/packages/deepinfra/src/deepinfra-chat-settings.ts b/packages/deepinfra/src/deepinfra-chat-settings.ts
index 8694d195e1e3..6c45c9f7dd0d 100644
--- a/packages/deepinfra/src/deepinfra-chat-settings.ts
+++ b/packages/deepinfra/src/deepinfra-chat-settings.ts
@@ -2,6 +2,8 @@ import { OpenAICompatibleChatSettings } from '@ai-sdk/openai-compatible';
// https://deepinfra.com/models/text-generation
export type DeepInfraChatModelId =
+ | 'meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8'
+ | 'meta-llama/Llama-4-Scout-17B-16E-Instruct'
| 'meta-llama/Llama-3.3-70B-Instruct'
| 'meta-llama/Llama-3.3-70B-Instruct-Turbo'
| 'meta-llama/Meta-Llama-3.1-70B-Instruct'