Describe the bug
when using the cross inference profileid, I receive this error: An error occurred while initializing the BedrockChatCompletionService: Unsupported model provider: us
When using a non cross inference profileid, just the modelid, I get this message from AWS: Invocation of model ID anthropic.claude-3-5-sonnet-20241022-v2:0 with on-demand throughput isn’t supported. Retry your request with the ID or ARN of an inference profile that contains this model.
To Reproduce
Steps to reproduce the behavior:
kernelBuilder.AddBedrockChatCompletionService("us.anthropic.claude-3-5-sonnet-20241022-v2:0");
Expected behavior
I would expect either the cross region profile id to work with the semanic kernal when ondemand does not work.
Screenshots
If applicable, add screenshots to help explain your problem.
Platform
- Language: C#
- Source: Microsoft.SemanticKernel.Connectors.Amazon 1.38.0-alpha
- AI model: us.anthropic.claude-3-5-sonnet-20241022-v2:0
- IDE: Visual Studio
- OS: Windows
Additional context
Using this with C#, dot net core 8, Blazor Server
Describe the bug
when using the cross inference profileid, I receive this error: An error occurred while initializing the BedrockChatCompletionService: Unsupported model provider: us
When using a non cross inference profileid, just the modelid, I get this message from AWS: Invocation of model ID anthropic.claude-3-5-sonnet-20241022-v2:0 with on-demand throughput isn’t supported. Retry your request with the ID or ARN of an inference profile that contains this model.
To Reproduce
Steps to reproduce the behavior:
kernelBuilder.AddBedrockChatCompletionService("us.anthropic.claude-3-5-sonnet-20241022-v2:0");
Expected behavior
I would expect either the cross region profile id to work with the semanic kernal when ondemand does not work.
Screenshots
If applicable, add screenshots to help explain your problem.
Platform
Additional context
Using this with C#, dot net core 8, Blazor Server