diff --git a/.github/workflows/docker.yml b/.github/workflows/docker.yml new file mode 100644 index 00000000..0882e992 --- /dev/null +++ b/.github/workflows/docker.yml @@ -0,0 +1,81 @@ +name: Build and Push Docker Image + +on: + push: + branches: + - main + - master + - ai-bankofai-patch-1 + tags: + - 'test' + pull_request: + branches: + - main + - master + - ai-bankofai-patch-1 + workflow_dispatch: + +env: + IMAGE_NAME: bankofai/docs + +jobs: + build-and-push: + runs-on: ubuntu-latest + permissions: + contents: read + + steps: + - name: Checkout repository + uses: actions/checkout@v4 + + - name: Set up Docker Buildx + uses: docker/setup-buildx-action@v3 + + - name: Debug Docker Hub secrets + if: github.event_name != 'pull_request' + env: + DOCKERHUB_USERNAME: ${{ secrets.DOCKERHUB_USERNAME }} + DOCKERHUB_TOKEN: ${{ secrets.DOCKERHUB_TOKEN }} + run: | + echo "username=[$DOCKERHUB_USERNAME]" + echo "token_length=${#DOCKERHUB_TOKEN}" + + - name: Log in to Docker Hub + if: github.event_name != 'pull_request' + env: + DOCKERHUB_USERNAME: ${{ secrets.DOCKERHUB_USERNAME }} + DOCKERHUB_TOKEN: ${{ secrets.DOCKERHUB_TOKEN }} + run: | + echo "$DOCKERHUB_TOKEN" | docker login -u "$DOCKERHUB_USERNAME" --password-stdin + + - name: Extract Docker metadata + id: meta + uses: docker/metadata-action@v5 + with: + images: ${{ env.IMAGE_NAME }} + tags: | + type=raw,value=test + + - name: Determine APP_ENV + id: app_env + run: | + if [[ "${{ github.ref }}" == "refs/heads/main" || "${{ github.ref }}" == "refs/heads/master" ]]; then + echo "APP_ENV=production" >> $GITHUB_OUTPUT + elif [[ "${{ github.ref }}" == refs/tags/* ]]; then + echo "APP_ENV=production" >> $GITHUB_OUTPUT + else + echo "APP_ENV=development" >> $GITHUB_OUTPUT + fi + + - name: Build and push Docker image + uses: docker/build-push-action@v6 + with: + context: . + push: ${{ github.event_name != 'pull_request' }} + tags: ${{ steps.meta.outputs.tags }} + labels: ${{ steps.meta.outputs.labels }} + build-args: | + APP_ENV=${{ steps.app_env.outputs.APP_ENV }} + cache-from: type=gha + cache-to: type=gha,mode=max + platforms: linux/amd64 \ No newline at end of file diff --git a/docs/llmservice/api/API.md b/docs/llmservice/api/API.md new file mode 100644 index 00000000..6f67b862 --- /dev/null +++ b/docs/llmservice/api/API.md @@ -0,0 +1,236 @@ +# AI API (OpenAI Compatible) +Chat completion. Auth: Bearer token. Non-stream: JSON with choices[].content. Stream: SSE chunks with choices[].delta.content. + +## Version: 1.0 + +### BaseURL +https://api.bankofai.io/ + +### Available authorizations +#### bearerAuth (HTTP, bearer) +Bearer ``, e.g. Bearer sk-xxx +Bearer format: JWT + +--- +## Model List + +### [GET] /v1/models +**List models (OpenAI compatible)** + +List available models. Auth: Bearer token. Response: object, success, data. + +#### Responses + +| Code | Description | Schema | +| ---- | ----------- | ------ | +| 200 | object: list; success: true; data: array of { id, object, created, owned_by } | **application/json**: [V1ModelsResponse](#v1modelsresponse)
| +| 400 | Bad Request - invalid parameters or malformed body | **application/json**: [ErrorResponse](#errorresponse)
| +| 401 | Unauthorized - invalid or missing authentication | **application/json**: [ErrorResponse](#errorresponse)
| +| 403 | Forbidden - access denied, insufficient quota, or banned | **application/json**: [ErrorResponse](#errorresponse)
| +| 429 | Too Many Requests - rate limit exceeded | **application/json**: [ErrorResponse](#errorresponse)
| +| 500 | Internal Server Error | **application/json**: [ErrorResponse](#errorresponse)
| + +##### Security + +| Security Schema | Scopes | +| --------------- | ------ | +| bearerAuth | | + +--- +## Chat Completions + +### [POST] /v1/chat/completions +**Create chat completion (OpenAI compatible)** + +Chat completion. Auth: Bearer token. Non-stream: JSON with choices[].content. Stream: SSE chunks with choices[].delta.content. + +#### Request Body + +| Required | Schema | +| -------- | ------ | +| Yes | **application/json**: [ChatCompletionsRequest](#chatcompletionsrequest)
| + +#### Responses + +| Code | Description | Schema | +| ---- | ----------- | ------ | +| 200 | Success. Schema differs by stream mode. | **application/json**: [ChatCompletionsResponse](#chatcompletionsresponse)
**text/event-stream**: [ChatCompletionsResponse](#chatcompletionsresponse)
| +| 400 | Bad Request - invalid parameters, malformed body, or invalid request | **application/json**: [ErrorResponse](#errorresponse)
| +| 401 | Unauthorized - invalid or missing authentication | **application/json**: [ErrorResponse](#errorresponse)
| +| 403 | Forbidden - access denied, insufficient quota, or model access restricted | **application/json**: [ErrorResponse](#errorresponse)
| +| 429 | Too Many Requests - rate limit exceeded | **application/json**: [ErrorResponse](#errorresponse)
| +| 500 | Internal Server Error | **application/json**: [ErrorResponse](#errorresponse)
| +| 502 | Bad Gateway - upstream service error | **application/json**: [ErrorResponse](#errorresponse)
| +| 503 | Service Unavailable - overloaded or no available channel | **application/json**: [ErrorResponse](#errorresponse)
| + +##### Security + +| Security Schema | Scopes | +| --------------- | ------ | +| bearerAuth | | + +--- +### Schemas + +#### ErrorResponse + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| error | { **"message"**: string, **"type"**: string, **"param"**: string, null, **"code"**: string, null } | | No | + +#### V1ModelsResponse + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| object | string | *Example:* `"list"` | No | +| success | boolean | *Example:* `true` | No | +| data | [ [V1ModelItem](#v1modelitem) ] | | No | + +#### V1ModelItem + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| id | string | *Example:* `"gpt-4"` | No | +| object | string | *Example:* `"model"` | No | +| created | integer | *Example:* `1626777600` | No | +| owned_by | string | *Example:* `"openai"` | No | + +#### ChatCompletionsRequest + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| model | string | ID of the model to use (e.g. gpt-4).
*Example:* `"gpt-4"` | Yes | +| messages | [ [ChatMessage](#chatmessage) ] | List of messages in the conversation. | Yes | +| stream | boolean | If true, partial message deltas will be sent as server-sent events. Default false. | No | +| max_tokens | integer | Maximum number of tokens that can be generated in the completion. | No | +| temperature | number | Sampling temperature between 0 and 2. Higher = more random. Default 1. | No | +| top_p | number | Nucleus sampling: consider tokens with top_p probability mass. Default 1. | No | +| stop | | Up to 4 sequences where the API will stop generating. String or array of strings. | No | +| n | integer | How many chat completion choices to generate. Default 1. | No | +| frequency_penalty | number | -2.0 to 2.0. Penalize repeated tokens. Default 0. | No | +| presence_penalty | number | -2.0 to 2.0. Penalize tokens that appear in the text so far. Default 0. | No | +| seed | integer | Random seed for deterministic sampling (if supported by model). | No | +| response_format | [ChatResponseFormat](#chatresponseformat) | | No | +| tools | [ [ChatTool](#chattool) ] | List of tools the model may call. Each has type "function" and function { name, description?, parameters? }. | No | +| tool_choice | | | No | +| user | string | Optional end-user identifier for abuse monitoring. | No | + +#### ChatMessage + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| role | string | "system" \| "user" \| "assistant" \| "tool". System sets behavior; user/assistant are conversation; tool is tool result.
*Example:* `"user"` | No | +| content | string | Message content. For tool role, the result of the tool call.
*Example:* `"Hello"` | No | +| name | string | Optional name for the message author (e.g. to disambiguate multiple users). | No | +| tool_call_id | string | When role is "tool", the id of the tool call this result is for. Required for tool messages. | No | +| tool_calls | [ [ChatToolCallItem](#chattoolcallitem) ] | When role is "assistant" and the model called tools, array of { id, type, function: { name, arguments } }. | No | + +#### ChatResponseFormat + +Specify output format: { "type": "text" } or { "type": "json_object" } or json_schema. + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| type | string | "text" or "json_object". | No | +| json_schema | | When type is json_schema, optional schema for the output. | No | + +#### ChatTool + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| type | string | Must be "function".
*Example:* `"function"` | No | +| function | [ChatToolFunction](#chattoolfunction) | Function definition (name, description, parameters). | No | + +#### ChatToolFunction + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| name | string | Name of the function. | No | +| description | string | Optional description for the model. | No | +| parameters | | Optional JSON schema for the function arguments. | No | + +#### ChatToolCallItem + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| id | string | ID of the tool call. | No | +| type | string | "function".
*Example:* `"function"` | No | +| function | [ChatToolCallFunction](#chattoolcallfunction) | Name and arguments of the call. | No | + +#### ChatToolCallFunction + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| name | string | Name of the function to call. | No | +| arguments | string | JSON string of the arguments. | No | + +#### ToolChoiceObject + +Precise mode: specifies the particular function to call. + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| type | string,
**Available values:** "function" | Must be "function".
*Enum:* `"function"`
*Example:* `"function"` | Yes | +| function | [ToolChoiceFunction](#toolchoicefunction) | Function definition to call. | Yes | + +#### ToolChoiceFunction + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| name | string | Name of the function to call. | Yes | + +#### ChatCompletionsResponse + +Non-stream: object=chat.completion, choices[].message, usage. Stream: object=chat.completion.chunk, choices[].delta; final chunk has usage. + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| id | string | *Example:* `"chatcmpl-xxx"` | No | +| object | string,
**Available values:** "chat.completion", "chat.completion.chunk" | "chat.completion" (non-stream) or "chat.completion.chunk" (stream).
*Enum:* `"chat.completion"`, `"chat.completion.chunk"` | No | +| created | integer | *Example:* `1677652288` | No | +| model | string | *Example:* `"gpt-4"` | No | +| service_tier | string | *Example:* `"default"` | No | +| system_fingerprint | string, null | | No | +| choices | [ [ChatChoice](#chatchoice) ] | Empty in final usage chunk. | No | +| usage | | Non-stream: always present. Stream: null until final chunk. | No | +| obfuscation | string | | No | + +#### ChatMessageContent + +Non-stream choices[].message. Full assistant message with role, content, refusal, annotations. + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| role | string | *Example:* `"assistant"` | No | +| content | string | Assistant reply text. | No | +| refusal | string, null | Refusal reason when model declines; null otherwise. | No | +| annotations | [ ] | Citations, references, etc. | No | + +#### ChatChoice + +Non-stream: message. Stream: delta. finish_reason null until last content chunk. + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| index | integer | | No | +| message | [ChatMessageContent](#chatmessagecontent) | Non-stream only. Full assistant message. | No | +| delta | [ChatChoiceDelta](#chatchoicedelta) | Stream only. Incremental content; empty {} on stop. | No | +| finish_reason | string, null | Null until done; e.g. "stop", "length", "tool_calls". | No | + +#### ChatChoiceDelta + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| content | string | | No | +| role | string | | No | +| tool_calls | [ [ChatToolCallItem](#chattoolcallitem) ] | | No | + +#### ChatUsage + +| Name | Type | Description | Required | +| ---- | ---- | ----------- | -------- | +| prompt_tokens | integer | Number of tokens in the prompt. | No | +| completion_tokens | integer | Number of tokens in the completion. | No | +| total_tokens | integer | Total tokens (prompt + completion). | No | +| prompt_tokens_details | { **"cached_tokens"**: integer, **"audio_tokens"**: integer } | | No | +| completion_tokens_details | { **"reasoning_tokens"**: integer, **"audio_tokens"**: integer, **"accepted_prediction_tokens"**: integer, **"rejected_prediction_tokens"**: integer } | | No | diff --git a/docs/llmservice/introduction.md b/docs/llmservice/introduction.md new file mode 100644 index 00000000..b6233320 --- /dev/null +++ b/docs/llmservice/introduction.md @@ -0,0 +1,19 @@ +# Welcome to LLM Service + +## About LLM Service + +LLM Service is a professional AI service module within the Bank of AI ecosystem, built on top-tier blockchain infrastructure. It is dedicated to providing users with an efficient, user-friendly, and creative AI interaction experience. As a core AI service infrastructure of Bank of AI, this service leverages the decentralization, security, and high efficiency of blockchain technology to introduce a brand-new AI service model. + +The service's core features include: + +* **Multi-Model AI Chat:** We integrate various industry-leading Large Language Models (LLMs), allowing users to select the most suitable model based on their specific needs. +* **Powerful Integrated AI Services:** We offer comprehensive AI-related API services, enabling users to access and integrate them rapidly and easily within the Bank of AI framework. +* **Web3 Native Experience:** Through seamless integration with mainstream Web3 wallets, we provide an end-to-end native experience, from login to payment. + +## Why Choose LLM Service? + +Choosing our LLM Service means enjoying the unique advantages of a secure blockchain ecosystem alongside meticulously designed features. + +* **Multi-chain Ecosystem Advantages:** As part of the Bank of AI ecosystem, users can make payments using mainstream tokens on supported chains, benefiting from fast transaction confirmations and low fees. +* **Low Cost & High Efficiency:** By optimizing resources and ensuring efficient on-chain interactions, we deliver highly cost-effective AI services to users. +* **Security & Privacy Protection:** We utilize a decentralized login method. Users can complete authentication simply by signing with their Web3 wallet, ensuring greater security and privacy for all AI interactions. diff --git a/docs/llmservice/models/chatgpt-5-2.md b/docs/llmservice/models/chatgpt-5-2.md new file mode 100644 index 00000000..a098e9c1 --- /dev/null +++ b/docs/llmservice/models/chatgpt-5-2.md @@ -0,0 +1,30 @@ +# ChatGPT-5.2 + +## Overview +ChatGPT-5.2 is the latest generation of the flagship large language model developed by OpenAI. Building upon the powerful capabilities of the 5.1 version, it further optimizes the speed of multimodal processing and the execution efficiency of complex tasks, making it the ideal choice for professional users seeking ultimate performance and efficiency. + +## Key Features +* **Efficient Multimodal Processing:** Significantly improves the parsing and generation speed of image and video content compared to 5.1, achieving a smoother multimodal interaction experience. +* **Enhanced Task Execution Efficiency:** Optimizes the internal reasoning engine, allowing for faster and more accurate conclusions when handling long-chain, multi-step complex tasks. +* **Stronger Interference Resistance:** Exhibits greater robustness and accuracy when processing inputs containing significant noise or ambiguous instructions. + +## Best Use Cases +* **Real-time Data Analysis and Visualization:** Capable of quickly processing real-time data streams and generating complex charts and visualization reports. +* **Complex Project Management and Planning:** Assists with task decomposition, resource allocation, and risk assessment for efficient decision support. +* **High-Frequency, High-Precision Professional Consulting:** Suitable for professional fields requiring fast and accurate responses, such as financial trading analysis and legal document retrieval. + +## Capabilities and Limitations + +| Capability | Detailed Description | +| :--- | :--- | +| **Reasoning Ability** | Extremely Strong. Maintains a leading position in complex logical reasoning and scientific computation, with improved efficiency. | +| **Creative Ability** | Extremely Strong. Can generate high-quality, in-depth content, particularly excelling in structured and professional texts. | +| **Multimodal Ability** | Comprehensive and Efficient. Supports input and understanding of images, videos, and audio, and can quickly generate high-quality image content. | +| **Response Speed** | Medium to Slow. Improved compared to 5.1, but still a deep analysis model, not suitable for extremely low-latency scenarios. | +| **Context Window** | Huge. Supports a context window of millions of tokens. | + +## Credits and Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | +| :--- | :--- | :--- | +| **ChatGPT-5.2** | 1.75 | 14.00 | diff --git a/docs/llmservice/models/chatgpt-5-mini.md b/docs/llmservice/models/chatgpt-5-mini.md new file mode 100644 index 00000000..2cfa3b82 --- /dev/null +++ b/docs/llmservice/models/chatgpt-5-mini.md @@ -0,0 +1,30 @@ +# ChatGPT-5-mini + +## Overview +ChatGPT-5-mini is an efficient and economical lightweight language model. It is optimized for fast, smooth daily conversations and general tasks, making it a premier choice for cost-effective AI interaction within the Bank of AI ecosystem. + +## Key Features +* **Extremely Fast Response:** Deeply optimized for low response latency, providing a near real-time conversation experience. +* **High Cost-Effectiveness:** While ensuring high-quality output, its computational cost is significantly lower than flagship models, achieving a balance between performance and cost. +* **Strong General Capabilities:** Covers a wide range of daily application scenarios, from quick Q&A to text organization, with stable and reliable performance. + +## Best Use Cases +* **Daily Conversation and Quick Q&A:** Acts as a smart assistant for quickly answering factual questions and engaging in casual chat. +* **Text Processing:** Summarizing, polishing, formatting, and extracting keywords from emails, articles, and documents. +* **Initial Draft Generation:** Quickly generating drafts for social media posts, product descriptions, and blog articles. + +## Capabilities and Limitations + +| Capability | Detailed Description | +| :--- | :--- | +| **Reasoning Ability** | **Medium.** Can handle simple logical reasoning, but may falter on multi-step complex problems. | +| **Creative Ability** | **Medium.** Generates fluent and coherent text, but is relatively limited in depth and professional creativity. | +| **Multimodal Ability** | **Not Supported.** This model focuses on text processing and does not have image or audio understanding capabilities. | +| **Response Speed** | **Fast.** One of the fastest responding models on the platform. | +| **Context Window** | **Standard.** Supports tens of thousands of tokens, sufficient for most daily conversation scenarios. | + +## Credits and Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | +| :--- | :--- | :--- | +| **ChatGPT-5-mini** | 0.25 | 2.00 | diff --git a/docs/llmservice/models/chatgpt-5-nano.md b/docs/llmservice/models/chatgpt-5-nano.md new file mode 100644 index 00000000..38dd9559 --- /dev/null +++ b/docs/llmservice/models/chatgpt-5-nano.md @@ -0,0 +1,30 @@ +# ChatGPT-5-nano + +## Overview +ChatGPT-5-nano is an advanced language model that strikes an excellent balance between performance, speed, and cost. It is designed to provide near-professional AI capabilities at a moderate cost within the Bank of AI ecosystem. + +## Key Features +* **Enhanced Reasoning Ability:** Nano shows significant improvements in logical reasoning, code generation, and multilingual processing compared to lighter models. +* **Efficient Performance:** Carefully tuned to maintain high output quality while sustaining a fast response speed. +* **Multifunctional Integration:** Capable of handling a diverse range of tasks, making it a powerful assistant for developers and content creators. + +## Best Use Cases +* **Code Assistance and Debugging:** Understanding and generating code across multiple programming languages, assisting with debugging and documentation. +* **Multilingual Translation and Writing:** Providing high-quality cross-language translation and creating content in authentic language styles. +* **Structured Content Generation:** Generating well-formatted reports, technical documents, tutorials, and other structured content. + +## Capabilities and Limitations + +| Capability | Detailed Description | +| :--- | :--- | +| **Reasoning Ability** | **Strong.** Can handle complex logical problems and programming tasks, performing well in specific domains. | +| **Creative Ability** | **Strong.** Generates creative and in-depth text content, meeting high writing requirements. | +| **Multimodal Ability** | **Limited Support.** Can understand and describe simple image content, but does not support deep multimodal analysis. | +| **Response Speed** | **Medium.** Faster than flagship models, though slightly slower than the mini model. | +| **Context Window** | **Large.** Supports a context window of hundreds of thousands of tokens for long document processing. | + +## Credits and Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | +| :--- | :--- | :--- | +| **ChatGPT-5-nano** | 0.05 | 0.40 | diff --git a/docs/llmservice/models/claude-haiku-4-5.md b/docs/llmservice/models/claude-haiku-4-5.md new file mode 100644 index 00000000..78fcf1f0 --- /dev/null +++ b/docs/llmservice/models/claude-haiku-4-5.md @@ -0,0 +1,30 @@ +# Claude Haiku 4.5 + +## Overview +Claude Haiku 4.5 is the fastest and most compact AI model developed by Anthropic, integrated into the Bank of AI platform. It is designed to provide near-instantaneous responses, making it the premier choice for building seamless, real-time AI experiences. + +## Key Features +* **Unmatched Response Speed:** As the fastest model in its intelligence class, it delivers extremely low latency, ideal for applications requiring immediate interaction. +* **Ultimate Cost-Effectiveness:** Highly competitive pricing makes it the most viable option for deploying AI at scale across massive user scenarios. +* **Enterprise-Grade Robustness:** Undergone rigorous security testing to ensure the reliability and safety required for professional enterprise applications. + +## Best Use Cases +* **Real-time Chatbots & Moderation:** Providing smooth, natural conversation experiences and quickly moderating user-generated content. +* **Mobile AI Applications:** Optimized for mobile environments where latency and resource consumption are critical factors. +* **Workflow Streamlining:** Automating routine tasks such as email classification, meeting summarization, and form data extraction to improve daily efficiency. + +## Capabilities and Limitations + +| Capability | Detailed Description | +| :--- | :--- | +| **Reasoning Ability** | **Medium.** Capable of handling general tasks, but limited in solving highly complex or multi-step reasoning problems. | +| **Creative Ability** | **Medium.** Generates concise and fluent text, best suited for information delivery rather than deep creative writing. | +| **Multimodal Ability** | **Supported.** Basic image understanding capabilities to identify and describe objects within visual inputs. | +| **Response Speed** | **Extremely Fast.** The fastest responding model on the platform, enabling near-instantaneous interaction. | +| **Context Window** | **Huge.** Supports an extra-long context window, capable of handling large documents and extensive conversation histories. | + +## Credits and Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | +| :--- | :--- | :--- | +| **Claude Haiku 4.5** | 1.00 | 5.00 | diff --git a/docs/llmservice/models/claude-opus-4-5.md b/docs/llmservice/models/claude-opus-4-5.md new file mode 100644 index 00000000..004e1da9 --- /dev/null +++ b/docs/llmservice/models/claude-opus-4-5.md @@ -0,0 +1,30 @@ +# Claude Opus 4.5 + +## Overview +Claude Opus 4.5 is the flagship AI model developed by Anthropic, integrated into the Bank of AI platform. It is renowned for its superior intelligence, top-tier performance, and profound ethical considerations, setting new benchmarks for professional AI applications. + +## Key Features +* **Unparalleled Intelligence:** Sets industry benchmarks in graduate-level reasoning, mathematics, and coding, capable of solving extremely complex and open-ended problems. +* **Powerful Vision Understanding:** Possesses top-tier visual analysis capabilities, accurately interpreting complex charts, scientific illustrations, and multi-image documents. +* **Extra-Long Context Processing:** Supports a massive context window, effortlessly handling and analyzing documents or entire codebases with hundreds of thousands of words. + +## Best Use Cases +* **Complex System Analysis:** Understanding complex flowcharts and API documentation, and autonomously writing code to automate cross-system tasks. +* **High-End Scientific Research:** Analyzing complex scientific papers, interpreting experimental data, and proposing innovative research hypotheses. +* **Enterprise Strategic Planning:** Deeply analyzing market trends, financial reports, and legal contracts to provide high-quality decision support. + +## Capabilities and Limitations + +| Capability | Detailed Description | +| :--- | :--- | +| **Reasoning Ability** | **Top-tier.** Outperforms other models in benchmark tests, especially adept at handling problems requiring deep thought. | +| **Creative Ability** | **Extremely Strong.** Generates logically coherent, insightful, and elegantly styled text across all genres. | +| **Multimodal Ability** | **Powerful.** Skilled at extracting insights from complex visual materials and visual data analysis. | +| **Response Speed** | **Slow.** Prioritizes the highest quality and depth over response time. | +| **Context Window** | **Huge.** Supports a context window of over 200K tokens, with potential for further expansion. | + +## Credits and Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | +| :--- | :--- | :--- | +| **Claude Opus 4.5** | 5.00 | 25.00 | diff --git a/docs/llmservice/models/claude-opus-4-6.md b/docs/llmservice/models/claude-opus-4-6.md new file mode 100644 index 00000000..96809673 --- /dev/null +++ b/docs/llmservice/models/claude-opus-4-6.md @@ -0,0 +1,32 @@ +# Claude Opus 4.6 + +## Overview +Claude Opus 4.6 is the latest iteration of Anthropic's flagship AI model series. Building upon the 4.5 version's exceptional intelligence and profound ethical considerations, it further enhances complex reasoning capabilities, multimodal understanding, and performance in specific professional domains, aiming to provide a more powerful and precise AI experience within the Bank of AI ecosystem. + +## Key Features +* **Enhanced Complex Reasoning:** Deeply optimized for multi-step, cross-domain, and abstract concept reasoning, performing exceptionally well in solving complex scientific, engineering, and legal problems. +* **Refined Multimodal Understanding:** Improved detail capture and contextual correlation for images, charts, and videos, enabling more accurate analysis of complex multimodal inputs. +* **Deepened Professional Knowledge:** Expanded breadth and depth in specific fields such as advanced financial analysis, drug discovery, and complex system design. +* **Distinction from 4.5:** Version 4.6 focuses on **depth and precision** improvements, providing more reliable and detailed outputs for tasks requiring ultimate accuracy. + +## Best Use Cases +* **Cutting-Edge Scientific Research:** Assisting in complex data analysis, theoretical validation, and exploration of new discoveries. +* **Advanced Strategic Consulting:** Providing strategic decision support based on vast information and deep analytical insights. +* **Legal and Compliance Review:** Processing complex legal texts, performing risk assessments, and compliance analysis. +* **Innovative Content Generation:** Creating research reports, technical whitepapers, and high-end market analysis requiring logical rigor. + +## Capabilities and Limitations + +| Capability | Detailed Description | +| :--- | :--- | +| **Reasoning Ability** | **Top-tier.** Further enhanced beyond 4.5, particularly adept at problems requiring deep thought and abstract logic. | +| **Creative Ability** | **Extremely Strong.** Generates logically rigorous, insightful, and elegantly styled professional texts. | +| **Multimodal Ability** | **Refined.** Enhanced ability to extract insights from complex visual materials and subtle multimodal information. | +| **Response Speed** | **Slow.** Prioritizes quality and depth; suitable for tasks that are not time-sensitive. | +| **Context Window** | **Huge.** Supports an ultra-long context window for processing massive documents and codebases. | + +## Credits and Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | +| :--- | :--- | :--- | +| **Claude Opus 4.6** | 5.00 | 25.00 | diff --git a/docs/llmservice/models/claude-sonnet-4-5.md b/docs/llmservice/models/claude-sonnet-4-5.md new file mode 100644 index 00000000..b5626474 --- /dev/null +++ b/docs/llmservice/models/claude-sonnet-4-5.md @@ -0,0 +1,30 @@ +# Claude Sonnet 4.5 + +## Overview +Claude Sonnet 4.5 is a balanced AI model developed by Anthropic, integrated into the Bank of AI platform. It is designed to provide near-flagship intelligence at a lower cost, making it the ideal choice for large-scale enterprise AI deployment. + +## Key Features +* **Ideal Intelligence & Speed:** Far surpasses similar models in intelligence while offering twice the response speed of the Opus model series. +* **Enterprise-Grade Optimization:** Specifically tuned for core business scenarios such as knowledge retrieval, sales automation, and complex data analysis. +* **Extra-Long Context & High Recall:** Supports a massive context window while maintaining extremely high information recall accuracy even in very long documents. + +## Best Use Cases +* **Enterprise Knowledge Management:** Quickly and accurately retrieving information from vast documentation to provide precise answers for customers or employees. +* **Sales & Marketing Automation:** Analyzing market data, generating personalized marketing copy, and automatically processing sales leads. +* **Code Quality Control:** Efficiently generating and reviewing code to help development teams improve both speed and code quality. + +## Capabilities and Limitations + +| Capability | Detailed Description | +| :--- | :--- | +| **Reasoning Ability** | **Strong.** Performs well in most reasoning and coding tasks, meeting the needs of professional business applications. | +| **Creative Ability** | **Strong.** Generates high-quality text that complies with rigorous business and professional standards. | +| **Multimodal Ability** | **Supported.** Strong image understanding capabilities for processing documents and reports containing charts. | +| **Response Speed** | **Medium.** Strikes an excellent balance between intelligence and speed, ideal for interactive applications. | +| **Context Window** | **Huge.** Shares the extra-long context processing capability with the flagship Opus model. | + +## Credits and Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | +| :--- | :--- | :--- | +| **Claude Sonnet 4.5** | 3.00 | 15.00 | diff --git a/docs/llmservice/models/claude-sonnet-4-6.md b/docs/llmservice/models/claude-sonnet-4-6.md new file mode 100644 index 00000000..fa6c270d --- /dev/null +++ b/docs/llmservice/models/claude-sonnet-4-6.md @@ -0,0 +1,32 @@ +# Claude Sonnet 4.6 + +## Overview +Claude Sonnet 4.6 is a major upgrade to Anthropic's Sonnet series, designed to deliver performance close to the flagship Opus model at a more economical cost. It achieves comprehensive improvements in coding, computer use, long-context reasoning, and knowledge work. For the first time in the Sonnet series, it introduces a **1,000,000 token context window (beta)**, making it an ideal choice for enterprise-level automation and complex task processing. + +## Key Features +* **Near-Opus Performance:** In multiple benchmarks, especially for processing enterprise documents (charts, PDFs, tables), its performance is comparable to the Opus 4.6. +* **1M Token Context Window:** Brings a million-token context window to the Sonnet series for the first time, capable of processing entire codebases or massive research papers. +* **Excellent Coding & Computer Use:** Demonstrates outstanding ability to use real software, reaching or exceeding human-level performance in complex code fixes and multi-step web workflows. +* **Distinction from 4.5:** A comprehensive skill upgrade over 4.5 with stronger instruction-following, fewer hallucinations, and more reliable multi-step task execution. + +## Best Use Cases +* **Enterprise Document Analysis:** Efficiently processing and reasoning with complex PDF documents containing charts and tables. +* **Large-Scale Codebase Refactoring:** Ideal for codebase-level analysis and modification thanks to its million-token context. +* **Cost-Effective Agentic Workflows:** Providing top-tier performance for automated tasks requiring multi-step planning. +* **Knowledge Work & Design:** Excelling in high-quality content generation and front-end design with fewer iterations required. + +## Capabilities and Limitations + +| Capability | Detailed Description | +| :--- | :--- | +| **Reasoning Ability** | **Strong.** Performance in long-context reasoning and complex problem-solving is close to the Opus level. | +| **Creative Ability** | **Strong.** Particularly adept at generating high-quality code and front-end pages with a good sense of design. | +| **Multimodal Ability** | **Supported.** Capable of processing and understanding multimodal inputs such as images and PDFs. | +| **Response Speed** | **Medium.** Achieves an excellent balance between performance and speed. | +| **Context Window** | **1,000,000 Tokens (beta).** Supports a max output of 128,000 tokens. | + +## Credits and Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | +| :--- | :--- | :--- | +| **Claude Sonnet 4.6** | 3.00 | 15.00 | diff --git a/docs/llmservice/models/gemini-3-1-pro.md b/docs/llmservice/models/gemini-3-1-pro.md new file mode 100644 index 00000000..b0ef042d --- /dev/null +++ b/docs/llmservice/models/gemini-3-1-pro.md @@ -0,0 +1,31 @@ +# Gemini 3.1 pro + +## Overview +Gemini 3.1 pro is a significant upgrade to Google's flagship multimodal model series. It is designed to deliver enhanced reasoning, more reliable output, and greater efficiency. While retaining its native multimodal capabilities, it significantly improves accuracy in handling complex tasks and resolves the output truncation issues present in previous versions. + +## Key Features +* **Doubled Reasoning Capability:** Achieved a groundbreaking score on the ARC-AGI-2 benchmark, demonstrating a performance increase of more than double in novel pattern recognition compared to its predecessor. +* **Ultra-Long Context & Output:** Supports a context window of up to **1,000,000 tokens** and can generate outputs of up to **64,000 tokens** in a single run, ensuring long responses are never truncated. +* **Native Multimodality:** Seamlessly processes and understands information across various formats, including text, images, audio, and video, within a single model architecture. +* **Higher Operational Efficiency:** Delivers more reliable results with optimized token usage, enhancing overall performance for complex workflows. + +## Best Use Cases +* **Complex Code Generation:** Ideal for software development tasks that require deep logical reasoning and complete, non-truncated code blocks. +* **Large-Scale Data Analysis:** Capable of processing and deeply analyzing vast amounts of documents or datasets in a single pass. +* **Creative Long-Form Writing:** Drafting detailed reports, scripts, or technical whitepapers that require sustained logical coherence over thousands of words. + +## Capabilities and Limitations + +| Capability | Detailed Description | +| :--- | :--- | +| **Reasoning Ability** | **Extremely Strong.** Significant improvements in abstract reasoning and novel problem-solving. | +| **Creative Ability** | **Extremely Strong.** Capable of generating high-quality, lengthy, and logically coherent content. | +| **Multimodal Ability** | **Native & Comprehensive.** Supports mixed inputs of text, images, audio, and video seamlessly. | +| **Response Speed** | **Medium to Slow.** Prioritizes the quality, depth, and completeness of professional-grade output. | +| **Context Window** | **1,000,000 Tokens.** (Supports a maximum output of 64,000 tokens). | + +## Credits and Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | +| :--- | :--- | :--- | +| **Gemini 3.1 pro** | 2.00 | 12.00 | diff --git a/docs/llmservice/models/gemini-3-flash.md b/docs/llmservice/models/gemini-3-flash.md new file mode 100644 index 00000000..78a88384 --- /dev/null +++ b/docs/llmservice/models/gemini-3-flash.md @@ -0,0 +1,30 @@ +# Gemini 3 flash + +## Overview +Gemini 3 flash is the fastest and most efficient model in the Gemini 3 series released by Google. It is designed for applications requiring rapid response and high throughput, retaining the native multimodal capabilities of the series while significantly optimizing for speed and cost. + +## Key Features +* **Rapid Response & High Throughput:** Optimized for low-latency and high-concurrency scenarios, making it the preferred choice for real-time AI applications. +* **Efficient Multimodality:** Inherits native multimodal capabilities to quickly process and understand information like images and audio, with lower computational requirements than the Pro version. +* **Excellent Cost-Effectiveness:** Ensures high speed and multimodal versatility while significantly reducing operating costs within the Bank of AI ecosystem. + +## Best Use Cases +* **Real-time Chatbots:** Providing smooth, instant, and multimodal interactive experiences for customer service and support. +* **Content Moderation:** Quickly identifying and filtering non-compliant content in both text and image formats. +* **Mobile & Edge Applications:** Ideal for latency-sensitive scenarios that require quick feedback and efficient resource usage. + +## Capabilities and Limitations + +| Capability | Detailed Description | +| :--- | :--- | +| **Reasoning Ability** | **Strong.** Handles most general and complex reasoning tasks, though less depth than the Pro version for niche professional problems. | +| **Creative Ability** | **Strong.** Quickly generates high-quality text and detailed multimodal content descriptions. | +| **Multimodal Ability** | **Native & Efficient.** Possesses strong multimodal understanding, optimized for speed over deep exhaustive analysis. | +| **Response Speed** | **Extremely Fast.** One of the fastest models on the platform, enabling near-instantaneous interaction. | +| **Context Window** | **Huge.** Supports an extremely long context window, consistent with the flagship Pro version. | + +## Credits and Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | +| :--- | :--- | :--- | +| **Gemini 3 flash** | 0.50 | 3.00 | diff --git a/docs/llmservice/models/glm-5.md b/docs/llmservice/models/glm-5.md new file mode 100644 index 00000000..cd7e1682 --- /dev/null +++ b/docs/llmservice/models/glm-5.md @@ -0,0 +1,47 @@ +# GLM-5 + +## Overview +**GLM-5** is Zhipu AI's new-generation flagship foundation model, specifically designed for **Coding** and **Agent** scenarios. It achieves State-Of-The-Art (SOTA) performance in open-source complex system engineering and long-horizon tasks, with a real-world coding experience approaching **Claude Opus** level. + +Based on a **744B** scale foundation model, combined with asynchronous reinforcement learning and sparse attention mechanisms, GLM-5 marks a paradigm shift from "writing code" to "building systems". + +--- + +## Key Features +* **Parameter Scale and Data Volume:** The base model's parameter scale has expanded to **744B** (with **40B** activated parameters), and pre-training data has increased to **28.5T**, significantly enhancing the model's breadth and depth of knowledge. +* **Ultra-Long Context and Output:** Supports a context window of up to **200K tokens** and a maximum output length of **128K tokens**, enabling excellent performance in handling complex code repositories and multi-step tasks. +* **Exceptional Coding & Agent Capabilities:** Systematically strengthened programming capabilities, excelling in code generation with low hallucination rates and efficient token utilization. +* **Multiple Thinking Modes:** Offers various thinking modes to support more flexible and in-depth problem-solving. + +--- + +## Best Use Cases +1. **Complex System Engineering:** Construction and management of complex software systems, assisting in system design and optimization. +2. **Long-Horizon Agent Tasks:** Agent tasks requiring multi-step planning, execution, and feedback (e.g., automated workflows). +3. **High-Precision Code Debugging:** Provides human-level coding assistance to improve development efficiency. +4. **Large-Scale Document Analysis:** Deep information extraction and summarization for massive document sets. + +--- + +## Capabilities and Limitations + +| Capability | Detailed Description | +| :--- | :--- | +| **Reasoning Ability** | **Extremely Strong.** Excels in complex logical reasoning and multi-step planning. | +| **Creative Ability** | **Extremely Strong.** Particularly adept at code generation and system design. | +| **Multimodal Ability** | Primarily focuses on text/code; can be integrated with visual tools on Zhipu platform. | +| **Response Speed** | **30-50 tokens/s.** Balances high-quality output with efficient speed. | +| **Context Window** | 200K Tokens | +| **Max Output** | 128K Tokens | + +--- + +## Credits and Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | +| :--- | :--- | :--- | +| **GLM-5** | 0.30 | 2.55 | + +--- + +> **Note:** For optimal performance in coding tasks, it is recommended to provide clear system prompts and utilize the 128K output capacity for building complete modules. diff --git a/docs/llmservice/models/kimi-k2.5.md b/docs/llmservice/models/kimi-k2.5.md new file mode 100644 index 00000000..9cdbe112 --- /dev/null +++ b/docs/llmservice/models/kimi-k2.5.md @@ -0,0 +1,48 @@ +# Kimi-K2.5 + +## Core Overview +**Kimi-K2.5** is Moonshot AI’s most versatile model to date, featuring a **native multimodal architecture** that simultaneously supports visual and text input, thinking and non-thinking modes, and conversational and Agent tasks. + +With its **256K ultra-long context window**, multimodal understanding, and advanced Tool Calling capabilities, it sets a new benchmark in open-source visual programming and Agent clusters, empowering developers to build next-generation AI applications. + +--- + +## Key Features +* **Native Multimodal Architecture:** Supports mixed input of visual and text, excels in image recognition and visual programming. +* **256K Ultra-Long Context Window:** Provides a **256,000 token** window, supporting long-form reasoning and processing of massive datasets. +* **Agent Clusters & Tool Calling:** Supports a preview version of Agent clusters (up to **100 sub-agents** and **1,500 tool calls**), operating 4.5x faster than single-agent configurations. +* **Exceptional Coding Capabilities:** Leading performance in SWE-Bench and LiveCodeBench, offering competitive programming skills at a fraction of the cost of comparable models. +* **Thinking Modes:** Flexible switching between quick response and deep reasoning/planning modes. + +--- + +## Best Use Cases +1. **Visual Programming & Automation:** Pixel-level webpage replication and expert-level office task automation. +2. **Ultra-Long Text Analysis:** Legal document review, massive research report analysis, and full codebase understanding. +3. **Multi-Agent Collaboration:** Building complex automated workflows involving multiple specialized sub-agents. +4. **Professional Code Generation:** High-efficiency code generation, optimization, and deep debugging for developers. + +--- + +## Capabilities and Limitations + +| Capability | Detailed Description | +| :--- | :--- | +| **Reasoning Ability** | **Extremely Strong.** Excels in long-context reasoning and Agent task planning. | +| **Creative Ability** | **Extremely Strong.** Adept at visual programming and multimodal content creation. | +| **Multimodal Ability** | **Native Multimodal.** Outstanding performance in visual understanding and input. | +| **Response Speed** | Fast in quick mode; highly efficient parallel processing in Agent cluster mode. | +| **Context Window** | 256,000 Tokens | +| **Max Output** | 256,000 Tokens | + +--- + +## Credits and Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | +| :--- | :--- | :--- | +| **Kimi-K2.5** | 0.23 | 3.00 | + +--- + +> **Pro Tip:** When using Kimi-K2.5 for complex system design, try leveraging its **Thinking Mode** for architecture planning before switching to standard mode for rapid code execution. diff --git a/docs/llmservice/models/minimax-m2.5.md b/docs/llmservice/models/minimax-m2.5.md new file mode 100644 index 00000000..fde22f68 --- /dev/null +++ b/docs/llmservice/models/minimax-m2.5.md @@ -0,0 +1,48 @@ +# MiniMax-M2.5 + +## Core Overview +**MiniMax-M2.5** is MiniMax's independently developed flagship multimodal general large model, designed for **high-throughput and low-latency** production environments. + +It achieves industry-leading performance in coding and Agent capabilities, with the ability to natively understand, generate, and integrate multiple modalities including **text, audio, images, video, and music**. M2.5 aims to provide top-tier performance at extremely low costs, excelling particularly in complex task processing and professional office scenarios. + +--- + +## Key Features +* **Industry-Leading Coding & Agent Capabilities:** Achieved best-in-industry performance on the **Multi-SWE-Bench** benchmark, demonstrating higher decision-making maturity and more efficient token utilization. +* **Efficient Multimodal Processing:** Natively supports the integration of text, audio, images, video, and music, providing a truly rich multimodal interactive experience. +* **Ultra-Long Context Processing:** Features a substantial context window of **197K tokens**, optimized through reinforcement learning for precise task decomposition. +* **High Throughput, Low Latency:** Optimized for production with **100 TPS** and **50 TPS** versions. Pricing is significantly lower (1/10 to 1/20) than comparable models. +* **Enhanced Office Scenarios:** Significant capability improvements in handling professional software tasks such as Word, PPT, and Excel financial modeling. + +--- + +## Best Use Cases +1. **Enterprise-Level Automated Workflows:** Ideal for automation requiring fast multimodal processing and complex Agent decision-making. +2. **Software Development & Code Assistance:** Industry-leading generation and debugging, especially in large, complex codebases. +3. **Multimodal Content Creation:** Innovative cross-modal generation (e.g., text-to-video or image-to-music integration). +4. **Advanced Office Document Processing:** High-efficiency information extraction and modeling for Word, Excel, and PPT. + +--- + +## Capabilities and Limitations + +| Capability | Detailed Description | +| :--- | :--- | +| **Reasoning Ability** | **Extremely Strong.** High decision-making maturity for multi-step Agent tasks. | +| **Creative Ability** | **Extremely Strong.** Proficient in multimodal creation and office document automation. | +| **Multimodal Ability** | **Native Multimodal.** Supports text, audio, images, video, and music. | +| **Response Speed** | **Extremely Fast.** Offers 100 TPS and 50 TPS versions for high-throughput needs. | +| **Context Window** | 197,000 Tokens | +| **Max Output** | 131,000 Tokens | + +--- + +## Credits and Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | +| :--- | :--- | :--- | +| **MiniMax-M2.5** | 0.30 | 1.20 | + +--- + +> **Pro Tip:** MiniMax-M2.5 is exceptionally cost-effective for **high-volume production API calls**. If your workflow requires processing thousands of documents per hour, M2.5 provides the best balance of speed and budget. diff --git a/docs/llmservice/openclaw/integration-guide.md b/docs/llmservice/openclaw/integration-guide.md new file mode 100644 index 00000000..041277e5 --- /dev/null +++ b/docs/llmservice/openclaw/integration-guide.md @@ -0,0 +1,353 @@ +# Integrating OpenClaw with Bank of AI +## From Zero to Private Agent: Deploying OpenClaw with Bank of AI in 15 Minutes + +OpenClaw (formerly ClawdBot or Moltbot) is an open-source personal AI assistant. Unlike cloud-based SaaS tools, OpenClaw runs locally on your own machine, giving you full control over your data and workflows. You can interact with it through familiar messaging platforms like WhatsApp, Telegram, Lark, and DingTalk to handle emails, manage calendars, write code, or even automate your smart home. + +OpenClaw is more than just a chatbot; it's a truly functional "agent" designed for real-world execution. It features persistent memory, access to your local file system and the internet, and the ability to grow more powerful by expanding its "skills". + +Because it's open-source and self-hosted, OpenClaw has attracted a vibrant community of developers and tech enthusiasts. The community has pioneered creative use cases, from automating business operations to managing personal life—showcasing the immense potential of a truly personal AI. + +This guide will walk you through everything from scratch: downloading, installing, and setting up OpenClaw, as well as connecting it to the Bank of AI API. By the end, you'll have built your very own AI assistant. + +--- + +## **Step 1: Get Your Bank of AI API Key** + +1. Log in to https://chat.bankofai.io/chat +2. Navigate to the API key management page and apply for your api_key + +--- + +## **Step 2: Prepare Your System** + +Before installing, make sure your system meets these basic requirements. OpenClaw is designed for Unix-like environments but runs perfectly on Windows via WSL2 (Windows Subsystem for Linux 2). + +| Requirement | Details | +| ---------------- | --------------------------------------------------------------------------------------------------------------- | +| Node.js | Version 22 or higher. This is the runtime environment for OpenClaw. | +| Operating System | macOS, Linux, or Windows (via WSL2). | +| Package Manager | pnpm is required to compile from source. For a standard install, npm (which comes with Node.js) is recommended. | + +To check your environment, open a terminal and run: + +```bash +node -v +``` + +If the version is lower than v22.0.0, or you see a "command not found" error, please install or upgrade Node.js from the official website: https://nodejs.org/ + +--- + +## **Step 3: Install OpenClaw** + +OpenClaw supports several installation methods. For beginners, the official one-line installation script is the best choice, as it automatically handles most of the setup. + +This method is the quickest and easiest as it will detect your OS, install dependencies, and make the openclaw command available globally. + +For macOS or Linux terminals, execute the following: + +```bash +npm install -g openclaw +``` + +For Windows users using PowerShell, run the same command: + +```bash +npm install -g openclaw +``` + +### **Troubleshooting Common Errors** + +#### **Problem 1: Sharp Module Error** + +On some systems—especially macOS where libvips was installed via Homebrew—you might encounter an error with the sharp module (an image processing library). To fix this, try forcing the installation of pre-built binaries, which bypasses local compilation: + +```bash +npm install -g openclaw --force +``` + +#### **Problem 2: "Command Not Found"** + +After installation, you might see: + +```bash +openclaw: command not found +``` + +This usually means your system can't find where globally installed npm packages are located. Find npm's global installation path by running: + +```bash +npm config get prefix +``` + +If the output is, for example, `/usr/local`, then your binaries live in `/usr/local/bin`. You will need to add this path to your shell profile (`~/.zshrc` or `~/.bashrc`): + +```bash +export PATH="/usr/local/bin:$PATH" +``` + +After saving the file, restart your terminal or run: + +```bash +source ~/.zshrc +``` + +The openclaw command should now work. + +--- + +## **Step 4: Complete the Initialization Wizard** + +After installation, the onboarding wizard should trigger automatically. + +If you accidentally closed the window, you can restart the wizard (and install the background daemon) by running: + +```bash +openclaw onboard +``` + +The wizard will walk you through three primary sections: + +- **AI Model Configuration** + The wizard will request an API key for the large language model service (Anthropic Claude, OpenAI GPT, etc.). + → For now, select **"Skip for now"**. We'll set this up manually in the next step. + +- **Communication Channels** + Choose which messaging apps you want to use to talk to OpenClaw (e.g., Telegram, WhatsApp). + +- **Skills** + We recommend selecting **No** (use the Spacebar to toggle selections and Enter to confirm). You can add them later. + +Once the wizard finishes and the OpenClaw UI launches, you will need to manually edit the configuration file to connect it to Bank of AI. + +--- + +## **Step 5: Configure the Bank of AI Model** + +After completing the onboarding wizard, you'll need to manually add your Bank of AI configuration to OpenClaw and set it as the default model. + +There are two ways to complete the configuration: + +- **One-Click Script**: https://docs.ainft.com/reference/openclaw-ainft-integration-one-click-script-tutorial +- **Manual Configuration**: Follow the instructions below + +--- + +### **5.1 Edit the Configuration File** + +Open the configuration file located at: + +```bash +vim ~/.openclaw/openclaw.json +``` + +OpenClaw reads this file at startup to load all its LLM configurations. + +Locate the `"models"` section and merge the following JSON snippet. Be sure to replace `{BANKOFAI_API_KEY}` with your actual API key from https://chat.bankofai.io/key. + +```json +{ + "models": { + "mode": "merge", + "providers": { + "bankofai": { + "baseUrl": "https://api.bankofai.io/v1/", + "apiKey": "{BANKOFAI_API_KEY}", + "api": "openai-completions", + "models": [ + { + "id": "gpt-5.2", + "name": "gpt-5.2" + }, + { + "id": "gpt-5-mini", + "name": "gpt-5-mini" + }, + { + "id": "gpt-5-nano", + "name": "gpt-5-nano" + }, + { + "id": "claude-opus-4.6", + "name": "claude-opus-4.6" + }, + { + "id": "claude-sonnet-4.6", + "name": "claude-sonnet-4.6" + }, + { + "id": "claude-haiku-4.5", + "name": "claude-haiku-4.5" + } + ] + } + } + } +} +``` + +--- + +### **5.2 Set the Default Model** + +In the same `openclaw.json` file, locate the `agents` section and set the default model: + +```json +{ + "agents": { + "default": { + "model": "bankofai/gpt-5-nano" + } + } +} +``` + +--- + +### **5.3 Restart the Gateway** + +For the configuration changes to take effect, restart the OpenClaw gateway: + +```bash +openclaw gateway restart +``` + +--- + +### **5.4 Test the Connection** + +Send a test message from your terminal: + +```bash +openclaw agent --agent main --message "How are you doing today?" +``` + +If you receive a coherent response, congratulations—you have successfully connected OpenClaw to Bank of AI! + +--- + +## **Step 6: Understand Gateway and Diagnostic Commands** + +If you encounter issues during configuration or while running the program, it helps to understand what the Gateway is and how to use the built-in diagnostic tools. + +### **What is the Gateway?** + +During setup, you'll frequently encounter the term "Gateway". + +| Action | Command | +| --------------------- | -------------------------- | +| Install the Gateway | openclaw gateway install | +| Start the Gateway | openclaw gateway start | +| Stop the Gateway | openclaw gateway stop | +| Restart the Gateway | openclaw gateway restart | +| Uninstall the Gateway | openclaw gateway uninstall | +| Check Gateway Status | openclaw gateway status | + +--- + +### **Diagnostic Commands** + +Once completing onboarding and updating your configuration file, run: + +```bash +openclaw doctor +``` + +You can also check status: + +```bash +openclaw gateway status +``` + +If the Gateway is functioning correctly, it will show a **Healthy** status. + +--- + +## **Step 7: Launch OpenClaw** + +With the configuration complete, you can interact with your AI assistant through either a web dashboard or a terminal interface. + +### **Option 1: Web Dashboard** + +```bash +openclaw dashboard +``` + +Then open: + +``` +http://127.0.0.1:18789 +``` + +You can: + +- Chat with your AI +- View history +- Configure models +- Monitor system status + +--- + +### **Option 2: Terminal UI (TUI)** + +```bash +openclaw tui +``` + +### Commands + +| Command | Description | +| ---------------- | --------------------------------------- | +| /status | View the current system status | +| /session `` | Switch to a specific chat session | +| /model `` | Switch the LLM | +| /help | View available commands | + +--- + +## **Step 8: Master Essential Commands** + +### 1. Check Model Status + +```bash +openclaw models status +``` + +--- + +### 2. Manage Channels + +```bash +openclaw channels list +``` + +--- + +### 3. Search Memory + +```bash +openclaw memory search "keyword" +``` + +--- + +### 4. View Docs + +```bash +openclaw docs +``` + +--- + +# ✅ Done + +You now have a fully working **OpenClaw + Bank of AI private AI agent**. + +You can now: + +- Build automation workflows +- Connect Telegram bots +- Execute on-chain operations +- Create your own AI agent product + +🚀 Welcome to your personal AI infrastructure powered by Bank of AI. diff --git a/docs/llmservice/openclaw/one-click-script-tutorial.md b/docs/llmservice/openclaw/one-click-script-tutorial.md new file mode 100644 index 00000000..6d1375e9 --- /dev/null +++ b/docs/llmservice/openclaw/one-click-script-tutorial.md @@ -0,0 +1,124 @@ +## Quick Start + +**Before running this script, please ensure:** + +1. Node.js 22 or higher is installed +2. OpenClaw is installed and initialized (you have run `openclaw onboard`) +3. Network connection is normal, and the BANK OF AI API is accessible + +### Script Commands + +**Linux & macOS:** + +Mac users: search for "Terminal" in Applications, open it, and enter the command below: + +```bash +curl https://chat.bankofai.io/scripts/openclaw-install-bankofai-provider.sh | bash +``` + +**Windows PowerShell:** + +Windows users: search for "PowerShell" in the Start menu, open it, and enter the command below (CMD is not supported) + +```powershell +iwr https://chat.bankofai.io/scripts/openclaw-install-bankofai-provider.ps1 | iex +``` + +*** + +## Detailed Steps + +### 1. Apply for an API Key + +1. Log in to the [BANK OF AI Platform](https://chat.bankofai.io/) +2. Go to the [API Key Management Page](https://chat.bankofai.io/key) +3. Click to apply for a new API Key + +![](https://files.readme.io/354a3d414f37e7df28f2cbf92dd055db9b67a20cab8738f6d5ac007226b6931b-image.png) + +
+ +*** + +### 2. Run the Installation Script + +Depending on your operating system, execute the corresponding command above. The script will automatically: + +- Check the environment (Node.js, OpenClaw, etc.) +- Prompt you to enter your API Key + +![](https://files.readme.io/ef091efb8911db673af8a5eade7b281a52f641c93d41fbd57cb898ee91893e76-Image_16-3-2026_at_7.00PM.png) + +*** + +### 3. Select a Default Model + +After validating the API Key, the script will fetch the list of available models and prompt you to select a default model: + +![](https://files.readme.io/be3b9162405e38988898261db3effa8adcf92b6d9e37181d36b32c1cf8bcd1e3-Image_16-3-2026_at_7.01PM.png) + +>**Note:** Gemini series models are currently largely unusable in OpenClaw due to client fingerprinting strictness for function calls. Please choose with caution. + +*** + +### 4. Complete Configuration + +Once the selection is complete, the script will automatically: + +- Back up the original configuration +- Update the OpenClaw configuration file +- Restart the Gateway + +![](https://files.readme.io/a08c7fcee0cbe906042ef52fefa15348750ad5cd2db8c4f555f92ecabba72761-Image_16-3-2026_at_7.03PM.png) + +
+ +*** + +### 5. Switch Models + +You can switch the currently used model in two ways: + +- Command Line + +```bash +openclaw models set bankofai/ +``` + +Or manually edit the `~/.openclaw/openclaw.json` configuration file. + +- Web UI - Dashboard + +Visit (18789 is the default port for OpenClaw) in your browser to access the OpenClaw Dashboard. Click "Agent" in the left navigation menu, and select the desired model in the Primary model dropdown: + +> **Note:** Once you change the model via the Dashboard, command-line model switching will no longer work because the Dashboard automatically adds a `list` field to the config file. + +![](https://files.readme.io/3668289b53d185d158dd8393f46c0e171c0d301b5bdc624792c8cf8e6f9c4936-16-3-26_6.56.png) + +
+ +*** + +## Compatibility Testing + +| Operating System | Status | +| :--------------- | :------- | +| Ubuntu 24.04 | ✅ Passed | +| Windows 11 25H2 | ✅ Passed | +| macOS 24.6.0 | ✅ Passed | + +*** + +## FAQ + +**Q: What should I do if the script execution fails?** + +A: Please ensure that: + +1. Node.js 22 or higher is installed +2. OpenClaw is installed and initialized (you have run `openclaw onboard`) +3. Network connection is normal, and the BANK OF AI API is accessible + +**Q: How do I switch models?** + +See point 5 in the **Detailed Steps** above. diff --git a/docs/llmservice/pricing-and-usage.md b/docs/llmservice/pricing-and-usage.md new file mode 100644 index 00000000..11020dc6 --- /dev/null +++ b/docs/llmservice/pricing-and-usage.md @@ -0,0 +1,49 @@ +# Pricing and Usage + +## Credits & Pricing + +The Bank of AI platform utilizes a unified credit system to measure and settle usage for all AI services. + +* **Credit Calculation Rules:** The number of tokens consumed in each interaction with the AI is converted into corresponding credits based on the pricing standards of different models and deducted from your account balance. +* **Token Consumption Details:** In the AI response details, the platform displays the breakdown of token consumption, helping you understand the specific sources of credit usage and optimize future usage patterns. +* **Pricing for Different Models:** Pricing varies due to differences in capabilities and computational costs among AI models. Generally, more capable models consume more credits. The web search feature incurs an additional fee and is charged on a pay-per-use basis. Some models do not support web search marked as "-". + +### Specific Model Pricing + +| Model | Input (Credits/Token) | Output (Credits/Token) | Web Search (Credits/Use) | +| :--- | :--- | :--- | :--- | +| ChatGPT-5.2 | 1.75 | 14.00 | 10,000 | +| ChatGPT-5-mini | 0.25 | 2.00 | 10,000 | +| ChatGPT-5-nano | 0.05 | 0.40 | - | +| Claude Opus 4.6 | 5.00 | 25.00 | 10,000 | +| Claude Opus 4.5 | 5.00 | 25.00 | 10,000 | +| Claude Sonnet 4.6 | 3.00 | 15.00 | 10,000 | +| Claude Sonnet 4.5 | 3.00 | 15.00 | 10,000 | +| Claude Haiku 4.5 | 1.00 | 5.00 | 10,000 | +| Gemini 3.1 Pro | 2.00 | 12.00 | 14,000 | +| Gemini 3 Flash | 0.50 | 3.00 | 14,000 | + +> **Calculation Example:** If you use a model with a rate of 1.25 (Input) and 10.00 (Output) to ask a question (10 input tokens) and receive a response (50 output tokens), the dialogue consumes **512.5 credits** (10 × 1.25 + 50 × 10). You can check the specific usage by hovering over the model name in the bottom right corner of the chat. + +## Usage Information + +You can view detailed data regarding all your consumption on the **Usage** page via the left navigation bar. + +* **Usage Overview:** Displays your credit balance and total consumption for the current month. +* **Monthly Usage Chart:** An intuitive bar chart to track usage fluctuations over the past year. +* **Usage Detail:** Every record corresponds precisely to a single AI interaction, including creation time, model used, token usage, credits consumed, and response time. + +## Deposit + +Bank of AI operates on a pre-paid model. Leveraging secure blockchain technology, the platform offers a convenient deposit experience. + +* **Deposit Process:** On the **Top up** page, the platform will guide you to pay using your connected Web3 wallet. Simply confirm the transaction in the wallet pop-up window to complete it. +* **Supported Token Types:** The platform supports various mainstream tokens on supported networks (including TRON and BNB Chain). +* **Arrival Time:** Once the transaction is confirmed on the blockchain, the system will automatically issue the equivalent value of credits to your account, typically within a few minutes. + +## Billing & Invoices + +View your complete deposit history under the **History** tab on the **Top up** page. + +* **Deposit Records:** Displays creation time, type, transaction hash, and token information for each deposit. +* **Transparency:** You can click the transaction hash to verify details on the corresponding blockchain explorer (e.g., TRONSCAN or BscScan). diff --git a/docs/llmservice/quick-start.md b/docs/llmservice/quick-start.md new file mode 100644 index 00000000..250895be --- /dev/null +++ b/docs/llmservice/quick-start.md @@ -0,0 +1,32 @@ +# Quick Start + +This chapter will guide you through the initial setup of the LLM Service within the Bank of AI platform. Follow these steps to kickstart your AI journey. + +## 1. Connect Wallet +LLM Service utilizes a decentralized login method. You can use supported Web3 wallets to connect and authorize the login. This method is secure and convenient, eliminating the need to remember complex usernames and passwords. + +### How to Login: +1. Visit the [Bank of AI Chat platform](https://chat.bankofai.io/chat). +2. Click the **Log in** button in the top right corner of the page. +3. In the pop-up window, select your wallet provider and authorize the connection. +4. Confirm the signature request within your wallet to complete the login. + +Upon successful login, your wallet address will be displayed in the top right corner, indicating that you have successfully accessed the platform. + +*Note: Please ensure you have a compatible Web3 wallet installed in your browser or mobile device. Always store your mnemonic phrase and private key securely.* + +## 2. Start Your First Conversation +Once inside the platform, you can immediately start interacting with the AI through our LLM Service. + +* **Select AI Model:** In the chat interface, you will see the current default AI model. Click the model name to expand the list and select the model (e.g., GPT-4o, Claude 3.5, Gemini) you wish to use. +* **Send Message:** In the input box at the bottom of the page, enter your prompt or task, then click the send button or press Enter. +* **Contextual Interaction:** The AI's response will be presented in a dialogue format. You can engage in multi-turn conversations, and the AI will respond based on the context. + +## 3. Credits and Usage +The LLM Service operates on a credit-based system. You can obtain credits to use AI services through seamless on-chain transactions. + +* **Top-up Interface:** Navigate to the **Top up** section in the dashboard to manage your balance. +* **Purchase Credits:** The platform supports mainstream tokens on supported blockchains. Simply confirm the transaction in your wallet to transfer tokens to the designated address. +* **Automatic Crediting:** Once the transaction is confirmed on the blockchain, the system will automatically credit the corresponding value to your account balance. + +After completing these steps, you can enjoy streamlined access to various industry-leading AI models. diff --git a/docusaurus.config.js b/docusaurus.config.js index 82003ae1..5e8fd252 100644 --- a/docusaurus.config.js +++ b/docusaurus.config.js @@ -5,7 +5,7 @@ require('dotenv').config() module.exports = { title: 'BANK OF AI | Developer Guide', tagline: 'HTTP 402 Payment Protocol for TRON', - url: 'https://docs.x402-tron.org/', + url: 'https://docs.bankofai.io', baseUrl: '/', trailingSlash: true, i18n: { @@ -97,17 +97,6 @@ module.exports = { ], plugins: [ require.resolve('./docusaurus-plugin-global-style'), - // [ - // '@docusaurus/plugin-client-redirects', - // { - // redirects: [ - // { - // from: '/overview', - // to: '/', - // }, - // ], - // }, - // ], function webpackFallbackPlugin() { return { name: 'custom-webpack-fallback-plugin', diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current.json b/i18n/zh-Hans/docusaurus-plugin-content-docs/current.json index 92a49a53..4de059a6 100644 --- a/i18n/zh-Hans/docusaurus-plugin-content-docs/current.json +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current.json @@ -43,5 +43,29 @@ "sidebar.docsSidebar.doc.OverView": { "message": "简介", "description": "The label for the doc item OverView in sidebar docsSidebar" + }, + "sidebar.docsSidebar.category.LLM Service": { + "message": "LLM 服务", + "description": "The label for category LLM Service in sidebar docsSidebar" + }, + "sidebar.docsSidebar.doc.Quick Start": { + "message": "快速入门", + "description": "The label for the doc item Quick Start in sidebar docsSidebar" + }, + "sidebar.docsSidebar.doc.Pricing and Usage": { + "message": "定价与用量", + "description": "The label for the doc item Pricing and Usage in sidebar docsSidebar" + }, + "sidebar.docsSidebar.category.Models": { + "message": "模型列表", + "description": "The label for category Models in sidebar docsSidebar" + }, + "sidebar.docsSidebar.category.OpenClaw": { + "message": "OpenClaw 集成", + "description": "The label for category OpenClaw in sidebar docsSidebar" + }, + "sidebar.docsSidebar.category.API": { + "message": "API 参考", + "description": "The label for category API in sidebar docsSidebar" } } diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/api/API.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/api/API.md new file mode 100644 index 00000000..b03f16f3 --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/api/API.md @@ -0,0 +1,221 @@ +# Bank of AI LLM API(兼容 OpenAI) + +本接口与 **OpenAI API 完全兼容**,你可以直接复用现有 OpenAI SDK,仅需替换 Base URL 即可接入。 + +支持接口: + +- `/v1/chat/completions`:对话生成 +- `/v1/models`:获取模型列表 + +--- + +## 基本信息 + +- **版本:** 1.0 +- **协议:** https +- **Host:** api.bankofai.io + +--- + +# /v1/chat/completions + +## POST + +### 功能说明 +创建对话补全(Chat Completion),接口格式与 OpenAI 一致。 + +### 认证方式 +在 Header 中传入: + +``` +Authorization: Bearer sk-xxxx +``` + +--- + +### 返回方式 + +- **普通模式**:返回完整 JSON(内容在 `choices[].content`) +- **流式模式**:SSE 流(内容在 `choices[].delta.content`) + +--- + +## 请求参数 + +| 参数 | 位置 | 说明 | 必填 | +|-----|-----|-----|-----| +| Authorization | header | Bearer Token,例如 `Bearer sk-xxx` | 是 | +| body | body | 请求体(必须包含 model 和 messages) | 是 | + +--- + +## 请求体字段说明(body) + +| 字段 | 类型 | 说明 | +|----|----|----| +| model | string | 模型 ID,例如 `gpt-4` | +| messages | array | 对话消息列表 | +| max_tokens | integer | 最大生成 token 数 | +| temperature | number | 随机性(0~2) | +| top_p | number | 核采样 | +| n | integer | 返回结果数量 | +| stream | boolean | 是否开启流式输出 | +| stop | string / array | 停止词 | +| presence_penalty | number | 主题惩罚 | +| frequency_penalty | number | 频率惩罚 | +| tools | array | 可调用工具 | +| tool_choice | string / object | 工具调用策略 | +| response_format | object | 输出格式 | +| user | string | 用户标识 | + +--- + +## 返回结果 + +| 状态码 | 说明 | +|------|------| +| 200 | 成功返回结果(普通 / 流式) | +| 401 | 认证失败 | + +--- + +# /v1/models + +## GET + +### 功能说明 +获取当前可用模型列表。 + +### 认证方式 + +``` +Authorization: Bearer sk-xxxx +``` + +--- + +## 返回结果 + +| 状态码 | 说明 | +|------|------| +| 200 | 返回模型列表 | +| 401 | 认证失败 | + +返回结构: + +```json +{ + "object": "list", + "success": true, + "data": [ + { + "id": "gpt-4", + "object": "model", + "created": 1626777600, + "owned_by": "openai" + } + ] +} +``` + +--- + +# 数据结构说明 + +## ChatChoice(生成结果) + +| 字段 | 类型 | 说明 | +|----|----|----| +| content | string | AI 返回内容 | +| finish_reason | string | 结束原因,例如 `stop` | +| index | integer | 结果序号 | + +--- + +## ChatMessage(对话消息) + +| 字段 | 类型 | 说明 | +|----|----|----| +| role | string | system / user / assistant / tool | +| content | string | 消息内容 | +| name | string | 可选用户名 | +| tool_call_id | string | 工具调用 ID | +| tool_calls | array | 工具调用信息 | + +--- + +## ChatCompletionsResponse(响应结构) + +| 字段 | 类型 | 说明 | +|----|----|----| +| id | string | 请求 ID | +| object | string | 类型 | +| created | integer | 时间戳 | +| model | string | 模型名称 | +| choices | array | 生成结果 | +| usage | object | Token 使用统计 | + +--- + +## Token 使用统计(Usage) + +| 字段 | 类型 | 说明 | +|----|----|----| +| prompt_tokens | integer | 输入 token | +| completion_tokens | integer | 输出 token | +| total_tokens | integer | 总 token | + +--- + +## 工具调用(Function Calling) + +支持 OpenAI 标准工具调用: + +### 工具定义 + +```json +{ + "type": "function", + "function": { + "name": "get_weather", + "description": "获取天气", + "parameters": {} + } +} +``` + +--- + +### 模型调用返回 + +```json +{ + "tool_calls": [ + { + "id": "call_xxx", + "type": "function", + "function": { + "name": "get_weather", + "arguments": "{ \"city\": \"Beijing\" }" + } + } + ] +} +``` + +--- + +# 总结 + +- 完全兼容 OpenAI API +- 支持流式输出 +- 支持工具调用(Function Calling) +- 支持多模型统一接入 + +只需将 Base URL 替换为: + +``` +https://api.bankofai.io/v1 +``` + +即可快速接入 Bank of AI 🚀 diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/introduction.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/introduction.md new file mode 100644 index 00000000..995f9af8 --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/introduction.md @@ -0,0 +1,19 @@ +# 欢迎使用 LLM Service + +## 关于 LLM Service + +LLM Service 是 Bank of AI 生态中的专业 AI 服务模块,基于顶级区块链基础设施构建,致力于为用户提供高效、易用且富有创造力的 AI 交互体验。作为 Bank of AI 的核心 AI 基础设施,该服务结合区块链的去中心化、安全性与高效率,打造全新的 AI 服务模式。 + +核心功能包括: + +- **多模型 AI 对话:** 集成多种行业领先的大语言模型(LLMs),用户可根据需求自由选择最合适的模型。 +- **强大的 AI 服务能力:** 提供完整的 AI API 服务,支持在 Bank of AI 框架内快速接入与集成。 +- **原生 Web3 体验:** 无缝集成主流 Web3 钱包,实现从登录到支付的一体化原生体验。 + +## 为什么选择 LLM Service? + +选择 LLM Service,即可同时享受区块链生态优势与高质量 AI 服务体验: + +- **多链生态优势:** 支持多链主流代币支付,具备快速确认与低手续费优势。 +- **低成本高效率:** 通过资源优化与高效链上交互,实现更具性价比的 AI 服务。 +- **安全与隐私保护:** 基于去中心化登录机制,用户仅需通过 Web3 钱包签名即可完成认证,保障交互过程的安全与隐私。 diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /chatgpt-5-2.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /chatgpt-5-2.md new file mode 100644 index 00000000..afdfbb6d --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /chatgpt-5-2.md @@ -0,0 +1,41 @@ +# ChatGPT-5.2 + +## 概述 + +ChatGPT-5.2 是 OpenAI 最新一代旗舰大语言模型。在 5.1 的基础上,进一步优化了多模态处理速度与复杂任务执行效率,适合追求高性能与高效率的专业用户。 + +--- + +## 核心特性 + +- **高效多模态处理:** 相比 5.1,大幅提升图像与视频内容的解析与生成速度,交互更流畅 +- **任务执行效率提升:** 优化推理引擎,在处理多步骤复杂任务时更快、更准确 +- **更强抗干扰能力:** 在噪声较多或指令模糊的情况下,依然保持较高准确性 + +--- + +## 适用场景 + +- **实时数据分析与可视化:** 快速处理数据流并生成复杂图表与分析报告 +- **复杂项目管理与规划:** 支持任务拆解、资源分配与风险评估 +- **高频高精度专业咨询:** 适用于金融分析、法律检索等需要快速准确响应的场景 + +--- + +## 能力与限制 + +| 能力 | 说明 | +| :--- | :--- | +| **推理能力** | 极强。在复杂逻辑推理与科学计算方面保持领先,并提升效率 | +| **创造能力** | 极强。可生成高质量、结构化的专业内容 | +| **多模态能力** | 全面且高效。支持图像、视频、音频输入与理解,并可生成高质量图像 | +| **响应速度** | 中等偏慢。相比 5.1 有提升,但仍属于深度推理模型 | +| **上下文窗口** | 超大。支持百万级 Token 上下文 | + +--- + +## 积分与定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | +| :--- | :--- | :--- | +| **ChatGPT-5.2** | 1.75 | 14.00 | diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /chatgpt-5-mini.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /chatgpt-5-mini.md new file mode 100644 index 00000000..dd550add --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /chatgpt-5-mini.md @@ -0,0 +1,41 @@ +# ChatGPT-5-mini + +## 概述 + +ChatGPT-5-mini 是一款高效、经济的轻量级语言模型,针对日常对话与通用任务进行了优化,是 Bank of AI 生态中性价比极高的选择。 + +--- + +## 核心特性 + +- **极快响应:** 针对低延迟优化,提供接近实时的对话体验 +- **高性价比:** 在保证输出质量的同时,大幅降低计算成本 +- **稳定通用能力:** 覆盖日常问答、文本处理等多种场景,表现稳定 + +--- + +## 适用场景 + +- **日常对话与快速问答:** 适合作为智能助手,快速回答问题与聊天 +- **文本处理:** 支持摘要、润色、格式整理、关键词提取 +- **内容初稿生成:** 快速生成社交媒体文案、产品描述、博客草稿 + +--- + +## 能力与限制 + +| 能力 | 说明 | +| :--- | :--- | +| **推理能力** | 中等。可处理基础逻辑,但复杂多步骤问题表现有限 | +| **创造能力** | 中等。文本流畅,但深度与专业性有限 | +| **多模态能力** | 不支持。仅支持文本处理 | +| **响应速度** | 快。平台中响应速度较快的模型之一 | +| **上下文窗口** | 标准。支持数万级 Token,满足大多数日常场景 | + +--- + +## 积分与定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | +| :--- | :--- | :--- | +| **ChatGPT-5-mini** | 0.25 | 2.00 | diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /chatgpt-5-nano.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /chatgpt-5-nano.md new file mode 100644 index 00000000..25620e8d --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /chatgpt-5-nano.md @@ -0,0 +1,41 @@ +# ChatGPT-5-nano + +## 概述 + +ChatGPT-5-nano 是一款在性能、速度和成本之间取得良好平衡的语言模型。在 Bank of AI 生态中,它以适中的成本提供接近专业级的 AI 能力。 + +--- + +## 核心特性 + +- **推理能力增强:** 相比更轻量的模型,在逻辑推理、代码生成和多语言处理方面有明显提升 +- **高效性能:** 在保持较高输出质量的同时,维持较快响应速度 +- **多功能能力:** 可处理多种任务类型,是开发者与内容创作者的实用助手 + +--- + +## 适用场景 + +- **代码辅助与调试:** 支持多语言代码理解与生成,帮助调试和编写技术文档 +- **多语言翻译与写作:** 提供高质量跨语言翻译,并生成自然流畅的内容 +- **结构化内容生成:** 生成格式规范的报告、技术文档、教程等 + +--- + +## 能力与限制 + +| 能力 | 说明 | +| :--- | :--- | +| **推理能力** | 强。能够处理复杂逻辑问题与编程任务,在特定领域表现优秀 | +| **创造能力** | 强。可生成富有创意且内容深入的文本 | +| **多模态能力** | 有限支持。可理解和描述简单图像,但不支持深度多模态分析 | +| **响应速度** | 中等。快于旗舰模型,但略慢于 mini 模型 | +| **上下文窗口** | 大。支持数十万级 Token 上下文,适合长文档处理 | + +--- + +## 积分与定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | +| :--- | :--- | :--- | +| **ChatGPT-5-nano** | 0.05 | 0.40 | diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-haiku-4-5.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-haiku-4-5.md new file mode 100644 index 00000000..3bf9282f --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-haiku-4-5.md @@ -0,0 +1,41 @@ +# Claude Haiku 4.5 + +## 概述 + +Claude Haiku 4.5 是 Anthropic 推出的轻量高速模型,已集成至 Bank of AI 平台。主打极低延迟与高并发,适用于实时交互场景。 + +--- + +## 核心特性 + +- **极致响应速度:** 超低延迟,适合即时交互应用 +- **高性价比:** 成本优势明显,适合大规模部署 +- **企业级稳定性:** 通过严格安全测试,满足企业级可靠性要求 + +--- + +## 适用场景 + +- **实时聊天与内容审核:** 提供流畅对话体验,快速处理用户内容 +- **移动端 AI 应用:** 针对低延迟与资源受限场景优化 +- **工作流自动化:** 邮件分类、会议总结、表单信息提取等 + +--- + +## 能力与限制 + +| 能力 | 说明 | +| :--- | :--- | +| **推理能力** | 中等。可处理常规任务,但复杂多步骤问题能力有限 | +| **创造能力** | 中等。文本简洁流畅,更适合信息表达 | +| **多模态能力** | 支持。具备基础图像理解与描述能力 | +| **响应速度** | 极快。平台中响应最快的模型之一 | +| **上下文窗口** | 超大。支持长上下文与大规模文本处理 | + +--- + +## 积分与定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | +| :--- | :--- | :--- | +| **Claude Haiku 4.5** | 1.00 | 5.00 | diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-opus-4-5.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-opus-4-5.md new file mode 100644 index 00000000..1c678bad --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-opus-4-5.md @@ -0,0 +1,41 @@ +# Claude Opus 4.5 + +## 概述 + +Claude Opus 4.5 是 Anthropic 推出的旗舰级 AI 模型,已集成至 Bank of AI 平台。以顶级智能、卓越性能和强伦理能力著称,适用于高端专业场景。 + +--- + +## 核心特性 + +- **顶级智能水平:** 在复杂推理、数学与编程等任务中表现领先,可解决高度复杂问题 +- **强大视觉理解:** 能准确解析复杂图表、科研图像及多图文档 +- **超长上下文处理:** 支持超大上下文窗口,可处理长文档及大型代码库 + +--- + +## 适用场景 + +- **复杂系统分析:** 理解流程图、API 文档,并自动生成跨系统代码 +- **高端科研分析:** 解读论文、分析实验数据、提出研究假设 +- **企业战略决策:** 分析市场、财务与法律信息,提供决策支持 + +--- + +## 能力与限制 + +| 能力 | 说明 | +| :--- | :--- | +| **推理能力** | 顶级。在复杂逻辑与深度思考任务中表现领先 | +| **创造能力** | 极强。可生成逻辑严密、富有洞察力的高质量文本 | +| **多模态能力** | 强。擅长复杂视觉信息理解与分析 | +| **响应速度** | 较慢。优先保证深度与质量 | +| **上下文窗口** | 超大。支持 200K+ Token,上限可扩展 | + +--- + +## 积分与定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | +| :--- | :--- | :--- | +| **Claude Opus 4.5** | 5.00 | 25.00 | diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-opus-4-6.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-opus-4-6.md new file mode 100644 index 00000000..9637dd92 --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-opus-4-6.md @@ -0,0 +1,43 @@ +# Claude Opus 4.6 + +## 概述 + +Claude Opus 4.6 是 Anthropic 旗舰模型系列的最新版本。在 4.5 强大智能与伦理能力的基础上,进一步强化复杂推理、多模态理解以及专业领域表现,为 Bank of AI 生态提供更精准、更强大的 AI 能力。 + +--- + +## 核心特性 + +- **复杂推理能力提升:** 针对多步骤、跨领域与抽象问题进行深度优化,适用于科学、工程与法律等复杂任务 +- **多模态理解增强:** 对图像、图表与视频的信息捕捉与上下文关联能力更强 +- **专业知识深化:** 在金融分析、药物研发、复杂系统设计等领域能力进一步增强 +- **相较 4.5 的提升:** 4.6 更注重 **深度与精度**,在高准确度任务中提供更可靠结果 + +--- + +## 适用场景 + +- **前沿科研分析:** 复杂数据分析、理论验证与新发现探索 +- **高级战略咨询:** 基于大规模信息进行深度分析与决策支持 +- **法律与合规审查:** 处理复杂法律文本并进行风险评估 +- **高端内容生成:** 撰写研究报告、技术白皮书与深度行业分析 + +--- + +## 能力与限制 + +| 能力 | 说明 | +| :--- | :--- | +| **推理能力** | 顶级。相比 4.5 进一步提升,擅长深度推理与抽象逻辑 | +| **创造能力** | 极强。可生成逻辑严谨、见解深入的专业文本 | +| **多模态能力** | 增强。更擅长分析复杂视觉信息与多模态数据 | +| **响应速度** | 较慢。优先保证输出质量与分析深度 | +| **上下文窗口** | 超大。支持超长上下文,用于处理大型文档与代码库 | + +--- + +## 积分与定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | +| :--- | :--- | :--- | +| **Claude Opus 4.6** | 5.00 | 25.00 | diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-sonnet-4-5.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-sonnet-4-5.md new file mode 100644 index 00000000..ba39935e --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-sonnet-4-5.md @@ -0,0 +1,41 @@ +# Claude Sonnet 4.5 + +## 概述 + +Claude Sonnet 4.5 是 Anthropic 推出的平衡型 AI 模型,已集成至 Bank of AI 平台。在接近旗舰级智能的同时提供更低成本,是企业规模化部署 AI 的理想选择。 + +--- + +## 核心特性 + +- **智能与速度兼顾:** 智能水平显著高于同类模型,响应速度约为 Opus 系列的两倍 +- **企业场景优化:** 针对知识检索、销售自动化和复杂数据分析等核心业务场景进行优化 +- **超长上下文与高召回:** 支持超大上下文窗口,在长文档中仍保持高信息召回率 + +--- + +## 适用场景 + +- **企业知识管理:** 从海量文档中快速检索信息,为客户或员工提供准确答案 +- **销售与营销自动化:** 分析市场数据、生成个性化营销文案、处理销售线索 +- **代码质量管理:** 高效生成与审查代码,帮助开发团队提升效率与质量 + +--- + +## 能力与限制 + +| 能力 | 说明 | +| :--- | :--- | +| **推理能力** | 强。可满足大多数商业与专业应用场景 | +| **创造能力** | 强。生成符合商业与专业标准的高质量文本 | +| **多模态能力** | 支持。具备较强图像理解能力,可处理包含图表的文档 | +| **响应速度** | 中等。在智能与速度之间取得良好平衡 | +| **上下文窗口** | 超大。与 Opus 系列相同,支持长上下文处理 | + +--- + +## 积分与定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | +| :--- | :--- | :--- | +| **Claude Sonnet 4.5** | 3.00 | 15.00 | diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-sonnet-4-6.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-sonnet-4-6.md new file mode 100644 index 00000000..3fe6c3eb --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /claude-sonnet-4-6.md @@ -0,0 +1,43 @@ +# Claude Sonnet 4.6 + +## 概述 + +Claude Sonnet 4.6 是 Anthropic Sonnet 系列的重要升级版本,在更低成本下提供接近 Opus 级别的性能。该版本在编程能力、计算机操作、长上下文推理和知识工作方面全面提升,并首次引入 **100 万 Token 上下文窗口(Beta)**,非常适合企业级自动化与复杂任务处理。 + +--- + +## 核心特性 + +- **接近 Opus 级性能:** 在多项基准测试中表现接近 Opus 4.6,尤其在企业文档处理(图表、PDF、表格)方面表现突出 +- **100 万 Token 上下文:** Sonnet 系列首次支持百万级上下文,可处理完整代码库或大型研究文档 +- **优秀编程与计算机操作能力:** 在复杂代码修复和多步骤网页工作流中表现接近甚至达到人类水平 +- **相较 4.5 的提升:** 指令遵循更强、幻觉更少,多步骤任务执行更加可靠 + +--- + +## 适用场景 + +- **企业文档分析:** 高效处理包含图表和表格的复杂 PDF 文档 +- **大型代码库重构:** 利用百万级上下文对整个代码库进行分析与修改 +- **自动化 Agent 工作流:** 适用于需要多步骤规划的自动化任务 +- **知识工作与设计:** 生成高质量内容和前端页面,减少反复迭代 + +--- + +## 能力与限制 + +| 能力 | 说明 | +| :--- | :--- | +| **推理能力** | 强。在长上下文推理与复杂问题处理方面接近 Opus | +| **创造能力** | 强。尤其擅长生成高质量代码与具有设计感的前端页面 | +| **多模态能力** | 支持。可处理图像、PDF 等多模态输入 | +| **响应速度** | 中等。在性能与速度之间保持良好平衡 | +| **上下文窗口** | 100 万 Token(Beta),最大输出 128,000 Token | + +--- + +## 积分与定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | +| :--- | :--- | :--- | +| **Claude Sonnet 4.6** | 3.00 | 15.00 | diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /gemini-3-1-pro.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /gemini-3-1-pro.md new file mode 100644 index 00000000..98d295d1 --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /gemini-3-1-pro.md @@ -0,0 +1,42 @@ +# Gemini 3.1 Pro + +## 概述 + +Gemini 3.1 Pro 是 Google 旗舰多模态模型的重要升级版本,在推理能力、输出可靠性和效率方面全面提升。在保留原生多模态能力的同时,显著增强复杂任务处理的准确性,并解决了旧版本输出截断的问题。 + +--- + +## 核心特性 + +- **推理能力大幅提升:** 在 ARC-AGI-2 基准测试中取得突破性成绩,复杂模式识别能力较前代提升超过一倍 +- **超长上下文与输出:** 支持最高 **100 万 Token 上下文**,单次输出可达 **64,000 Token**,避免长内容被截断 +- **原生多模态能力:** 支持文本、图像、音频、视频的统一理解与处理 +- **更高运行效率:** 优化 Token 使用,在复杂任务中提供更稳定、更高质量输出 + +--- + +## 适用场景 + +- **复杂代码生成:** 适用于需要深度推理和完整代码输出的软件开发任务 +- **大规模数据分析:** 一次性处理并分析大量文档或数据集 +- **长文本内容创作:** 撰写报告、剧本或技术白皮书,保持长篇逻辑一致性 + +--- + +## 能力与限制 + +| 能力 | 说明 | +| :--- | :--- | +| **推理能力** | 极强。在抽象推理与新问题解决方面显著提升 | +| **创造能力** | 极强。可生成高质量、长篇且逻辑连贯的内容 | +| **多模态能力** | 原生支持。可无缝处理文本、图像、音频、视频混合输入 | +| **响应速度** | 中等偏慢。优先保证输出质量与完整性 | +| **上下文窗口** | 100 万 Token(最大输出 64,000 Token) | + +--- + +## 积分与定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | +| :--- | :--- | :--- | +| **Gemini 3.1 Pro** | 2.00 | 12.00 | diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /gemini-3-flash.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /gemini-3-flash.md new file mode 100644 index 00000000..e850ddd4 --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /gemini-3-flash.md @@ -0,0 +1,41 @@ +# Gemini 3 Flash + +## 概述 + +Gemini 3 Flash 是 Google Gemini 3 系列中速度最快、效率最高的模型。针对低延迟与高并发场景优化,在保留原生多模态能力的同时,大幅提升响应速度并降低成本。 + +--- + +## 核心特性 + +- **极速响应与高吞吐:** 适用于低延迟、高并发场景,是实时 AI 应用的首选 +- **高效多模态:** 支持图像、音频等多模态处理,相比 Pro 版本计算成本更低 +- **高性价比:** 在保证速度与多模态能力的同时显著降低使用成本 + +--- + +## 适用场景 + +- **实时聊天机器人:** 提供流畅、即时的多模态交互体验 +- **内容审核:** 快速识别与过滤文本和图像中的违规内容 +- **移动与边缘应用:** 适用于对延迟敏感、资源受限的场景 + +--- + +## 能力与限制 + +| 能力 | 说明 | +| :--- | :--- | +| **推理能力** | 强。可处理大多数通用与复杂任务,但深度略低于 Pro | +| **创造能力** | 强。可快速生成高质量文本与多模态描述 | +| **多模态能力** | 原生支持且高效。侧重速度而非深度分析 | +| **响应速度** | 极快。平台中响应速度最快的模型之一 | +| **上下文窗口** | 超大。与 Pro 版本一致,支持长上下文 | + +--- + +## 积分与定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | +| :--- | :--- | :--- | +| **Gemini 3 Flash** | 0.50 | 3.00 | diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /glm-5.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /glm-5.md new file mode 100644 index 00000000..6b14389b --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /glm-5.md @@ -0,0 +1,50 @@ +# GLM-5 + +## 概述 + +**GLM-5** 是智谱 AI 推出的新一代旗舰基础模型,专为 **编程(Coding)** 与 **Agent** 场景设计。在开源复杂系统工程与长周期任务中达到 SOTA 表现,实际编码能力接近 **Claude Opus** 水平。 + +基于 **744B** 规模模型,结合异步强化学习与稀疏注意力机制,实现从“写代码”到“构建系统”的能力跃迁。 + +--- + +## 核心特性 + +- **大规模参数与数据:** 模型规模达 **744B(激活 40B)**,预训练数据 **28.5T**,显著提升知识广度与深度 +- **超长上下文与输出:** 支持 **200K Token 上下文**,最大输出 **128K Token**,适合复杂代码与多步骤任务 +- **强大编程与 Agent 能力:** 编程能力系统性增强,低幻觉率,高 Token 利用效率 +- **多种思维模式:** 支持不同推理模式,提升问题解决的灵活性与深度 + +--- + +## 适用场景 + +1. **复杂系统工程:** 软件系统构建、架构设计与优化 +2. **长周期 Agent 任务:** 多步骤规划与执行的自动化流程 +3. **高精度代码调试:** 提供接近人类水平的编程辅助 +4. **大规模文档分析:** 海量文档的信息提取与总结 + +--- + +## 能力与限制 + +| 能力 | 说明 | +| :--- | :--- | +| **推理能力** | 极强。擅长复杂逻辑推理与多步骤规划 | +| **创造能力** | 极强。尤其在代码生成与系统设计方面表现突出 | +| **多模态能力** | 以文本与代码为主,可结合智谱平台视觉能力 | +| **响应速度** | 30–50 tokens/s,在质量与速度之间取得平衡 | +| **上下文窗口** | 200K Token | +| **最大输出** | 128K Token | + +--- + +## 积分与定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | +| :--- | :--- | :--- | +| **GLM-5** | 0.30 | 2.55 | + +--- + +> **说明:** 在编程场景中建议提供清晰的系统提示词,并充分利用 128K 输出能力构建完整模块。 diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /kimi-k2.5.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /kimi-k2.5.md new file mode 100644 index 00000000..329eae0b --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /kimi-k2.5.md @@ -0,0 +1,51 @@ +# Kimi-K2.5 + +## 概述 + +**Kimi-K2.5** 是 Moonshot AI 推出的多功能旗舰模型,采用 **原生多模态架构**,同时支持图像与文本输入、思考模式与快速模式,以及对话与 Agent 任务。 + +凭借 **256K 超长上下文窗口**、多模态理解能力以及强大的 Tool Calling 能力,Kimi-K2.5 为视觉编程与多 Agent 系统提供强大支持,帮助开发者构建下一代 AI 应用。 + +--- + +## 核心特性 + +- **原生多模态架构:** 支持图像与文本混合输入,在视觉理解与视觉编程方面表现突出 +- **256K 超长上下文:** 支持 **256,000 Token** 上下文,适合长文本推理与大规模数据处理 +- **Agent 集群与工具调用:** 支持 Agent 集群预览功能(最多 **100 个子 Agent** 与 **1,500 次工具调用**),速度比单 Agent 快约 4.5 倍 +- **强大编程能力:** 在 SWE-Bench 与 LiveCodeBench 等基准测试中表现领先 +- **思考模式:** 可在快速响应模式与深度推理模式之间灵活切换 + +--- + +## 适用场景 + +1. **视觉编程与自动化:** 网页像素级复刻与办公自动化任务 +2. **超长文本分析:** 法律文档审查、大规模研究报告分析、完整代码库理解 +3. **多 Agent 协作:** 构建复杂自动化流程与多 Agent 协同系统 +4. **专业代码生成:** 高效生成、优化与调试代码 + +--- + +## 能力与限制 + +| 能力 | 说明 | +| :--- | :--- | +| **推理能力** | 极强。擅长长上下文推理与 Agent 任务规划 | +| **创造能力** | 极强。尤其在视觉编程与多模态内容生成方面表现突出 | +| **多模态能力** | 原生多模态。视觉理解能力优秀 | +| **响应速度** | 快速模式响应迅速;Agent 集群模式支持高效并行 | +| **上下文窗口** | 256,000 Token | +| **最大输出** | 256,000 Token | + +--- + +## 积分与定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | +| :--- | :--- | :--- | +| **Kimi-K2.5** | 0.23 | 3.00 | + +--- + +> **提示:** 在复杂系统设计任务中,建议先使用 **思考模式**进行架构规划,再切换到标准模式执行代码生成。 diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /minimax-m2.5.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /minimax-m2.5.md new file mode 100644 index 00000000..f1ab94be --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/models /minimax-m2.5.md @@ -0,0 +1,51 @@ +# MiniMax-M2.5 + +## 概述 + +**MiniMax-M2.5** 是 MiniMax 自主研发的旗舰多模态通用大模型,专为 **高吞吐、低延迟** 的生产环境设计。 + +该模型在编程与 Agent 能力方面表现领先,可原生理解和生成多种模态内容,包括 **文本、音频、图像、视频和音乐**。M2.5 在保持顶级性能的同时大幅降低成本,特别适用于复杂任务处理与专业办公场景。 + +--- + +## 核心特性 + +- **领先的编程与 Agent 能力:** 在 **Multi-SWE-Bench** 基准测试中表现领先,决策能力更成熟,Token 使用效率更高 +- **高效多模态处理:** 原生支持文本、音频、图像、视频与音乐的融合处理 +- **超长上下文能力:** 支持 **197K Token 上下文窗口**,通过强化学习优化复杂任务拆解 +- **高吞吐低延迟:** 提供 **100 TPS** 与 **50 TPS** 两种版本,适合高并发生产环境 +- **办公场景增强:** 在 Word、PPT、Excel 财务建模等专业办公任务中能力显著提升 + +--- + +## 适用场景 + +1. **企业自动化工作流:** 适合需要多模态处理与 Agent 决策的自动化系统 +2. **软件开发与代码辅助:** 在大型复杂代码库中提供高效生成与调试能力 +3. **多模态内容创作:** 支持跨模态内容生成,例如文本生成视频或图像生成音乐 +4. **办公文档处理:** 高效处理 Word、Excel、PPT 等文档的信息提取与分析 + +--- + +## 能力与限制 + +| 能力 | 说明 | +| :--- | :--- | +| **推理能力** | 极强。适合多步骤 Agent 任务与复杂决策 | +| **创造能力** | 极强。擅长多模态创作与办公自动化 | +| **多模态能力** | 原生多模态。支持文本、音频、图像、视频和音乐 | +| **响应速度** | 极快。提供 100 TPS 与 50 TPS 高吞吐版本 | +| **上下文窗口** | 197,000 Token | +| **最大输出** | 131,000 Token | + +--- + +## 积分与定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | +| :--- | :--- | :--- | +| **MiniMax-M2.5** | 0.30 | 1.20 | + +--- + +> **提示:** 对于需要处理大量请求的生产环境(例如每小时处理成千上万份文档),MiniMax-M2.5 在速度与成本之间提供了极佳平衡。 diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/openclaw /integration-guide.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/openclaw /integration-guide.md new file mode 100644 index 00000000..12ec1fd1 --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/openclaw /integration-guide.md @@ -0,0 +1,228 @@ +# 集成 OpenClaw 与 Bank of AI +## 从零到私人 AI Agent:15 分钟部署 OpenClaw + Bank of AI + +OpenClaw(原名 ClawdBot / Moltbot)是一个开源的个人 AI 助手。与云端 SaaS 工具不同,OpenClaw 运行在本地设备上,让你完全掌控数据与工作流程。 + +你可以通过 WhatsApp、Telegram、Lark、钉钉等消息平台与其交互,用于处理邮件、管理日历、编写代码,甚至自动化智能家居。 + +OpenClaw 不只是聊天工具,而是具备实际执行能力的 **AI Agent**,支持: + +- 持久化记忆 +- 本地文件访问 +- 互联网访问 +- 技能扩展 + +本指南将带你完成安装、配置,并连接 Bank of AI API,快速搭建属于你的 AI Agent。 + +--- + +# Step 1 获取 Bank of AI API Key + +1. 打开:https://chat.bankofai.io/chat +2. 登录账户 +3. 进入 API Key 管理页面 +4. 创建并复制 API Key + +--- + +# Step 2 准备系统环境 + +| 要求 | 说明 | +|---|---| +| Node.js | ≥ v22 | +| 系统 | macOS / Linux / Windows(WSL2) | +| 包管理 | npm | + +检查版本: + +```bash +node -v +``` + +若版本过低,请前往:https://nodejs.org/ + +--- + +# Step 3 安装 OpenClaw + +```bash +npm install -g openclaw +``` + +--- + +## 常见问题 + +### Sharp 报错 + +```bash +npm install -g openclaw --force +``` + +### 命令找不到 + +```bash +npm config get prefix +``` + +例如输出: + +``` +/usr/local +``` + +加入环境变量: + +```bash +export PATH="/usr/local/bin:$PATH" +source ~/.zshrc +``` + +--- + +# Step 4 初始化 + +```bash +openclaw onboard +``` + +操作建议: + +- 模型配置:选择 **Skip for now** +- 渠道:可跳过 +- Skills:选择 No + +--- + +# Step 5 配置 Bank of AI + +## 5.1 编辑配置文件 + +```bash +vim ~/.openclaw/openclaw.json +``` + +加入: + +```json +{ + "models": { + "mode": "merge", + "providers": { + "bankofai": { + "baseUrl": "https://api.bankofai.io/v1/", + "apiKey": "{BANKOFAI_API_KEY}", + "api": "openai-completions", + "models": [ + { "id": "gpt-5.2", "name": "gpt-5.2" }, + { "id": "gpt-5-mini", "name": "gpt-5-mini" }, + { "id": "gpt-5-nano", "name": "gpt-5-nano" }, + { "id": "claude-opus-4.6", "name": "claude-opus-4.6" }, + { "id": "claude-sonnet-4.6", "name": "claude-sonnet-4.6" }, + { "id": "claude-haiku-4.5", "name": "claude-haiku-4.5" } + ] + } + } + } +} +``` + +--- + +## 5.2 设置默认模型 + +```json +{ + "agents": { + "default": { + "model": "bankofai/gpt-5-nano" + } + } +} +``` + +--- + +## 5.3 重启服务 + +```bash +openclaw gateway restart +``` + +--- + +## 5.4 测试 + +```bash +openclaw agent --agent main --message "Hello" +``` + +--- + +# Step 6 Gateway 与诊断 + +常用命令: + +| 操作 | 命令 | +|---|---| +| 启动 | openclaw gateway start | +| 停止 | openclaw gateway stop | +| 重启 | openclaw gateway restart | +| 状态 | openclaw gateway status | + +诊断: + +```bash +openclaw doctor +``` + +--- + +# Step 7 启动 + +## Web UI + +```bash +openclaw dashboard +``` + +打开: + +``` +http://127.0.0.1:18789 +``` + +--- + +## 终端 UI + +```bash +openclaw tui +``` + +--- + +# Step 8 常用命令 + +```bash +openclaw models status +openclaw channels list +openclaw memory search "keyword" +openclaw docs +``` + +--- + +# 完成 + +你已经成功部署: + +**OpenClaw + Bank of AI AI Agent** + +你可以: + +- 自动化流程 +- 接入 Telegram +- 构建 AI 产品 + +🚀 diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/openclaw /one-click-script-tutorial.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/openclaw /one-click-script-tutorial.md new file mode 100644 index 00000000..ee4e9327 --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/openclaw /one-click-script-tutorial.md @@ -0,0 +1,135 @@ +## 快速开始 + +**在运行脚本前,请确保:** + +1. 已安装 Node.js 22 或以上版本 +2. 已安装并初始化 OpenClaw(已执行 `openclaw onboard`) +3. 网络正常,可访问 Bank of AI API + +--- + +## 脚本命令 + +### Linux & macOS + +Mac 用户:打开「终端」,输入: + +```bash +curl https://chat.bankofai.io/scripts/openclaw-install-bankofai-provider.sh | bash +``` + +--- + +### Windows PowerShell + +Windows 用户:打开「PowerShell」(不支持 CMD),输入: + +```powershell +iwr https://chat.bankofai.io/scripts/openclaw-install-bankofai-provider.ps1 | iex +``` + +--- + +## 详细步骤 + +### 1 获取 API Key + +1. 登录 Bank of AI 平台 + https://chat.bankofai.io/ + +2. 进入 API Key 页面 + https://chat.bankofai.io/key + +3. 创建新的 API Key + +![](https://files.readme.io/354a3d414f37e7df28f2cbf92dd055db9b67a20cab8738f6d5ac007226b6931b-image.png) + +--- + +### 2 运行安装脚本 + +根据系统执行对应命令,脚本会自动: + +- 检查环境(Node.js、OpenClaw 等) +- 提示输入 API Key + +![](https://files.readme.io/ef091efb8911db673af8a5eade7b281a52f641c93d41fbd57cb898ee91893e76-Image_16-3-2026_at_7.00PM.png) + +--- + +### 3 选择默认模型 + +验证 API Key 后,脚本会获取模型列表并提示选择默认模型。 + +![](https://files.readme.io/be3b9162405e38988898261db3effa8adcf92b6d9e37181d36b32c1cf8bcd1e3-Image_16-3-2026_at_7.01PM.png) + +> 注意:Gemini 系列模型在 OpenClaw 中存在兼容性问题(函数调用限制),建议谨慎选择。 + +--- + +### 4 完成配置 + +脚本会自动执行: + +- 备份原配置 +- 更新 OpenClaw 配置文件 +- 重启 Gateway + +![](https://files.readme.io/a08c7fcee0cbe906042ef52fefa15348750ad5cd2db8c4f555f92ecabba72761-Image_16-3-2026_at_7.03PM.png) + +--- + +### 5 切换模型 + +#### 方法一:命令行 + +```bash +openclaw models set bankofai/ +``` + +--- + +#### 方法二:Web 界面 + +打开: + +``` +http://127.0.0.1:18789/ +``` + +操作: + +- 左侧点击 Agent +- 在 Primary model 中选择模型 + +![](https://files.readme.io/3668289b53d185d158dd8393f46c0e171c0d301b5bdc624792c8cf8e6f9c4936-16-3-26_6.56.png) + +> 注意:如果通过 Dashboard 修改模型,配置文件会新增 `list` 字段,此时命令行切换将失效。 + +--- + +## 兼容性测试 + +| 系统 | 状态 | +|---|---| +| Ubuntu 24.04 | ✅ 通过 | +| Windows 11 | ✅ 通过 | +| macOS | ✅ 通过 | + +--- + +## 常见问题 + +### Q:脚本执行失败怎么办? + +请确认: + +1. Node.js ≥ 22 +2. 已执行 `openclaw onboard` +3. 网络正常,可访问 Bank of AI + +--- + +### Q:如何切换模型? + +参考上方「步骤 5 切换模型」。 diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/pricing-and-usage.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/pricing-and-usage.md new file mode 100644 index 00000000..ee8d59e7 --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/pricing-and-usage.md @@ -0,0 +1,56 @@ +# 计费与使用 + +## 积分与定价 + +Bank of AI 平台采用统一的积分(Credits)体系,对所有 AI 服务进行计量与结算。 + +- **计费规则:** 每次与 AI 交互所消耗的 Token,会根据不同模型的定价标准换算为对应积分,并从账户余额中扣除 +- **Token 消耗明细:** 在 AI 回复详情中可查看 Token 使用拆分,帮助理解消耗来源并优化使用策略 +- **模型定价差异:** 不同模型因能力与计算成本不同,价格有所差异。一般来说,能力越强的模型消耗越高 +- **联网搜索费用:** Web 搜索为额外收费功能,按次计费。部分模型不支持该功能(标记为 "-") + +### 模型定价 + +| 模型 | 输入(积分/Token) | 输出(积分/Token) | 联网搜索(积分/次) | +| :--- | :--- | :--- | :--- | +| ChatGPT-5.2 | 1.75 | 14.00 | 10,000 | +| ChatGPT-5-mini | 0.25 | 2.00 | 10,000 | +| ChatGPT-5-nano | 0.05 | 0.40 | - | +| Claude Opus 4.6 | 5.00 | 25.00 | 10,000 | +| Claude Opus 4.5 | 5.00 | 25.00 | 10,000 | +| Claude Sonnet 4.6 | 3.00 | 15.00 | 10,000 | +| Claude Sonnet 4.5 | 3.00 | 15.00 | 10,000 | +| Claude Haiku 4.5 | 1.00 | 5.00 | 10,000 | +| Gemini 3.1 Pro | 2.00 | 12.00 | 14,000 | +| Gemini 3 Flash | 0.50 | 3.00 | 14,000 | + +> **计算示例:** 若使用输入价格 1.25、输出价格 10.00 的模型,提问消耗 10 个输入 Token,回复消耗 50 个输出 Token,则总消耗为 **512.5 积分**(10 × 1.25 + 50 × 10)。可在聊天界面右下角悬停模型名称查看具体消耗。 + +--- + +## 使用数据 + +可在左侧导航栏进入 **使用情况** 页面查看详细数据: + +- **使用概览:** 当前积分余额与本月总消耗 +- **月度使用图:** 展示过去一年的使用趋势 +- **使用明细:** 每条记录对应一次 AI 交互,包括时间、模型、Token 用量、积分消耗和响应时间 + +--- + +## 充值 + +Bank of AI 采用预付费模式,基于区块链实现安全便捷的充值体验。 + +- **充值流程:** 在 **充值** 页面,通过已连接的钱包完成支付确认 +- **支持代币:** 支持多链主流代币(包括 TRON、BNB Chain 等) +- **到账时间:** 区块链确认后自动到账,通常为几分钟内 + +--- + +## 账单与记录 + +可在 **充值** 页面中的 **历史记录** 查看完整充值信息: + +- **充值记录:** 包括时间、类型、交易哈希、代币信息 +- **透明可查:** 可点击交易哈希,在区块链浏览器(如 TRONSCAN、BscScan)中查看详情 diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/quick-start.md b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/quick-start.md new file mode 100644 index 00000000..72a3d567 --- /dev/null +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/llmservice/quick-start.md @@ -0,0 +1,40 @@ +# 快速开始 + +本章节将引导你在 Bank of AI 平台上完成 LLM Service 的初始配置,快速开启 AI 使用体验。 + +## 1. 连接钱包 + +LLM Service 采用去中心化登录方式,支持通过 Web3 钱包进行授权登录,无需用户名和密码,安全便捷。 + +### 登录步骤: + +1. 访问 [Bank of AI 对话平台](https://chat.bankofai.io/chat) +2. 点击右上角 **登录** +3. 选择你的钱包并授权连接 +4. 在钱包中确认签名完成登录 + +登录成功后,右上角会显示你的钱包地址。 + +> 注意:请确保浏览器或移动设备中已安装兼容的 Web3 钱包,并妥善保管助记词和私钥。 + +--- + +## 2. 开始首次对话 + +进入平台后即可直接使用 LLM Service 进行 AI 交互。 + +- **选择模型:** 点击当前模型名称,从列表中选择需要使用的模型(如 GPT、Claude、Gemini 等) +- **发送消息:** 在输入框输入内容,点击发送按钮或按 Enter +- **多轮对话:** 支持连续上下文对话,AI 会根据历史内容进行响应 + +--- + +## 3. 余额与使用 + +LLM Service 采用积分(Credits)计费,通过链上支付获取。 + +- **充值入口:** 在控制台页面进入 **充值** +- **购买积分:** 使用支持的链上代币支付,在钱包中确认交易 +- **自动到账:** 区块链交易确认后,系统会自动为账户增加对应积分 + +完成以上步骤后,你即可使用多种主流 AI 模型。 diff --git a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/sidebars.js b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/sidebars.js index 4df37998..c154bf8f 100644 --- a/i18n/zh-Hans/docusaurus-plugin-content-docs/current/sidebars.js +++ b/i18n/zh-Hans/docusaurus-plugin-content-docs/current/sidebars.js @@ -154,6 +154,49 @@ const sidebars = { collapsed: false, items: ['Openclaw-extension/Intro', 'Openclaw-extension/QuickStart', 'Openclaw-extension/FAQ'], }, + + { + type: 'category', + label: 'LLM Service', + collapsed: false, + items: [ + { type: 'doc', id: 'llmservice/introduction', label: '简介' }, + { type: 'doc', id: 'llmservice/quick-start', label: '快速开始' }, + { type: 'doc', id: 'llmservice/pricing-and-usage', label: '定价与使用' }, + { + type: 'category', + label: '模型', + collapsed: true, + items: [ + 'llmservice/models/chatgpt-5-2', + 'llmservice/models/chatgpt-5-mini', + 'llmservice/models/chatgpt-5-nano', + 'llmservice/models/claude-haiku-4-5', + 'llmservice/models/claude-opus-4-5', + 'llmservice/models/claude-opus-4-6', + 'llmservice/models/claude-sonnet-4-5', + 'llmservice/models/claude-sonnet-4-6', + 'llmservice/models/gemini-3-1-pro', + 'llmservice/models/gemini-3-flash', + 'llmservice/models/glm-5', + 'llmservice/models/kimi-k2.5', + 'llmservice/models/minimax-m2.5', + ], + }, + { + type: 'category', + label: 'OpenClaw', + collapsed: true, + items: ['llmservice/openclaw/integration-guide', 'llmservice/openclaw/one-click-script-tutorial'], + }, + { + type: 'category', + label: 'API', + collapsed: true, + items: ['llmservice/api/API'], + }, + ], + }, ], } diff --git a/sidebars.js b/sidebars.js index 79785b69..8c6b3e56 100644 --- a/sidebars.js +++ b/sidebars.js @@ -70,7 +70,6 @@ const sidebars = { type: 'doc', id: '8004/SupportedNetworks', }, - { type: 'category', label: 'Usage', @@ -151,6 +150,49 @@ const sidebars = { collapsed: false, items: ['Openclaw-extension/Intro', 'Openclaw-extension/QuickStart', 'Openclaw-extension/FAQ'], }, + + { + type: 'category', + label: 'LLM Service', + collapsed: false, + items: [ + { type: 'doc', id: 'llmservice/introduction', label: 'Introduction' }, + { type: 'doc', id: 'llmservice/quick-start', label: 'Quick Start' }, + { type: 'doc', id: 'llmservice/pricing-and-usage', label: 'Pricing and Usage' }, + { + type: 'category', + label: 'Models', + collapsed: true, + items: [ + 'llmservice/models/chatgpt-5-2', + 'llmservice/models/chatgpt-5-mini', + 'llmservice/models/chatgpt-5-nano', + 'llmservice/models/claude-haiku-4-5', + 'llmservice/models/claude-opus-4-5', + 'llmservice/models/claude-opus-4-6', + 'llmservice/models/claude-sonnet-4-5', + 'llmservice/models/claude-sonnet-4-6', + 'llmservice/models/gemini-3-1-pro', + 'llmservice/models/gemini-3-flash', + 'llmservice/models/glm-5', + 'llmservice/models/kimi-k2.5', + 'llmservice/models/minimax-m2.5', + ], + }, + { + type: 'category', + label: 'OpenClaw', + collapsed: true, + items: ['llmservice/openclaw/integration-guide', 'llmservice/openclaw/one-click-script-tutorial'], + }, + { + type: 'category', + label: 'API', + collapsed: true, + items: ['llmservice/api/API'], + }, + ], + }, ], } diff --git a/src/pages/llmservice.js b/src/pages/llmservice.js new file mode 100644 index 00000000..a524ade6 --- /dev/null +++ b/src/pages/llmservice.js @@ -0,0 +1,14 @@ +import { useEffect } from 'react' +import { useHistory, useLocation } from '@docusaurus/router' + +export default function LlmServiceRedirect() { + const history = useHistory() + const { pathname } = useLocation() + + useEffect(() => { + const isZh = pathname.startsWith('/zh-Hans') + history.replace(isZh ? '/zh-Hans/llmservice/introduction' : '/llmservice/introduction') + }, []) + + return null +}