diff --git a/poetry.lock b/poetry.lock index 9093781..85aab65 100644 --- a/poetry.lock +++ b/poetry.lock @@ -151,29 +151,29 @@ python-dateutil = ">=2.7" [[package]] name = "h11" -version = "0.14.0" +version = "0.16.0" description = "A pure-Python, bring-your-own-I/O implementation of HTTP/1.1" optional = false -python-versions = ">=3.7" +python-versions = ">=3.8" files = [ - {file = "h11-0.14.0-py3-none-any.whl", hash = "sha256:e3fe4ac4b851c468cc8363d500db52c2ead036020723024a109d37346efaa761"}, - {file = "h11-0.14.0.tar.gz", hash = "sha256:8f19fbbe99e72420ff35c00b27a34cb9937e902a8b810e2c88300c6f0a3b699d"}, + {file = "h11-0.16.0-py3-none-any.whl", hash = "sha256:63cf8bbe7522de3bf65932fda1d9c2772064ffb3dae62d55932da54b31cb6c86"}, + {file = "h11-0.16.0.tar.gz", hash = "sha256:4e35b956cf45792e4caa5885e69fba00bdbc6ffafbfa020300e549b208ee5ff1"}, ] [[package]] name = "httpcore" -version = "1.0.7" +version = "1.0.9" description = "A minimal low-level HTTP client." optional = false python-versions = ">=3.8" files = [ - {file = "httpcore-1.0.7-py3-none-any.whl", hash = "sha256:a3fff8f43dc260d5bd363d9f9cf1830fa3a458b332856f34282de498ed420edd"}, - {file = "httpcore-1.0.7.tar.gz", hash = "sha256:8551cb62a169ec7162ac7be8d4817d561f60e08eaa485234898414bb5a8a0b4c"}, + {file = "httpcore-1.0.9-py3-none-any.whl", hash = "sha256:2d400746a40668fc9dec9810239072b40b4484b640a8c38fd654a024c7a1bf55"}, + {file = "httpcore-1.0.9.tar.gz", hash = "sha256:6e34463af53fd2ab5d807f399a9b45ea31c3dfa2276f15a2c3f00afff6e176e8"}, ] [package.dependencies] certifi = "*" -h11 = ">=0.13,<0.15" +h11 = ">=0.16" [package.extras] asyncio = ["anyio (>=4.0,<5.0)"] diff --git a/pyproject.toml b/pyproject.toml index 3614316..c984e8f 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [tool.poetry] name = "workflowai" -version = "0.6.9" +version = "0.6.10" description = "Python SDK for WorkflowAI" authors = ["Guillaume Aquilina "] readme = "README.md" diff --git a/workflowai/core/domain/model.py b/workflowai/core/domain/model.py index 60e5408..48cd18b 100644 --- a/workflowai/core/domain/model.py +++ b/workflowai/core/domain/model.py @@ -65,9 +65,11 @@ class Model(str, Enum): # -------------------------------------------------------------------------- # Gemini Models # -------------------------------------------------------------------------- + GEMINI_2_0_FLASH_LATEST = "gemini-2.0-flash-latest" + GEMINI_2_5_FLASH_PREVIEW_0417 = "gemini-2.5-flash-preview-04-17" + GEMINI_2_5_FLASH_THINKING_PREVIEW_0417 = "gemini-2.5-flash-thinking-preview-04-17" GEMINI_2_5_PRO_PREVIEW_0325 = "gemini-2.5-pro-preview-03-25" GEMINI_2_5_PRO_EXP_0325 = "gemini-2.5-pro-exp-03-25" - GEMINI_2_0_FLASH_LATEST = "gemini-2.0-flash-latest" GEMINI_2_0_FLASH_001 = "gemini-2.0-flash-001" GEMINI_2_0_FLASH_LITE_001 = "gemini-2.0-flash-lite-001" GEMINI_2_0_FLASH_LITE_PREVIEW_2502 = "gemini-2.0-flash-lite-preview-02-05" @@ -108,8 +110,10 @@ class Model(str, Enum): # -------------------------------------------------------------------------- # Llama Models # -------------------------------------------------------------------------- - LLAMA_4_SCOUT_BASIC = "llama4-scout-instruct-basic" + LLAMA_4_MAVERICK_FAST = "llama4-maverick-instruct-fast" + LLAMA_4_SCOUT_FAST = "llama4-scout-instruct-fast" LLAMA_4_MAVERICK_BASIC = "llama4-maverick-instruct-basic" + LLAMA_4_SCOUT_BASIC = "llama4-scout-instruct-basic" LLAMA_3_3_70B = "llama-3.3-70b" LLAMA_3_2_90B = "llama-3.2-90b" LLAMA_3_2_11B = "llama-3.2-11b" @@ -151,6 +155,7 @@ class Model(str, Enum): # -------------------------------------------------------------------------- # Qwen Models # -------------------------------------------------------------------------- + QWEN_QWQ_32B = "qwen-qwq-32b" QWEN_QWQ_32B_PREVIEW = "qwen-v3p2-32b-instruct" # --------------------------------------------------------------------------