From def87547772efd23628e2ae97c5406369c903fdf Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 17 Mar 2025 08:49:45 +0000 Subject: [PATCH] Bump litellm from 1.63.0 to 1.63.11 Bumps [litellm](https://github.com/BerriAI/litellm) from 1.63.0 to 1.63.11. - [Release notes](https://github.com/BerriAI/litellm/releases) - [Commits](https://github.com/BerriAI/litellm/commits) --- updated-dependencies: - dependency-name: litellm dependency-type: direct:production update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] --- poetry.lock | 16 ++++++++-------- pyproject.toml | 4 ++-- 2 files changed, 10 insertions(+), 10 deletions(-) diff --git a/poetry.lock b/poetry.lock index a4a6f862..066a16e1 100644 --- a/poetry.lock +++ b/poetry.lock @@ -1546,14 +1546,14 @@ files = [ [[package]] name = "litellm" -version = "1.63.0" +version = "1.63.11" description = "Library to easily interface with LLM API providers" optional = false python-versions = "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8" groups = ["main", "dev"] files = [ - {file = "litellm-1.63.0-py3-none-any.whl", hash = "sha256:38961eaeb81fa2500c2725e01be898fb5d6347e73286b6d13d2f4d2f006d99e9"}, - {file = "litellm-1.63.0.tar.gz", hash = "sha256:872fb3fa4c8875d82fe998a5e4249c21a15bb08800286f03f90ed1700203f62e"}, + {file = "litellm-1.63.11-py3-none-any.whl", hash = "sha256:f3915dc35309b164ef2419ad05e5241ddd97f3f47aa036df28365bf889d8ea23"}, + {file = "litellm-1.63.11.tar.gz", hash = "sha256:89930895121d0cbf5553e560ed886c45be480ceec0eca3c53ae441473d5d46a4"}, ] [package.dependencies] @@ -1563,7 +1563,7 @@ httpx = ">=0.23.0" importlib-metadata = ">=6.8.0" jinja2 = ">=3.1.2,<4.0.0" jsonschema = ">=4.22.0,<5.0.0" -openai = ">=1.61.0" +openai = ">=1.66.1" pydantic = ">=2.0.0,<3.0.0" python-dotenv = ">=0.2.0" tiktoken = ">=0.7.0" @@ -2137,14 +2137,14 @@ sympy = "*" [[package]] name = "openai" -version = "1.61.1" +version = "1.66.3" description = "The official Python library for the openai API" optional = false python-versions = ">=3.8" groups = ["main", "dev"] files = [ - {file = "openai-1.61.1-py3-none-any.whl", hash = "sha256:72b0826240ce26026ac2cd17951691f046e5be82ad122d20a8e1b30ca18bd11e"}, - {file = "openai-1.61.1.tar.gz", hash = "sha256:ce1851507218209961f89f3520e06726c0aa7d0512386f0f977e3ac3e4f2472e"}, + {file = "openai-1.66.3-py3-none-any.whl", hash = "sha256:a427c920f727711877ab17c11b95f1230b27767ba7a01e5b66102945141ceca9"}, + {file = "openai-1.66.3.tar.gz", hash = "sha256:8dde3aebe2d081258d4159c4cb27bdc13b5bb3f7ea2201d9bd940b9a89faf0c9"}, ] [package.dependencies] @@ -4276,4 +4276,4 @@ type = ["pytest-mypy"] [metadata] lock-version = "2.1" python-versions = ">=3.12,<3.13" -content-hash = "11d90797bbc8dee54226f9c44d922333558c96820801cbe2cf677e313ff58fd0" +content-hash = "a9b5c2cadc3b38398935336a4cd08078d9bf2a4084dfd8f693a254dc193c01dd" diff --git a/pyproject.toml b/pyproject.toml index fdfd9b05..30762f68 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -14,7 +14,7 @@ PyYAML = "==6.0.2" fastapi = "==0.115.11" uvicorn = "==0.34.0" structlog = "==25.2.0" -litellm = "==1.63.0" +litellm = "==1.63.11" llama_cpp_python = "==0.3.5" cryptography = "==44.0.2" sqlalchemy = "==2.0.39" @@ -50,7 +50,7 @@ ruff = "==0.11.0" bandit = "==1.8.3" build = "==1.2.2.post1" wheel = "==0.45.1" -litellm = "==1.63.0" +litellm = "==1.63.11" pytest-asyncio = "==0.25.3" llama_cpp_python = "==0.3.5" scikit-learn = "==1.6.1"