From 17d44353d19a204ff364a19bddad95514a16ed3f Mon Sep 17 00:00:00 2001 From: Yue Fei Date: Tue, 21 Jan 2025 10:15:26 +0800 Subject: [PATCH] Bump llama-index-core to 0.12.12 (#348) * update to llama-index 0.12.12 * update gpu deps --------- Co-authored-by: Yue Fei --- poetry.lock | 8 ++++---- pyproject.toml | 2 +- pyproject_gpu.toml | 2 +- 3 files changed, 6 insertions(+), 6 deletions(-) diff --git a/poetry.lock b/poetry.lock index 6df66557..8499a69c 100644 --- a/poetry.lock +++ b/poetry.lock @@ -3990,13 +3990,13 @@ openai = ">=1.14.0" [[package]] name = "llama-index-core" -version = "0.12.11" +version = "0.12.12" description = "Interface between LLMs and your data" optional = false python-versions = "<4.0,>=3.9" files = [ - {file = "llama_index_core-0.12.11-py3-none-any.whl", hash = "sha256:3b1e019c899e9e011dfa01c96b7e3f666e0c161035fbca6cb787b4c61e0c94db"}, - {file = "llama_index_core-0.12.11.tar.gz", hash = "sha256:9a41ca91167ea5eec9ebaac7f5e958b7feddbd8af3bfbf7c393a5edfb994d566"}, + {file = "llama_index_core-0.12.12-py3-none-any.whl", hash = "sha256:cea491e87f65e6b775b5aef95720de302b85af1bdc67d779c4b09170a30e5b98"}, + {file = "llama_index_core-0.12.12.tar.gz", hash = "sha256:068b755bbc681731336e822f5977d7608585e8f759c6293ebd812e2659316a37"}, ] [package.dependencies] @@ -10564,4 +10564,4 @@ testing = ["coverage[toml]", "zope.event", "zope.testing"] [metadata] lock-version = "2.0" python-versions = ">=3.11.0,<3.12" -content-hash = "67daf40cf054d8f4ed68673d1a9f790fbc86ae1264202abbb597490a8410dc1d" +content-hash = "aba41ab55b4b9bc3287ceba30c825282733b2b86e00c73baa7748a858d67cc76" diff --git a/pyproject.toml b/pyproject.toml index 7352f717..a065e01f 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -13,7 +13,7 @@ readme = "README.md" python = ">=3.11.0,<3.12" fastapi = "0.115.2" uvicorn = "^0.29.0" -llama-index-core = "0.12.11" +llama-index-core = "0.12.12" llama-index-embeddings-openai = "^0.3.0" llama-index-embeddings-azure-openai = "^0.3.0" llama-index-embeddings-dashscope = "^0.3.0" diff --git a/pyproject_gpu.toml b/pyproject_gpu.toml index 507a9a00..d43855d3 100644 --- a/pyproject_gpu.toml +++ b/pyproject_gpu.toml @@ -13,7 +13,7 @@ readme = "README.md" python = ">=3.11.0,<3.12" fastapi = "0.115.2" uvicorn = "^0.29.0" -llama-index-core = "^0.12.11" +llama-index-core = "0.12.12" llama-index-embeddings-openai = "^0.3.0" llama-index-embeddings-azure-openai = "^0.3.0" llama-index-embeddings-dashscope = "^0.3.0"