diff --git a/poetry.lock b/poetry.lock index 90b4fdd754..67fc4b7708 100644 --- a/poetry.lock +++ b/poetry.lock @@ -3865,19 +3865,19 @@ tenacity = ">=8.5.0,<10.0" [[package]] name = "llama-index" -version = "0.13.2" +version = "0.14.3" description = "Interface between LLMs and your data" optional = false python-versions = "<4.0,>=3.9" groups = ["main"] files = [ - {file = "llama_index-0.13.2-py3-none-any.whl", hash = "sha256:8de8eefffcfa64a9225267d7813fcb55b8ea12181d4044efe5b22642d91d2294"}, - {file = "llama_index-0.13.2.tar.gz", hash = "sha256:110e5e8e077aab7643eecb0962bcdb927bdea6a2c9897606b4b26e498d93dd5b"}, + {file = "llama_index-0.14.3-py3-none-any.whl", hash = "sha256:bdfb42068712bd4c80015c4af41f3208b8b5ebb3115e438d14f56b476ac35e4c"}, + {file = "llama_index-0.14.3.tar.gz", hash = "sha256:316b3d6d80a836f8bb59bac17af63289c022f56f3ed3ff82fb6f87bba6cf4091"}, ] [package.dependencies] llama-index-cli = ">=0.5.0,<0.6" -llama-index-core = ">=0.13.2,<0.14" +llama-index-core = ">=0.14.3,<0.15" llama-index-embeddings-openai = ">=0.5.0,<0.6" llama-index-indices-managed-llama-cloud = ">=0.4.0" llama-index-llms-openai = ">=0.5.0,<0.6" @@ -3887,31 +3887,31 @@ nltk = ">3.8.1" [[package]] name = "llama-index-cli" -version = "0.5.0" +version = "0.5.1" description = "llama-index cli" optional = false python-versions = "<4.0,>=3.9" groups = ["main"] files = [ - {file = "llama_index_cli-0.5.0-py3-none-any.whl", hash = "sha256:e331ca98005c370bfe58800fa5eed8b10061d0b9c656b84a1f5f6168733a2a7b"}, - {file = "llama_index_cli-0.5.0.tar.gz", hash = "sha256:2eb9426232e8d89ffdf0fa6784ff8da09449d920d71d0fcc81d07be93cf9369f"}, + {file = "llama_index_cli-0.5.1-py3-none-any.whl", hash = "sha256:5429b2fd7960df7724c2955b6e6901f6fa910b7b5ecef411c979a8b545a6b7e2"}, + {file = "llama_index_cli-0.5.1.tar.gz", hash = "sha256:0446159d85c56c29022c1c830c9886f670d5f59d69343c3c029a3b20eda1a9d8"}, ] [package.dependencies] -llama-index-core = ">=0.13.0,<0.14" +llama-index-core = ">=0.13.0,<0.15" llama-index-embeddings-openai = ">=0.5.0,<0.6" llama-index-llms-openai = ">=0.5.0,<0.6" [[package]] name = "llama-index-core" -version = "0.13.2" +version = "0.14.14" description = "Interface between LLMs and your data" optional = false python-versions = "<4.0,>=3.9" groups = ["main"] files = [ - {file = "llama_index_core-0.13.2-py3-none-any.whl", hash = "sha256:335481fdd1bd0c230a8c6136671fe9a2dcabdc760166f1df91b92e7907784396"}, - {file = "llama_index_core-0.13.2.tar.gz", hash = "sha256:3a9496d5ec8c476f1e8da42cde806cc49262fa21c76bdc936e0a02513e843f50"}, + {file = "llama_index_core-0.14.14-py3-none-any.whl", hash = "sha256:b62512a4ba1911b50b30d25d721535b30c9242f5f2dd6a4d5a4a712cca7b570e"}, + {file = "llama_index_core-0.14.14.tar.gz", hash = "sha256:505a252c7e357cb12fb9bef41498ae2be3f2baa07666a939f8107f3faef2823b"}, ] [package.dependencies] @@ -3924,7 +3924,7 @@ dirtyjson = ">=1.0.8,<2" filetype = ">=1.2.0,<2" fsspec = ">=2023.5.0" httpx = "*" -llama-index-workflows = ">=1.0.1,<2" +llama-index-workflows = ">=2,<2.9.0 || >2.9.0,<3" nest-asyncio = ">=1.5.8,<2" networkx = ">=3.0" nltk = ">3.8.1" @@ -3945,36 +3945,36 @@ wrapt = "*" [[package]] name = "llama-index-embeddings-openai" -version = "0.5.0" +version = "0.5.1" description = "llama-index embeddings openai integration" optional = false python-versions = "<4.0,>=3.9" groups = ["main"] files = [ - {file = "llama_index_embeddings_openai-0.5.0-py3-none-any.whl", hash = "sha256:d817edb22e3ff475e8cd1833faf1147028986bc1d688f7894ef947558864b728"}, - {file = "llama_index_embeddings_openai-0.5.0.tar.gz", hash = "sha256:ac587839a111089ea8a6255f9214016d7a813b383bbbbf9207799be1100758eb"}, + {file = "llama_index_embeddings_openai-0.5.1-py3-none-any.whl", hash = "sha256:a2fcda3398bbd987b5ce3f02367caee8e84a56b930fdf43cc1d059aa9fd20ca5"}, + {file = "llama_index_embeddings_openai-0.5.1.tar.gz", hash = "sha256:1c89867a48b0d0daa3d2d44f5e76b394b2b2ef9935932daf921b9e77939ccda8"}, ] [package.dependencies] -llama-index-core = ">=0.13.0,<0.14" +llama-index-core = ">=0.13.0,<0.15" openai = ">=1.1.0" [[package]] name = "llama-index-indices-managed-llama-cloud" -version = "0.9.2" +version = "0.9.4" description = "llama-index indices llama-cloud integration" optional = false python-versions = "<4.0,>=3.9" groups = ["main"] files = [ - {file = "llama_index_indices_managed_llama_cloud-0.9.2-py3-none-any.whl", hash = "sha256:d139602d1c268ae4ce2003838e857c5ef5527fb1adc78add2d9ce18fef4b4bfa"}, - {file = "llama_index_indices_managed_llama_cloud-0.9.2.tar.gz", hash = "sha256:19af55da8f1218d80390fcbe5cdfef6100acc755b0177a4077a924c31a2fb345"}, + {file = "llama_index_indices_managed_llama_cloud-0.9.4-py3-none-any.whl", hash = "sha256:535a08811046803ca6ab7f8e9d510e926aa5306608b02201ad3d9d21701383bc"}, + {file = "llama_index_indices_managed_llama_cloud-0.9.4.tar.gz", hash = "sha256:b5e00752ab30564abf19c57595a2107f5697c3b03b085817b4fca84a38ebbd59"}, ] [package.dependencies] deprecated = "1.2.18" llama-cloud = "0.1.35" -llama-index-core = ">=0.13.0,<0.14" +llama-index-core = ">=0.13.0,<0.15" [[package]] name = "llama-index-instrumentation" @@ -3994,38 +3994,38 @@ pydantic = ">=2.11.5" [[package]] name = "llama-index-llms-openai" -version = "0.5.4" +version = "0.5.6" description = "llama-index llms openai integration" optional = false python-versions = "<4.0,>=3.9" groups = ["main"] files = [ - {file = "llama_index_llms_openai-0.5.4-py3-none-any.whl", hash = "sha256:8d42fbfa56b5f281ad0dfcb2915916c188b5876625f9f8d27016b7dc4366cc24"}, - {file = "llama_index_llms_openai-0.5.4.tar.gz", hash = "sha256:9e36b6d2fc5f056b00ee655901b3bb7e7060b23f7b19439889fb78d696340f54"}, + {file = "llama_index_llms_openai-0.5.6-py3-none-any.whl", hash = "sha256:a93a897fe733a6d7b668cbc6cca546e644054ddf5497821141b2d4b5ffb6ea80"}, + {file = "llama_index_llms_openai-0.5.6.tar.gz", hash = "sha256:92533e83be2eb321d84a01a84fb2bf4506bf684c410cd94ccb29ae6c949a27d4"}, ] [package.dependencies] -llama-index-core = ">=0.13.0,<0.14" +llama-index-core = ">=0.13.0,<0.15" openai = ">=1.81.0,<2" [[package]] name = "llama-index-readers-file" -version = "0.5.2" +version = "0.5.6" description = "llama-index readers file integration" optional = false python-versions = "<4.0,>=3.9" groups = ["main"] files = [ - {file = "llama_index_readers_file-0.5.2-py3-none-any.whl", hash = "sha256:c85e20022f3da3f635d38401863243c557d4e9ad6872acaafc9281e41a0c9460"}, - {file = "llama_index_readers_file-0.5.2.tar.gz", hash = "sha256:049d971ac4c936edbf4832915ba7128cfee8f5ead435266792b71edd87f5305c"}, + {file = "llama_index_readers_file-0.5.6-py3-none-any.whl", hash = "sha256:32e83f9adb4e4803e6c7cef746c44fa0949013b1cb76f06f422e9491d198dbda"}, + {file = "llama_index_readers_file-0.5.6.tar.gz", hash = "sha256:1c08b14facc2dfe933622aaa26dc7d2a7a6023c42d3db896a2c948789edaf1ea"}, ] [package.dependencies] beautifulsoup4 = ">=4.12.3,<5" defusedxml = ">=0.7.1" -llama-index-core = ">=0.13.0,<0.14" -pandas = "<2.3.0" -pypdf = ">=5.1.0,<7" +llama-index-core = ">=0.13.0,<0.15" +pandas = ">=2.0.0,<3" +pypdf = ">=6.1.3,<7" striprtf = ">=0.0.26,<0.0.27" [package.extras] @@ -4033,30 +4033,30 @@ pymupdf = ["pymupdf (>=1.23.21,<2)"] [[package]] name = "llama-index-readers-llama-parse" -version = "0.5.0" +version = "0.5.1" description = "llama-index readers llama-parse integration" optional = false python-versions = "<4.0,>=3.9" groups = ["main"] files = [ - {file = "llama_index_readers_llama_parse-0.5.0-py3-none-any.whl", hash = "sha256:e63ebf2248c4a726b8a1f7b029c90383d82cdc142942b54dbf287d1f3aee6d75"}, - {file = "llama_index_readers_llama_parse-0.5.0.tar.gz", hash = "sha256:891b21fb63fe1fe722e23cfa263a74d9a7354e5d8d7a01f2d4040a52f8d8feef"}, + {file = "llama_index_readers_llama_parse-0.5.1-py3-none-any.whl", hash = "sha256:0d41450ed29b0c49c024e206ef6c8e662b1854e77a1c5faefed3b958be54f880"}, + {file = "llama_index_readers_llama_parse-0.5.1.tar.gz", hash = "sha256:2b78b73faa933e30e6c69df351e4e9f36dfe2ae142e2ab3969ddd2ac48930e37"}, ] [package.dependencies] -llama-index-core = ">=0.13.0,<0.14" +llama-index-core = ">=0.13.0,<0.15" llama-parse = ">=0.5.0" [[package]] name = "llama-index-workflows" -version = "1.3.0" +version = "2.14.2" description = "An event-driven, async-first, step-based way to control the execution flow of AI applications like Agents." optional = false python-versions = ">=3.9" groups = ["main"] files = [ - {file = "llama_index_workflows-1.3.0-py3-none-any.whl", hash = "sha256:328cc25d92b014ef527f105a2f2088c0924fff0494e53d93decb951f14fbfe47"}, - {file = "llama_index_workflows-1.3.0.tar.gz", hash = "sha256:9c1688e237efad384f16485af71c6f9456a2eb6d85bf61ff49e5717f10ff286d"}, + {file = "llama_index_workflows-2.14.2-py3-none-any.whl", hash = "sha256:1360b5ae97459aa990048be0854788e33fe8a07d58e3519617ffc98c7b7326b3"}, + {file = "llama_index_workflows-2.14.2.tar.gz", hash = "sha256:b3f17a804c277901d30bc6e1b56398163db182ec48384f79ab16fdc9573b8877"}, ] [package.dependencies] @@ -4065,7 +4065,8 @@ pydantic = ">=2.11.5" typing-extensions = ">=4.6.0" [package.extras] -server = ["starlette (>=0.39.0)", "uvicorn (>=0.32.0)"] +client = ["llama-agents-client (>=0.1.0,<1.0.0)"] +server = ["llama-agents-server (>=0.1.0,<1.0.0)"] [[package]] name = "llama-parse" @@ -6711,20 +6712,20 @@ diagrams = ["jinja2", "railroad-diagrams"] [[package]] name = "pypdf" -version = "5.5.0" +version = "6.7.0" description = "A pure-python PDF library capable of splitting, merging, cropping, and transforming PDF files" optional = false -python-versions = ">=3.8" +python-versions = ">=3.9" groups = ["main"] files = [ - {file = "pypdf-5.5.0-py3-none-any.whl", hash = "sha256:2f61f2d32dde00471cd70b8977f98960c64e84dd5ba0d070e953fcb4da0b2a73"}, - {file = "pypdf-5.5.0.tar.gz", hash = "sha256:8ce6a18389f7394fd09a1d4b7a34b097b11c19088a23cfd09e5008f85893e254"}, + {file = "pypdf-6.7.0-py3-none-any.whl", hash = "sha256:62e85036d50839cbdf45b8067c2c1a1b925517514d7cba4cbe8755a6c2829bc9"}, + {file = "pypdf-6.7.0.tar.gz", hash = "sha256:eb95e244d9f434e6cfd157272283339ef586e593be64ee699c620f756d5c3f7e"}, ] [package.extras] crypto = ["cryptography"] cryptodome = ["PyCryptodome"] -dev = ["black", "flit", "pip-tools", "pre-commit", "pytest-cov", "pytest-socket", "pytest-timeout", "pytest-xdist", "wheel"] +dev = ["flit", "pip-tools", "pre-commit", "pytest-cov", "pytest-socket", "pytest-timeout", "pytest-xdist", "wheel"] docs = ["myst_parser", "sphinx", "sphinx_rtd_theme"] full = ["Pillow (>=8.0.0)", "cryptography"] image = ["Pillow (>=8.0.0)"] @@ -9488,4 +9489,4 @@ cffi = ["cffi (>=1.11)"] [metadata] lock-version = "2.1" python-versions = "~3.12" -content-hash = "aa40d98372564b86b8f2a559a5244da8e2c4bc24108d755335d5297ea74045b9" +content-hash = "0ef20dec4659ea0da7a31889f4ab2b80de364442acc8ab38c18409653b0ea4fc" diff --git a/pyproject.toml b/pyproject.toml index bda0e7e187..6a9e14c39f 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -61,7 +61,7 @@ langchain = "^0.3.11" langchain-experimental = "^0.3.4" langchain-openai = "^0.3.2" litellm = "1.79.3" -llama-index = "^0.13.0" +llama-index = "^0.14.0" llama-index-llms-openai = "^0.5.0" lxml = "^6.0.0" markdown = "^3.7"