diff --git a/pdm.lock b/pdm.lock index 508ea71a..f41de072 100644 --- a/pdm.lock +++ b/pdm.lock @@ -5,7 +5,7 @@ groups = ["default", "dev"] strategy = ["inherit_metadata"] lock_version = "4.4.2" -content_hash = "sha256:aa5a61326e032b4b9649f6e946a99930405bfc2bc4f995e8d7124f79fe2ee7ce" +content_hash = "sha256:b9fc0555a8757b71893d99349328c636f8be7276a7ec15c5b46bb804345ce1ea" [[package]] name = "absl-py" @@ -682,6 +682,16 @@ files = [ {file = "filelock-3.16.1.tar.gz", hash = "sha256:c249fbfcd5db47e5e2d6d62198e565475ee65e4831e2561c8e313fa7eb961435"}, ] +[[package]] +name = "filetype" +version = "1.2.0" +summary = "Infer file type and MIME type of any file/buffer. No external dependencies." +groups = ["default"] +files = [ + {file = "filetype-1.2.0-py2.py3-none-any.whl", hash = "sha256:7ce71b6880181241cf7ac8697a2f1eb6a8bd9b429f7ad6d27b8db9ba5f1c2d25"}, + {file = "filetype-1.2.0.tar.gz", hash = "sha256:66b56cd6474bf41d8c54660347d37afcc3f7d1970648de365c102ef77548aadb"}, +] + [[package]] name = "findpython" version = "0.6.2" @@ -1452,7 +1462,7 @@ files = [ [[package]] name = "llama-cloud" -version = "0.0.9" +version = "0.1.5" requires_python = "<4,>=3.8" summary = "" groups = ["default"] @@ -1461,71 +1471,72 @@ dependencies = [ "pydantic>=1.10", ] files = [ - {file = "llama_cloud-0.0.9-py3-none-any.whl", hash = "sha256:3d6e6b051d63f466bee49bc8bbd03fdec658902ff20a52342d38b7824fc90759"}, - {file = "llama_cloud-0.0.9.tar.gz", hash = "sha256:c876ac3f37b6ddf395d29c5fbc93ffc6e2d1165107cb236ec15b279e67937328"}, + {file = "llama_cloud-0.1.5-py3-none-any.whl", hash = "sha256:15605022520d04bd6ef6a46c0cbde833f301d652286d34fca02b4c44e2a7a2aa"}, + {file = "llama_cloud-0.1.5.tar.gz", hash = "sha256:8ce1db36754a6a46c8511561dbc040a2e89ba4ca1cf4edfb6ce382a5240f6cb6"}, ] [[package]] name = "llama-index" -version = "0.10.62" -requires_python = "<4.0,>=3.8.1" +version = "0.12.2" +requires_python = "<4.0,>=3.9" summary = "Interface between LLMs and your data" groups = ["default"] dependencies = [ - "llama-index-agent-openai<0.3.0,>=0.1.4", - "llama-index-cli<0.2.0,>=0.1.2", - "llama-index-core==0.10.62", - "llama-index-embeddings-openai<0.2.0,>=0.1.5", - "llama-index-indices-managed-llama-cloud>=0.2.0", + "llama-index-agent-openai<0.5.0,>=0.4.0", + "llama-index-cli<0.5.0,>=0.4.0", + "llama-index-core<0.13.0,>=0.12.2", + "llama-index-embeddings-openai<0.4.0,>=0.3.0", + "llama-index-indices-managed-llama-cloud>=0.4.0", "llama-index-legacy<0.10.0,>=0.9.48", - "llama-index-llms-openai<0.2.0,>=0.1.27", - "llama-index-multi-modal-llms-openai<0.2.0,>=0.1.3", - "llama-index-program-openai<0.2.0,>=0.1.3", - "llama-index-question-gen-openai<0.2.0,>=0.1.2", - "llama-index-readers-file<0.2.0,>=0.1.4", - "llama-index-readers-llama-parse>=0.1.2", + "llama-index-llms-openai<0.4.0,>=0.3.0", + "llama-index-multi-modal-llms-openai<0.4.0,>=0.3.0", + "llama-index-program-openai<0.4.0,>=0.3.0", + "llama-index-question-gen-openai<0.4.0,>=0.3.0", + "llama-index-readers-file<0.5.0,>=0.4.0", + "llama-index-readers-llama-parse>=0.4.0", + "nltk>3.8.1", ] files = [ - {file = "llama_index-0.10.62-py3-none-any.whl", hash = "sha256:13af83c70860ba570e4ff34e57b8b3e48cf4967c925456f5526c77c52004fb44"}, - {file = "llama_index-0.10.62.tar.gz", hash = "sha256:b649a645bb5281a30077b74671132734f360c77370b6ef453d91a065c0029867"}, + {file = "llama_index-0.12.2-py3-none-any.whl", hash = "sha256:971528db7889f5a0d15fd9039a403bc6f92bfafc2d4e1bab2d166657728ae94c"}, + {file = "llama_index-0.12.2.tar.gz", hash = "sha256:da9738dd666e219689839c7451c9df8bed72e6510a6f7d6f7d9907bfdd4588eb"}, ] [[package]] name = "llama-index-agent-openai" -version = "0.2.3" -requires_python = "<4.0,>=3.8.1" +version = "0.4.0" +requires_python = "<4.0,>=3.9" summary = "llama-index agent openai integration" groups = ["default"] dependencies = [ - "llama-index-core<0.11.0,>=0.10.30", - "llama-index-llms-openai<0.2.0,>=0.1.5", + "llama-index-core<0.13.0,>=0.12.0", + "llama-index-llms-openai<0.4.0,>=0.3.0", "openai>=1.14.0", ] files = [ - {file = "llama_index_agent_openai-0.2.3-py3-none-any.whl", hash = "sha256:3782b24dd611364e391672dadc8308efd58d731a097c34a40e29f28c3abc5034"}, - {file = "llama_index_agent_openai-0.2.3.tar.gz", hash = "sha256:c899d90b32036656a8ef86d0f0378d4168e00eb2d75a10901eab58ba5b2656a4"}, + {file = "llama_index_agent_openai-0.4.0-py3-none-any.whl", hash = "sha256:71b2f46bb24813129ab6bc2d5bcebb9aebf323403ebf1e6cc9840687a34a6169"}, + {file = "llama_index_agent_openai-0.4.0.tar.gz", hash = "sha256:31d2675dbd84489756dd062a7ffed330b2abdca3b7715d511674f5b5075e4dd6"}, ] [[package]] name = "llama-index-cli" -version = "0.1.12" -requires_python = "<4.0,>=3.8.1" +version = "0.4.0" +requires_python = "<4.0,>=3.9" summary = "llama-index cli" groups = ["default"] dependencies = [ - "llama-index-core<0.11.0,>=0.10.11.post1", - "llama-index-embeddings-openai<0.2.0,>=0.1.1", - "llama-index-llms-openai<0.2.0,>=0.1.1", + "llama-index-core<0.13.0,>=0.12.0", + "llama-index-embeddings-openai<0.4.0,>=0.3.0", + "llama-index-llms-openai<0.4.0,>=0.3.0", ] files = [ - {file = "llama_index_cli-0.1.12-py3-none-any.whl", hash = "sha256:d80d546786f02d3f16f6183b8e86b22b8b5c33a1500923659f2ccbff8d5df634"}, - {file = "llama_index_cli-0.1.12.tar.gz", hash = "sha256:3cf1f706c3c69c6b1aab07fca7faad3959db1709808efd50491b669d38b0b580"}, + {file = "llama_index_cli-0.4.0-py3-none-any.whl", hash = "sha256:60d12f89e6b85e80a0cc3a8b531f05a911b5eebaebc37314411476d1ba685904"}, + {file = "llama_index_cli-0.4.0.tar.gz", hash = "sha256:d6ab201359962a8a34368aeda3a49bbbe67e9e009c59bd925c4fb2be4ace3906"}, ] [[package]] name = "llama-index-core" -version = "0.10.62" -requires_python = "<4.0,>=3.8.1" +version = "0.12.2" +requires_python = "<4.0,>=3.9" summary = "Interface between LLMs and your data" groups = ["default"] dependencies = [ @@ -1535,15 +1546,15 @@ dependencies = [ "dataclasses-json", "deprecated>=1.2.9.3", "dirtyjson<2.0.0,>=1.0.8", + "filetype<2.0.0,>=1.2.0", "fsspec>=2023.5.0", "httpx", "nest-asyncio<2.0.0,>=1.5.8", "networkx>=3.0", - "nltk<4.0.0,>=3.8.1", - "numpy<2.0.0", - "openai>=1.1.0", - "pandas", + "nltk>3.8.1", + "numpy", "pillow>=9.0.0", + "pydantic<2.10.0,>=2.7.0", "requests>=2.31.0", "tenacity!=8.4.0,<9.0.0,>=8.2.0", "tiktoken>=0.3.3", @@ -1553,53 +1564,54 @@ dependencies = [ "wrapt", ] files = [ - {file = "llama_index_core-0.10.62-py3-none-any.whl", hash = "sha256:c48c4b8bdd0ad6eec3f7c4ca129509cdbe5614f3d2ed76bec30999899a38b962"}, - {file = "llama_index_core-0.10.62.tar.gz", hash = "sha256:227f011829497e654bb32ab6907318f613c3a9a6809e08c20163395c26838606"}, + {file = "llama_index_core-0.12.2-py3-none-any.whl", hash = "sha256:27a5548523435a5c2b84f75c15894a44522b7f968e9f29a03f9a301ca09fb7fa"}, + {file = "llama_index_core-0.12.2.tar.gz", hash = "sha256:a48b2de9c3a09608ab5c03c5a313428f119c86946acdefde555992b7c0b8a38e"}, ] [[package]] name = "llama-index-embeddings-huggingface" -version = "0.2.2" -requires_python = "<4.0,>=3.8.1" +version = "0.4.0" +requires_python = "<4.0,>=3.9" summary = "llama-index embeddings huggingface integration" groups = ["default"] dependencies = [ "huggingface-hub[inference]>=0.19.0", - "llama-index-core<0.11.0,>=0.10.1", + "llama-index-core<0.13.0,>=0.12.0", "sentence-transformers>=2.6.1", ] files = [ - {file = "llama_index_embeddings_huggingface-0.2.2-py3-none-any.whl", hash = "sha256:3445b1c7823cdb45622f90e79f2540db870ea55b226ec7538be963d340f43240"}, - {file = "llama_index_embeddings_huggingface-0.2.2.tar.gz", hash = "sha256:43b2978740d29291ae4c7566922d2b1c7543dc979e268794b578e1a2adfb4319"}, + {file = "llama_index_embeddings_huggingface-0.4.0-py3-none-any.whl", hash = "sha256:a5890bab349b118398054138b298a9e429776b85bcf8017fdf01cd5d60fbba12"}, + {file = "llama_index_embeddings_huggingface-0.4.0.tar.gz", hash = "sha256:ce8f8b30b29cff85401aba2118285fb63fb8147a56b656ee20f7e8510ca085a2"}, ] [[package]] name = "llama-index-embeddings-openai" -version = "0.1.9" -requires_python = "<4.0,>=3.8.1" +version = "0.3.1" +requires_python = "<4.0,>=3.9" summary = "llama-index embeddings openai integration" groups = ["default"] dependencies = [ - "llama-index-core<0.11.0,>=0.10.1", + "llama-index-core<0.13.0,>=0.12.0", + "openai>=1.1.0", ] files = [ - {file = "llama_index_embeddings_openai-0.1.9-py3-none-any.whl", hash = "sha256:fbd16d6197b91f4dbdc6d0707e573cc224ac2b0a48d5b370c6232dd8a2282473"}, - {file = "llama_index_embeddings_openai-0.1.9.tar.gz", hash = "sha256:0fd292b2f9a0ad4534a790d6374726bc885853188087eb018167dcf239643924"}, + {file = "llama_index_embeddings_openai-0.3.1-py3-none-any.whl", hash = "sha256:f15a3d13da9b6b21b8bd51d337197879a453d1605e625a1c6d45e741756c0290"}, + {file = "llama_index_embeddings_openai-0.3.1.tar.gz", hash = "sha256:1368aad3ce24cbaed23d5ad251343cef1eb7b4a06d6563d6606d59cb347fef20"}, ] [[package]] name = "llama-index-indices-managed-llama-cloud" -version = "0.2.5" -requires_python = "<4.0,>=3.8.1" +version = "0.6.3" +requires_python = "<4.0,>=3.9" summary = "llama-index indices llama-cloud integration" groups = ["default"] dependencies = [ - "llama-cloud>=0.0.9", - "llama-index-core<0.11.0,>=0.10.48.post1", + "llama-cloud>=0.1.5", + "llama-index-core<0.13.0,>=0.12.0", ] files = [ - {file = "llama_index_indices_managed_llama_cloud-0.2.5-py3-none-any.whl", hash = "sha256:13329cc64289bc2d485bea8353b7205f01b21bbb7373bb8081a77008ee0bdf58"}, - {file = "llama_index_indices_managed_llama_cloud-0.2.5.tar.gz", hash = "sha256:af1242239b8e7870a2191c97a5248ea9e09f369e2c6dd02285315a496ae151a6"}, + {file = "llama_index_indices_managed_llama_cloud-0.6.3-py3-none-any.whl", hash = "sha256:7f125602f624a2d321b6a4130cd98df35eb8c15818a159390755b2c13068f4ce"}, + {file = "llama_index_indices_managed_llama_cloud-0.6.3.tar.gz", hash = "sha256:f09e4182cbc2a2bd75ae85cebb1681075247f0d91b931b094cac4315386ce87a"}, ] [[package]] @@ -1635,123 +1647,127 @@ files = [ [[package]] name = "llama-index-llms-openai" -version = "0.1.27" -requires_python = "<4.0,>=3.8.1" +version = "0.3.2" +requires_python = "<4.0,>=3.9" summary = "llama-index llms openai integration" groups = ["default"] dependencies = [ - "llama-index-core<0.11.0,>=0.10.57", + "llama-index-core<0.13.0,>=0.12.0", + "openai<2.0.0,>=1.40.0", ] files = [ - {file = "llama_index_llms_openai-0.1.27-py3-none-any.whl", hash = "sha256:8da0e90d4a558667d2b9cf1b3f577a4cb7723b7680ed6d22027b0baf9cd5999e"}, - {file = "llama_index_llms_openai-0.1.27.tar.gz", hash = "sha256:37c2d1159b56607d3a807d90260ee25b4f002086d6251c7272afbc53f2514603"}, + {file = "llama_index_llms_openai-0.3.2-py3-none-any.whl", hash = "sha256:439b8ac8183168156a9724d03e1b3aeeb95d8d3c605b866a6b803b84fae131f6"}, + {file = "llama_index_llms_openai-0.3.2.tar.gz", hash = "sha256:8a443a564e7d12779a9f030cb82fe3243803e217d72410764ac116dd43554fe5"}, ] [[package]] name = "llama-index-multi-modal-llms-openai" -version = "0.1.5" -requires_python = "<4.0,>=3.8.1" +version = "0.3.0" +requires_python = "<4.0,>=3.9" summary = "llama-index multi-modal-llms openai integration" groups = ["default"] dependencies = [ - "llama-index-core<0.11.0,>=0.10.1", - "llama-index-llms-openai<0.2.0,>=0.1.1", + "llama-index-core<0.13.0,>=0.12.0", + "llama-index-llms-openai<0.4.0,>=0.3.0", ] files = [ - {file = "llama_index_multi_modal_llms_openai-0.1.5-py3-none-any.whl", hash = "sha256:bb332580e7e4b5f2f87488b3649d2ceb53ee82c848e59694578a982c3982ce0b"}, - {file = "llama_index_multi_modal_llms_openai-0.1.5.tar.gz", hash = "sha256:9a237f4f886d1e20c27e9493e80b3e1f8753859481ff1b58fe25b7aa39b198a2"}, + {file = "llama_index_multi_modal_llms_openai-0.3.0-py3-none-any.whl", hash = "sha256:9b7e3e39b19b2668b9c75014bcb90795bb546f0f9e1af8b7f1087f8687805763"}, + {file = "llama_index_multi_modal_llms_openai-0.3.0.tar.gz", hash = "sha256:71e983c7771c39088e4058cd78029219315a0fb631b9e12b903e53243b9a3fd6"}, ] [[package]] name = "llama-index-program-openai" -version = "0.1.6" -requires_python = "<4.0,>=3.8.1" +version = "0.3.1" +requires_python = "<4.0,>=3.9" summary = "llama-index program openai integration" groups = ["default"] dependencies = [ - "llama-index-agent-openai<0.3.0,>=0.1.1", - "llama-index-core<0.11.0,>=0.10.1", - "llama-index-llms-openai<0.2.0,>=0.1.1", + "llama-index-agent-openai<0.5.0,>=0.4.0", + "llama-index-core<0.13.0,>=0.12.0", + "llama-index-llms-openai<0.4.0,>=0.3.0", ] files = [ - {file = "llama_index_program_openai-0.1.6-py3-none-any.whl", hash = "sha256:4660b338503537c5edca1e0dab606af6ce372b4f1b597e2833c6b602447c5d8d"}, - {file = "llama_index_program_openai-0.1.6.tar.gz", hash = "sha256:c6a4980c5ea826088b28b4dee3367edb20221e6d05eb0e05019049190131d772"}, + {file = "llama_index_program_openai-0.3.1-py3-none-any.whl", hash = "sha256:93646937395dc5318fd095153d2f91bd632b25215d013d14a87c088887d205f9"}, + {file = "llama_index_program_openai-0.3.1.tar.gz", hash = "sha256:6039a6cdbff62c6388c07e82a157fe2edd3bbef0c5adf292ad8546bf4ec75b82"}, ] [[package]] name = "llama-index-question-gen-openai" -version = "0.1.3" -requires_python = ">=3.8.1,<4.0" +version = "0.3.0" +requires_python = "<4.0,>=3.9" summary = "llama-index question_gen openai integration" groups = ["default"] dependencies = [ - "llama-index-core<0.11.0,>=0.10.1", - "llama-index-llms-openai<0.2.0,>=0.1.1", - "llama-index-program-openai<0.2.0,>=0.1.1", + "llama-index-core<0.13.0,>=0.12.0", + "llama-index-llms-openai<0.4.0,>=0.3.0", + "llama-index-program-openai<0.4.0,>=0.3.0", ] files = [ - {file = "llama_index_question_gen_openai-0.1.3-py3-none-any.whl", hash = "sha256:1f83b49e8b2e665030d1ec8c54687d6985d9fa8426147b64e46628a9e489b302"}, - {file = "llama_index_question_gen_openai-0.1.3.tar.gz", hash = "sha256:4486198117a45457d2e036ae60b93af58052893cc7d78fa9b6f47dd47b81e2e1"}, + {file = "llama_index_question_gen_openai-0.3.0-py3-none-any.whl", hash = "sha256:9b60ec114273a63b50349948666e5744a8f58acb645824e07c979041e8fec598"}, + {file = "llama_index_question_gen_openai-0.3.0.tar.gz", hash = "sha256:efd3b468232808e9d3474670aaeab00e41b90f75f52d0c9bfbf11207e0963d62"}, ] [[package]] name = "llama-index-readers-file" -version = "0.1.19" -requires_python = "<4.0,>=3.8.1" +version = "0.4.0" +requires_python = "<4.0,>=3.9" summary = "llama-index readers file integration" groups = ["default"] dependencies = [ "beautifulsoup4<5.0.0,>=4.12.3", - "llama-index-core<0.11.0,>=0.10.1", - "pypdf<5.0.0,>=4.0.1", + "llama-index-core<0.13.0,>=0.12.0", + "pandas", + "pypdf<6.0.0,>=5.1.0", "striprtf<0.0.27,>=0.0.26", ] files = [ - {file = "llama_index_readers_file-0.1.19-py3-none-any.whl", hash = "sha256:25eb6d066dc38753de435e876ef8511c68d84102302c053b7dcb0776db68fced"}, - {file = "llama_index_readers_file-0.1.19.tar.gz", hash = "sha256:194c1b9b85c265159b7302c7d80adba937aab06f05c170af7fd95c4e7a8aec35"}, + {file = "llama_index_readers_file-0.4.0-py3-none-any.whl", hash = "sha256:437a38d63d4e254168980dd17c6eccde18cb97876fb9fffae9da3dfe6737d0fe"}, + {file = "llama_index_readers_file-0.4.0.tar.gz", hash = "sha256:7828dec1feb7c53e6d3140385f8499c0e7ac746265299384714ddfd163f9d15a"}, ] [[package]] name = "llama-index-readers-llama-parse" -version = "0.1.4" -requires_python = "<4.0,>=3.8.1" +version = "0.4.0" +requires_python = "<4.0,>=3.9" summary = "llama-index readers llama-parse integration" groups = ["default"] dependencies = [ - "llama-index-core<0.11.0,>=0.10.7", - "llama-parse<0.5.0,>=0.4.0", + "llama-index-core<0.13.0,>=0.12.0", + "llama-parse>=0.5.0", ] files = [ - {file = "llama_index_readers_llama_parse-0.1.4-py3-none-any.whl", hash = "sha256:c4914b37d12cceee56fbd185cca80f87d60acbf8ea7a73f9719610180be1fcdd"}, - {file = "llama_index_readers_llama_parse-0.1.4.tar.gz", hash = "sha256:78608b193c818894aefeee0aa303f02b7f80f2e4caf13866c2fd3b0b1023e2c0"}, + {file = "llama_index_readers_llama_parse-0.4.0-py3-none-any.whl", hash = "sha256:574e48386f28d2c86c3f961ca4a4906910312f3400dd0c53014465bfbc6b32bf"}, + {file = "llama_index_readers_llama_parse-0.4.0.tar.gz", hash = "sha256:e99ec56f4f8546d7fda1a7c1ae26162fb9acb7ebcac343b5abdb4234b4644e0f"}, ] [[package]] name = "llama-index-vector-stores-faiss" -version = "0.1.2" -requires_python = ">=3.8.1,<4.0" +version = "0.3.0" +requires_python = "<4.0,>=3.9" summary = "llama-index vector_stores faiss integration" groups = ["default"] dependencies = [ - "llama-index-core<0.11.0,>=0.10.1", + "llama-index-core<0.13.0,>=0.12.0", ] files = [ - {file = "llama_index_vector_stores_faiss-0.1.2-py3-none-any.whl", hash = "sha256:6db18f4f11fdcce9294d7b1725805992006094766b9dcbd3967065f8997f0be9"}, - {file = "llama_index_vector_stores_faiss-0.1.2.tar.gz", hash = "sha256:e462641e4f99ae140a4725103a3d5cad2caf1849cbb782ca405b1a6eb5de65dc"}, + {file = "llama_index_vector_stores_faiss-0.3.0-py3-none-any.whl", hash = "sha256:2148163dba1222c855bd367a7b796bc35d46dc2e77d57bafd321ba14aac00177"}, + {file = "llama_index_vector_stores_faiss-0.3.0.tar.gz", hash = "sha256:c9df99dd00fe7058606ef4fce113535fa30b73edd650136be87c9b5b240df3f9"}, ] [[package]] name = "llama-parse" -version = "0.4.2" +version = "0.5.15" requires_python = "<4.0,>=3.8.1" summary = "Parse files into RAG-Optimized formats." groups = ["default"] dependencies = [ - "llama-index-core>=0.10.29", + "click<9.0.0,>=8.1.7", + "llama-index-core>=0.11.0", + "pydantic!=2.10", ] files = [ - {file = "llama_parse-0.4.2-py3-none-any.whl", hash = "sha256:5ce0390141f216dcd88c1123fea7f2a4f561d177f791a97217a3db3509dec4ff"}, - {file = "llama_parse-0.4.2.tar.gz", hash = "sha256:fa04c09730b102155f6505de9cf91998c86d334581f0f12597c5eb47ca5db859"}, + {file = "llama_parse-0.5.15-py3-none-any.whl", hash = "sha256:7a3506c7d3ae5a8e68c70a457a7213d2698e26abcef1d7a989eb9771cd73ae60"}, + {file = "llama_parse-0.5.15.tar.gz", hash = "sha256:ecb009f71c8b4c657085ca81808a922c80785810e38b10f3b46f03cfd29ba92a"}, ] [[package]] @@ -2666,13 +2682,13 @@ files = [ [[package]] name = "pypdf" -version = "4.2.0" -requires_python = ">=3.6" +version = "5.1.0" +requires_python = ">=3.8" summary = "A pure-python PDF library capable of splitting, merging, cropping, and transforming PDF files" groups = ["default"] files = [ - {file = "pypdf-4.2.0-py3-none-any.whl", hash = "sha256:dc035581664e0ad717e3492acebc1a5fc23dba759e788e3d4a9fc9b1a32e72c1"}, - {file = "pypdf-4.2.0.tar.gz", hash = "sha256:fe63f3f7d1dcda1c9374421a94c1bba6c6f8c4a62173a59b64ffd52058f846b1"}, + {file = "pypdf-5.1.0-py3-none-any.whl", hash = "sha256:3bd4f503f4ebc58bae40d81e81a9176c400cbbac2ba2d877367595fb524dfdfc"}, + {file = "pypdf-5.1.0.tar.gz", hash = "sha256:425a129abb1614183fd1aca6982f650b47f8026867c0ce7c4b9f281c443d2740"}, ] [[package]] diff --git a/pyproject.toml b/pyproject.toml index 09759049..8ea7a31e 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -87,9 +87,10 @@ dependencies = [ "fastapi==0.115.4", "langchain==0.3.6", "langchain-ibm==0.3.2", - "llama-index==0.10.62", - "llama-index-vector-stores-faiss==0.1.2", - "llama-index-embeddings-huggingface==0.2.2", + "llama-index==0.12.2", + "llama-index-core==0.12.2", + "llama-index-vector-stores-faiss==0.3.0", + "llama-index-embeddings-huggingface==0.4.0", "uvicorn==0.32.0", "redis==5.0.8", "faiss-cpu==1.8.0.post1",