diff --git a/CHANGELOG.md b/CHANGELOG.md index 8ffa502582604..53bb5f470b276 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,95 @@ # ChangeLog +## [2024-12-17] + +### `llama-index-core` [0.12.6] + +- [bug fix] Ensure that StopEvent gets cleared from Context.\_in_progress["_done"] after a Workflow run (#17300) +- fix: add a timeout to langchain callback handler (#17296) +- tweak User vs tool in react prompts (#17273) +- refact: Refactor Document to be natively multimodal (#17204) +- fix: make ImageDocument derive from Document, backward compatible (#17259) +- fix: accept already base64-encoded data in ImageBlock (#17244) +- fix(metrics): fixed NDCG calculation and updated previous tests (#17236) +- fix: remove llama-index-legacy dependency in llama-index-core (#17231) +- Refined the default documentation generation for function tools (#17208) + +### `llama-index-embeddings-voyageai` [0.3.3] + +- add support for voyageai >=0.3.0 (#17120) +- Introducting VoyageAI's new multimodal embeddings model (#17261) +- VoyageAI multimodal embedding, correction (#17284) + +### `llama-index-experimental` [0.5.2] + +- Fixed import errors for experimental JSONalyzeQueryEngine (#17228) + +### `llama-index-grapg-stores-neo4j` [0.4.4] + +- Add cypher corrector and allow graph schema filtering (#17223) +- Add timeout config to neo4j graph (#17267) +- Add text and embedding type to neo4j enhanced schema (#17289) + +### `llama-index-llms-anthropic` [0.6.3] + +- add content blocks to anthropic (#17274) +- Do not send blank content to anthropic (#17278) +- Update anthropic type imports for v0.41.0 release (#17299) +- Fix Anthropic tokenizer protocol (fix by Devin) (#17201) + +### `llama-index-llms-bedrock` [0.3.3] + +- Add Amazon bedrock guardrails (#17281) + +### `llama-index-llms-bedrock-converse` [0.4.2] + +- Add Amazon bedrock guardrails (#17281) + +### `llama-index-llms-gemini` [0.4.1] + +- Gemini 2.0 support (#17249) + +### `llama-index-llms-mistralai` [0.3.1] + +- add tool call id/name to mistral chat messages (#17280) + +### `llama-index-llms-nvidia` [0.3.1] + +- Adding llama 3.3-70b as function-calling-capable (#17253) + +### `llama-index-llms-openai` [0.3.10] + +- fix openai message dicts for tool calls (#17254) + +### `llama-index-llms-text-generation-inference` [0.3.1] + +- Fix: TGI context window (#17252) + +### `llama-index-multi-modal-llms-anthropic` [0.3.1] + +- handle more response types in anthropic multi modal llms (#17302) + +### `llama-index-readers-confluence` [0.3.1] + +- Support Confluence cookies (#17276) + +### `llama-index-vector-stores-milvus` [0.4.0] + +- Parse "milvus_search_config" out of "vector_store_kwargs" (#17221) +- refactor and optimize milvus code (#17229) + +### `llama-index-vector-stores-pinecone` [0.4.2] + +- Handle empty retrieved Pinecone index values (#17242) + +### `llama-index-vector-stores-qdrant` [0.4.1] + +- feat: Add NOT filter condition to MetadataFilter and QdrantVectorStore (#17270) + +### `llama-index-vector-stores-weaviate` [1.3.0] + +- Add async support to weaviate vector store integration (#17220) + ## [2024-12-09] ### `llama-index-core` [0.12.5] @@ -14,6 +104,10 @@ - feat: integration on pinecone hosted rerankers (#17192) +### `llama-index-tools-scrapegraph` [0.1.0] + +- Add Scrapegraph tool integration (#17238) + ### `llama-index-vector-stores-postgres` [0.3.3] - Update pgvector dependency to version 0.3.6 (#17195) diff --git a/docs/docs/CHANGELOG.md b/docs/docs/CHANGELOG.md index 8ffa502582604..792f06c8a6367 100644 --- a/docs/docs/CHANGELOG.md +++ b/docs/docs/CHANGELOG.md @@ -1,5 +1,95 @@ # ChangeLog +## [2024-12-17] + +### `llama-index-core` [0.12.6] + +- [bug fix] Ensure that StopEvent gets cleared from Context._in_progress["_done"] after a Workflow run (#17300) +- fix: add a timeout to langchain callback handler (#17296) +- tweak User vs tool in react prompts (#17273) +- refact: Refactor Document to be natively multimodal (#17204) +- fix: make ImageDocument derive from Document, backward compatible (#17259) +- fix: accept already base64-encoded data in ImageBlock (#17244) +- fix(metrics): fixed NDCG calculation and updated previous tests (#17236) +- fix: remove llama-index-legacy dependency in llama-index-core (#17231) +- Refined the default documentation generation for function tools (#17208) + +### `llama-index-embeddings-voyageai` [0.3.3] + +- add support for voyageai >=0.3.0 (#17120) +- Introducting VoyageAI's new multimodal embeddings model (#17261) +- VoyageAI multimodal embedding, correction (#17284) + +### `llama-index-experimental` [0.5.2] + +- Fixed import errors for experimental JSONalyzeQueryEngine (#17228) + +### `llama-index-grapg-stores-neo4j` [0.4.4] + +- Add cypher corrector and allow graph schema filtering (#17223) +- Add timeout config to neo4j graph (#17267) +- Add text and embedding type to neo4j enhanced schema (#17289) + +### `llama-index-llms-anthropic` [0.6.3] + +- add content blocks to anthropic (#17274) +- Do not send blank content to anthropic (#17278) +- Update anthropic type imports for v0.41.0 release (#17299) +- Fix Anthropic tokenizer protocol (fix by Devin) (#17201) + +### `llama-index-llms-bedrock` [0.3.3] + +- Add Amazon bedrock guardrails (#17281) + +### `llama-index-llms-bedrock-converse` [0.4.2] + +- Add Amazon bedrock guardrails (#17281) + +### `llama-index-llms-gemini` [0.4.1] + +- Gemini 2.0 support (#17249) + +### `llama-index-llms-mistralai` [0.3.1] + +- add tool call id/name to mistral chat messages (#17280) + +### `llama-index-llms-nvidia` [0.3.1] + +- Adding llama 3.3-70b as function-calling-capable (#17253) + +### `llama-index-llms-openai` [0.3.10] + +- fix openai message dicts for tool calls (#17254) + +### `llama-index-llms-text-generation-inference` [0.3.1] + +- Fix: TGI context window (#17252) + +### `llama-index-multi-modal-llms-anthropic` [0.3.1] + +- handle more response types in anthropic multi modal llms (#17302) + +### `llama-index-readers-confluence` [0.3.1] + +- Support Confluence cookies (#17276) + +### `llama-index-vector-stores-milvus` [0.4.0] + +- Parse "milvus_search_config" out of "vector_store_kwargs" (#17221) +- refactor and optimize milvus code (#17229) + +### `llama-index-vector-stores-pinecone` [0.4.2] + +- Handle empty retrieved Pinecone index values (#17242) + +### `llama-index-vector-stores-qdrant` [0.4.1] + +- feat: Add NOT filter condition to MetadataFilter and QdrantVectorStore (#17270) + +### `llama-index-vector-stores-weaviate` [1.3.0] + +- Add async support to weaviate vector store integration (#17220) + ## [2024-12-09] ### `llama-index-core` [0.12.5] @@ -14,6 +104,10 @@ - feat: integration on pinecone hosted rerankers (#17192) +### `llama-index-tools-scrapegraph` [0.1.0] + +- Add Scrapegraph tool integration (#17238) + ### `llama-index-vector-stores-postgres` [0.3.3] - Update pgvector dependency to version 0.3.6 (#17195) diff --git a/docs/docs/api_reference/tools/scrapegraph.md b/docs/docs/api_reference/tools/scrapegraph.md new file mode 100644 index 0000000000000..8cca961a5502b --- /dev/null +++ b/docs/docs/api_reference/tools/scrapegraph.md @@ -0,0 +1,4 @@ +::: llama_index.tools.scrapegraph + options: + members: + - ScrapegraphToolSpec diff --git a/docs/mkdocs.yml b/docs/mkdocs.yml index c34877e952aa2..01307c0862fbb 100644 --- a/docs/mkdocs.yml +++ b/docs/mkdocs.yml @@ -151,6 +151,7 @@ nav: - ./examples/cookbooks/contextual_retrieval.ipynb - ./examples/cookbooks/crewai_llamaindex.ipynb - ./examples/cookbooks/llama3_cookbook.ipynb + - ./examples/cookbooks/llama3_cookbook_gaudi.ipynb - ./examples/cookbooks/llama3_cookbook_groq.ipynb - ./examples/cookbooks/llama3_cookbook_ollama_replicate.ipynb - ./examples/cookbooks/mistralai.ipynb @@ -256,7 +257,6 @@ nav: - ./examples/embeddings/nomic.ipynb - ./examples/embeddings/nvidia.ipynb - ./examples/embeddings/oci_genai.ipynb - - ./examples/embeddings/octoai.ipynb - ./examples/embeddings/ollama_embedding.ipynb - ./examples/embeddings/openvino.ipynb - ./examples/embeddings/optimum_intel.ipynb @@ -312,6 +312,7 @@ nav: - ./examples/ingestion/document_management_pipeline.ipynb - ./examples/ingestion/ingestion_gdrive.ipynb - ./examples/ingestion/parallel_execution_ingestion_pipeline.ipynb + - ./examples/ingestion/redis_ingestion_pipeline.ipynb - LLMs: - ./examples/llm/ai21.ipynb - ./examples/llm/alephalpha.ipynb @@ -384,9 +385,7 @@ nav: - ./examples/llm/rungpt.ipynb - ./examples/llm/sagemaker_endpoint_llm.ipynb - ./examples/llm/sambanovasystems.ipynb - - ./examples/llm/solar.ipynb - ./examples/llm/together.ipynb - - ./examples/llm/unify.ipynb - ./examples/llm/upstage.ipynb - ./examples/llm/vertex.ipynb - ./examples/llm/vicuna.ipynb @@ -642,7 +641,6 @@ nav: - ./examples/vector_stores/LanternAutoRetriever.ipynb - ./examples/vector_stores/LanternIndexDemo.ipynb - ./examples/vector_stores/LindormDemo.ipynb - - ./examples/vector_stores/MetalIndexDemo.ipynb - ./examples/vector_stores/MilvusHybridIndexDemo.ipynb - ./examples/vector_stores/MilvusIndexDemo.ipynb - ./examples/vector_stores/MilvusOperatorFunctionDemo.ipynb @@ -1647,6 +1645,7 @@ nav: - ./api_reference/tools/requests.md - ./api_reference/tools/retriever.md - ./api_reference/tools/salesforce.md + - ./api_reference/tools/scrapegraph.md - ./api_reference/tools/shopify.md - ./api_reference/tools/slack.md - ./api_reference/tools/tavily_research.md @@ -2308,6 +2307,7 @@ plugins: - ../llama-index-integrations/llms/llama-index-llms-nebius - ../llama-index-integrations/postprocessor/llama-index-postprocessor-bedrock-rerank - ../llama-index-integrations/postprocessor/llama-index-postprocessor-pinecone-native-rerank + - ../llama-index-integrations/tools/llama-index-tools-scrapegraph - redirects: redirect_maps: ./api/llama_index.vector_stores.MongoDBAtlasVectorSearch.html: api_reference/storage/vector_store/mongodb.md diff --git a/llama-index-core/llama_index/core/__init__.py b/llama-index-core/llama_index/core/__init__.py index dbf98b1a2b8ef..d9b1df7e718e5 100644 --- a/llama-index-core/llama_index/core/__init__.py +++ b/llama-index-core/llama_index/core/__init__.py @@ -1,6 +1,6 @@ """Init file of LlamaIndex.""" -__version__ = "0.12.5" +__version__ = "0.12.6" import logging from logging import NullHandler diff --git a/llama-index-core/pyproject.toml b/llama-index-core/pyproject.toml index 13cf7ef93b6fb..bea2c3679bc0c 100644 --- a/llama-index-core/pyproject.toml +++ b/llama-index-core/pyproject.toml @@ -46,7 +46,7 @@ name = "llama-index-core" packages = [{include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.12.5" +version = "0.12.6" [tool.poetry.dependencies] SQLAlchemy = {extras = ["asyncio"], version = ">=1.4.49"} diff --git a/llama-index-integrations/vector_stores/llama-index-vector-stores-nile/pyproject.toml b/llama-index-integrations/vector_stores/llama-index-vector-stores-nile/pyproject.toml index 7d03f26100db1..c97d58c8c0dd5 100644 --- a/llama-index-integrations/vector_stores/llama-index-vector-stores-nile/pyproject.toml +++ b/llama-index-integrations/vector_stores/llama-index-vector-stores-nile/pyproject.toml @@ -27,11 +27,11 @@ exclude = ["**/BUILD"] license = "MIT" name = "llama-index-vector-stores-nile" readme = "README.md" -version = "0.2.0" +version = "0.2.1" [tool.poetry.dependencies] python = ">=3.9,<4.0" -llama-index-core = "^0.12.0" +llama-index-core = "^0.12.6" psycopg = "^3.2" [tool.poetry.group.dev.dependencies] diff --git a/llama-index-integrations/vector_stores/llama-index-vector-stores-postgres/pyproject.toml b/llama-index-integrations/vector_stores/llama-index-vector-stores-postgres/pyproject.toml index 3b3c1d9922893..86ffd5af4aa9d 100644 --- a/llama-index-integrations/vector_stores/llama-index-vector-stores-postgres/pyproject.toml +++ b/llama-index-integrations/vector_stores/llama-index-vector-stores-postgres/pyproject.toml @@ -27,14 +27,14 @@ exclude = ["**/BUILD"] license = "MIT" name = "llama-index-vector-stores-postgres" readme = "README.md" -version = "0.4.0" +version = "0.4.1" [tool.poetry.dependencies] python = ">=3.9,<4.0" pgvector = ">=0.3.6,<1.0.0" psycopg2-binary = ">=2.9.9,<3.0.0" asyncpg = ">=0.29.0,<1.0.0" -llama-index-core = "^0.12.0" +llama-index-core = "^0.12.6" [tool.poetry.dependencies.sqlalchemy] extras = ["asyncio"] diff --git a/llama-index-integrations/vector_stores/llama-index-vector-stores-qdrant/pyproject.toml b/llama-index-integrations/vector_stores/llama-index-vector-stores-qdrant/pyproject.toml index edbef55571db3..3a1c012c8f8fb 100644 --- a/llama-index-integrations/vector_stores/llama-index-vector-stores-qdrant/pyproject.toml +++ b/llama-index-integrations/vector_stores/llama-index-vector-stores-qdrant/pyproject.toml @@ -27,13 +27,13 @@ exclude = ["**/BUILD"] license = "MIT" name = "llama-index-vector-stores-qdrant" readme = "README.md" -version = "0.4.0" +version = "0.4.1" [tool.poetry.dependencies] python = ">=3.9,<3.13" qdrant-client = ">=1.7.1" grpcio = "^1.60.0" -llama-index-core = "^0.12.0" +llama-index-core = "^0.12.6" [tool.poetry.extras] fastembed = ["fastembed"] diff --git a/poetry.lock b/poetry.lock index 03a9ab55b6ef3..3ba1832f8d7ae 100644 --- a/poetry.lock +++ b/poetry.lock @@ -111,13 +111,13 @@ speedups = ["Brotli", "aiodns (>=3.2.0)", "brotlicffi"] [[package]] name = "aiosignal" -version = "1.3.1" +version = "1.3.2" description = "aiosignal: a list of registered asynchronous callbacks" optional = false -python-versions = ">=3.7" +python-versions = ">=3.9" files = [ - {file = "aiosignal-1.3.1-py3-none-any.whl", hash = "sha256:f8376fb07dd1e86a584e4fcdec80b36b7f81aac666ebc724e2c090300dd83b17"}, - {file = "aiosignal-1.3.1.tar.gz", hash = "sha256:54cd96e15e1649b75d6c87526a6ff0b6c1b0dd3459f43d9ca11d48c339b68cfc"}, + {file = "aiosignal-1.3.2-py2.py3-none-any.whl", hash = "sha256:45cde58e409a301715980c2b01d0c28bdde3770d8290b5eb2173759d9acb31a5"}, + {file = "aiosignal-1.3.2.tar.gz", hash = "sha256:a8c255c66fafb1e499c9351d0bf32ff2d8a0321595ebac3b93713656d2436f54"}, ] [package.dependencies] @@ -214,19 +214,19 @@ files = [ [[package]] name = "attrs" -version = "24.2.0" +version = "24.3.0" description = "Classes Without Boilerplate" optional = false -python-versions = ">=3.7" +python-versions = ">=3.8" files = [ - {file = "attrs-24.2.0-py3-none-any.whl", hash = "sha256:81921eb96de3191c8258c199618104dd27ac608d9366f5e35d011eae1867ede2"}, - {file = "attrs-24.2.0.tar.gz", hash = "sha256:5cfb1b9148b5b086569baec03f20d7b6bf3bcacc9a42bebf87ffaaca362f6346"}, + {file = "attrs-24.3.0-py3-none-any.whl", hash = "sha256:ac96cd038792094f438ad1f6ff80837353805ac950cd2aa0e0625ef19850c308"}, + {file = "attrs-24.3.0.tar.gz", hash = "sha256:8f5c07333d543103541ba7be0e2ce16eeee8130cb0b3f9238ab904ce1e85baff"}, ] [package.extras] benchmark = ["cloudpickle", "hypothesis", "mypy (>=1.11.1)", "pympler", "pytest (>=4.3.0)", "pytest-codspeed", "pytest-mypy-plugins", "pytest-xdist[psutil]"] cov = ["cloudpickle", "coverage[toml] (>=5.3)", "hypothesis", "mypy (>=1.11.1)", "pympler", "pytest (>=4.3.0)", "pytest-mypy-plugins", "pytest-xdist[psutil]"] -dev = ["cloudpickle", "hypothesis", "mypy (>=1.11.1)", "pre-commit", "pympler", "pytest (>=4.3.0)", "pytest-mypy-plugins", "pytest-xdist[psutil]"] +dev = ["cloudpickle", "hypothesis", "mypy (>=1.11.1)", "pre-commit-uv", "pympler", "pytest (>=4.3.0)", "pytest-mypy-plugins", "pytest-xdist[psutil]"] docs = ["cogapp", "furo", "myst-parser", "sphinx", "sphinx-notfound-page", "sphinxcontrib-towncrier", "towncrier (<24.7)"] tests = ["cloudpickle", "hypothesis", "mypy (>=1.11.1)", "pympler", "pytest (>=4.3.0)", "pytest-mypy-plugins", "pytest-xdist[psutil]"] tests-mypy = ["mypy (>=1.11.1)", "pytest-mypy-plugins"] @@ -333,13 +333,13 @@ css = ["tinycss2 (>=1.1.0,<1.5)"] [[package]] name = "certifi" -version = "2024.8.30" +version = "2024.12.14" description = "Python package for providing Mozilla's CA Bundle." optional = false python-versions = ">=3.6" files = [ - {file = "certifi-2024.8.30-py3-none-any.whl", hash = "sha256:922820b53db7a7257ffbda3f597266d435245903d80737e34f8a45ff3e3230d8"}, - {file = "certifi-2024.8.30.tar.gz", hash = "sha256:bec941d2aa8195e248a60b31ff9f0558284cf01a52591ceda73ea9afffd69fd9"}, + {file = "certifi-2024.12.14-py3-none-any.whl", hash = "sha256:1275f7a45be9464efc1173084eaa30f866fe2e47d389406136d332ed4967ec56"}, + {file = "certifi-2024.12.14.tar.gz", hash = "sha256:b650d30f370c2b724812bee08008be0c4163b163ddaec3f2546c1caf65f191db"}, ] [[package]] @@ -674,37 +674,37 @@ typing-inspect = ">=0.4.0,<1" [[package]] name = "debugpy" -version = "1.8.9" +version = "1.8.11" description = "An implementation of the Debug Adapter Protocol for Python" optional = false python-versions = ">=3.8" files = [ - {file = "debugpy-1.8.9-cp310-cp310-macosx_14_0_x86_64.whl", hash = "sha256:cfe1e6c6ad7178265f74981edf1154ffce97b69005212fbc90ca22ddfe3d017e"}, - {file = "debugpy-1.8.9-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ada7fb65102a4d2c9ab62e8908e9e9f12aed9d76ef44880367bc9308ebe49a0f"}, - {file = "debugpy-1.8.9-cp310-cp310-win32.whl", hash = "sha256:c36856343cbaa448171cba62a721531e10e7ffb0abff838004701454149bc037"}, - {file = "debugpy-1.8.9-cp310-cp310-win_amd64.whl", hash = "sha256:17c5e0297678442511cf00a745c9709e928ea4ca263d764e90d233208889a19e"}, - {file = "debugpy-1.8.9-cp311-cp311-macosx_14_0_universal2.whl", hash = "sha256:b74a49753e21e33e7cf030883a92fa607bddc4ede1aa4145172debc637780040"}, - {file = "debugpy-1.8.9-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:62d22dacdb0e296966d7d74a7141aaab4bec123fa43d1a35ddcb39bf9fd29d70"}, - {file = "debugpy-1.8.9-cp311-cp311-win32.whl", hash = "sha256:8138efff315cd09b8dcd14226a21afda4ca582284bf4215126d87342bba1cc66"}, - {file = "debugpy-1.8.9-cp311-cp311-win_amd64.whl", hash = "sha256:ff54ef77ad9f5c425398efb150239f6fe8e20c53ae2f68367eba7ece1e96226d"}, - {file = "debugpy-1.8.9-cp312-cp312-macosx_14_0_universal2.whl", hash = "sha256:957363d9a7a6612a37458d9a15e72d03a635047f946e5fceee74b50d52a9c8e2"}, - {file = "debugpy-1.8.9-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5e565fc54b680292b418bb809f1386f17081d1346dca9a871bf69a8ac4071afe"}, - {file = "debugpy-1.8.9-cp312-cp312-win32.whl", hash = "sha256:3e59842d6c4569c65ceb3751075ff8d7e6a6ada209ceca6308c9bde932bcef11"}, - {file = "debugpy-1.8.9-cp312-cp312-win_amd64.whl", hash = "sha256:66eeae42f3137eb428ea3a86d4a55f28da9bd5a4a3d369ba95ecc3a92c1bba53"}, - {file = "debugpy-1.8.9-cp313-cp313-macosx_14_0_universal2.whl", hash = "sha256:957ecffff80d47cafa9b6545de9e016ae8c9547c98a538ee96ab5947115fb3dd"}, - {file = "debugpy-1.8.9-cp313-cp313-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1efbb3ff61487e2c16b3e033bc8595aea578222c08aaf3c4bf0f93fadbd662ee"}, - {file = "debugpy-1.8.9-cp313-cp313-win32.whl", hash = "sha256:7c4d65d03bee875bcb211c76c1d8f10f600c305dbd734beaed4077e902606fee"}, - {file = "debugpy-1.8.9-cp313-cp313-win_amd64.whl", hash = "sha256:e46b420dc1bea64e5bbedd678148be512442bc589b0111bd799367cde051e71a"}, - {file = "debugpy-1.8.9-cp38-cp38-macosx_14_0_x86_64.whl", hash = "sha256:472a3994999fe6c0756945ffa359e9e7e2d690fb55d251639d07208dbc37caea"}, - {file = "debugpy-1.8.9-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:365e556a4772d7d0d151d7eb0e77ec4db03bcd95f26b67b15742b88cacff88e9"}, - {file = "debugpy-1.8.9-cp38-cp38-win32.whl", hash = "sha256:54a7e6d3014c408eb37b0b06021366ee985f1539e12fe49ca2ee0d392d9ceca5"}, - {file = "debugpy-1.8.9-cp38-cp38-win_amd64.whl", hash = "sha256:8e99c0b1cc7bf86d83fb95d5ccdc4ad0586d4432d489d1f54e4055bcc795f693"}, - {file = "debugpy-1.8.9-cp39-cp39-macosx_14_0_x86_64.whl", hash = "sha256:7e8b079323a56f719977fde9d8115590cb5e7a1cba2fcee0986ef8817116e7c1"}, - {file = "debugpy-1.8.9-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6953b335b804a41f16a192fa2e7851bdcfd92173cbb2f9f777bb934f49baab65"}, - {file = "debugpy-1.8.9-cp39-cp39-win32.whl", hash = "sha256:7e646e62d4602bb8956db88b1e72fe63172148c1e25c041e03b103a25f36673c"}, - {file = "debugpy-1.8.9-cp39-cp39-win_amd64.whl", hash = "sha256:3d9755e77a2d680ce3d2c5394a444cf42be4a592caaf246dbfbdd100ffcf7ae5"}, - {file = "debugpy-1.8.9-py2.py3-none-any.whl", hash = "sha256:cc37a6c9987ad743d9c3a14fa1b1a14b7e4e6041f9dd0c8abf8895fe7a97b899"}, - {file = "debugpy-1.8.9.zip", hash = "sha256:1339e14c7d980407248f09824d1b25ff5c5616651689f1e0f0e51bdead3ea13e"}, + {file = "debugpy-1.8.11-cp310-cp310-macosx_14_0_x86_64.whl", hash = "sha256:2b26fefc4e31ff85593d68b9022e35e8925714a10ab4858fb1b577a8a48cb8cd"}, + {file = "debugpy-1.8.11-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:61bc8b3b265e6949855300e84dc93d02d7a3a637f2aec6d382afd4ceb9120c9f"}, + {file = "debugpy-1.8.11-cp310-cp310-win32.whl", hash = "sha256:c928bbf47f65288574b78518449edaa46c82572d340e2750889bbf8cd92f3737"}, + {file = "debugpy-1.8.11-cp310-cp310-win_amd64.whl", hash = "sha256:8da1db4ca4f22583e834dcabdc7832e56fe16275253ee53ba66627b86e304da1"}, + {file = "debugpy-1.8.11-cp311-cp311-macosx_14_0_universal2.whl", hash = "sha256:85de8474ad53ad546ff1c7c7c89230db215b9b8a02754d41cb5a76f70d0be296"}, + {file = "debugpy-1.8.11-cp311-cp311-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8ffc382e4afa4aee367bf413f55ed17bd91b191dcaf979890af239dda435f2a1"}, + {file = "debugpy-1.8.11-cp311-cp311-win32.whl", hash = "sha256:40499a9979c55f72f4eb2fc38695419546b62594f8af194b879d2a18439c97a9"}, + {file = "debugpy-1.8.11-cp311-cp311-win_amd64.whl", hash = "sha256:987bce16e86efa86f747d5151c54e91b3c1e36acc03ce1ddb50f9d09d16ded0e"}, + {file = "debugpy-1.8.11-cp312-cp312-macosx_14_0_universal2.whl", hash = "sha256:84e511a7545d11683d32cdb8f809ef63fc17ea2a00455cc62d0a4dbb4ed1c308"}, + {file = "debugpy-1.8.11-cp312-cp312-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ce291a5aca4985d82875d6779f61375e959208cdf09fcec40001e65fb0a54768"}, + {file = "debugpy-1.8.11-cp312-cp312-win32.whl", hash = "sha256:28e45b3f827d3bf2592f3cf7ae63282e859f3259db44ed2b129093ca0ac7940b"}, + {file = "debugpy-1.8.11-cp312-cp312-win_amd64.whl", hash = "sha256:44b1b8e6253bceada11f714acf4309ffb98bfa9ac55e4fce14f9e5d4484287a1"}, + {file = "debugpy-1.8.11-cp313-cp313-macosx_14_0_universal2.whl", hash = "sha256:8988f7163e4381b0da7696f37eec7aca19deb02e500245df68a7159739bbd0d3"}, + {file = "debugpy-1.8.11-cp313-cp313-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c1f6a173d1140e557347419767d2b14ac1c9cd847e0b4c5444c7f3144697e4e"}, + {file = "debugpy-1.8.11-cp313-cp313-win32.whl", hash = "sha256:bb3b15e25891f38da3ca0740271e63ab9db61f41d4d8541745cfc1824252cb28"}, + {file = "debugpy-1.8.11-cp313-cp313-win_amd64.whl", hash = "sha256:d8768edcbeb34da9e11bcb8b5c2e0958d25218df7a6e56adf415ef262cd7b6d1"}, + {file = "debugpy-1.8.11-cp38-cp38-macosx_14_0_x86_64.whl", hash = "sha256:ad7efe588c8f5cf940f40c3de0cd683cc5b76819446abaa50dc0829a30c094db"}, + {file = "debugpy-1.8.11-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:189058d03a40103a57144752652b3ab08ff02b7595d0ce1f651b9acc3a3a35a0"}, + {file = "debugpy-1.8.11-cp38-cp38-win32.whl", hash = "sha256:32db46ba45849daed7ccf3f2e26f7a386867b077f39b2a974bb5c4c2c3b0a280"}, + {file = "debugpy-1.8.11-cp38-cp38-win_amd64.whl", hash = "sha256:116bf8342062246ca749013df4f6ea106f23bc159305843491f64672a55af2e5"}, + {file = "debugpy-1.8.11-cp39-cp39-macosx_14_0_x86_64.whl", hash = "sha256:654130ca6ad5de73d978057eaf9e582244ff72d4574b3e106fb8d3d2a0d32458"}, + {file = "debugpy-1.8.11-cp39-cp39-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:23dc34c5e03b0212fa3c49a874df2b8b1b8fda95160bd79c01eb3ab51ea8d851"}, + {file = "debugpy-1.8.11-cp39-cp39-win32.whl", hash = "sha256:52d8a3166c9f2815bfae05f386114b0b2d274456980d41f320299a8d9a5615a7"}, + {file = "debugpy-1.8.11-cp39-cp39-win_amd64.whl", hash = "sha256:52c3cf9ecda273a19cc092961ee34eb9ba8687d67ba34cc7b79a521c1c64c4c0"}, + {file = "debugpy-1.8.11-py2.py3-none-any.whl", hash = "sha256:0e22f846f4211383e6a416d04b4c13ed174d24cc5d43f5fd52e7821d0ebc8920"}, + {file = "debugpy-1.8.11.tar.gz", hash = "sha256:6ad2688b69235c43b020e04fecccdf6a96c8943ca9c2fb340b8adc103c655e57"}, ] [[package]] @@ -1559,13 +1559,13 @@ files = [ [[package]] name = "jupytext" -version = "1.16.4" +version = "1.16.6" description = "Jupyter notebooks as Markdown documents, Julia, Python or R scripts" optional = false python-versions = ">=3.8" files = [ - {file = "jupytext-1.16.4-py3-none-any.whl", hash = "sha256:76989d2690e65667ea6fb411d8056abe7cd0437c07bd774660b83d62acf9490a"}, - {file = "jupytext-1.16.4.tar.gz", hash = "sha256:28e33f46f2ce7a41fb9d677a4a2c95327285579b64ca104437c4b9eb1e4174e9"}, + {file = "jupytext-1.16.6-py3-none-any.whl", hash = "sha256:900132031f73fee15a1c9ebd862e05eb5f51e1ad6ab3a2c6fdd97ce2f9c913b4"}, + {file = "jupytext-1.16.6.tar.gz", hash = "sha256:dbd03f9263c34b737003f388fc069e9030834fb7136879c4c32c32473557baa0"}, ] [package.dependencies] @@ -1577,11 +1577,11 @@ pyyaml = "*" tomli = {version = "*", markers = "python_version < \"3.11\""} [package.extras] -dev = ["autopep8", "black", "flake8", "gitpython", "ipykernel", "isort", "jupyter-fs (>=1.0)", "jupyter-server (!=2.11)", "nbconvert", "pre-commit", "pytest", "pytest-cov (>=2.6.1)", "pytest-randomly", "pytest-xdist", "sphinx-gallery (<0.8)"] +dev = ["autopep8", "black", "flake8", "gitpython", "ipykernel", "isort", "jupyter-fs (>=1.0)", "jupyter-server (!=2.11)", "nbconvert", "pre-commit", "pytest", "pytest-cov (>=2.6.1)", "pytest-randomly", "pytest-xdist", "sphinx (<8)", "sphinx-gallery (<0.8)"] docs = ["myst-parser", "sphinx", "sphinx-copybutton", "sphinx-rtd-theme"] test = ["pytest", "pytest-randomly", "pytest-xdist"] test-cov = ["ipykernel", "jupyter-server (!=2.11)", "nbconvert", "pytest", "pytest-cov (>=2.6.1)", "pytest-randomly", "pytest-xdist"] -test-external = ["autopep8", "black", "flake8", "gitpython", "ipykernel", "isort", "jupyter-fs (>=1.0)", "jupyter-server (!=2.11)", "nbconvert", "pre-commit", "pytest", "pytest-randomly", "pytest-xdist", "sphinx-gallery (<0.8)"] +test-external = ["autopep8", "black", "flake8", "gitpython", "ipykernel", "isort", "jupyter-fs (>=1.0)", "jupyter-server (!=2.11)", "nbconvert", "pre-commit", "pytest", "pytest-randomly", "pytest-xdist", "sphinx (<8)", "sphinx-gallery (<0.8)"] test-functional = ["pytest", "pytest-randomly", "pytest-xdist"] test-integration = ["ipykernel", "jupyter-server (!=2.11)", "nbconvert", "pytest", "pytest-randomly", "pytest-xdist"] test-ui = ["calysto-bash"] @@ -1681,13 +1681,13 @@ llama-index-llms-openai = ">=0.3.0,<0.4.0" [[package]] name = "llama-index-core" -version = "0.12.5" +version = "0.12.6" description = "Interface between LLMs and your data" optional = false python-versions = "<4.0,>=3.9" files = [ - {file = "llama_index_core-0.12.5-py3-none-any.whl", hash = "sha256:1fe6dd39b2dc5a945b4702d780a2f5962a553e187524a255429461dc92a664db"}, - {file = "llama_index_core-0.12.5.tar.gz", hash = "sha256:1d967323891920579fad3d6497587c137894df3f76718a3ec134f9201f2f4fc0"}, + {file = "llama_index_core-0.12.6-py3-none-any.whl", hash = "sha256:31d72b61c9f582bb879bf110e09a97b0106acdee155ce92cbbeed4df74ce0b8c"}, + {file = "llama_index_core-0.12.6.tar.gz", hash = "sha256:fb380db6472d1d5e25297676b4e82a7644890fcc4d264ab10ac12490ebf359d4"}, ] [package.dependencies] @@ -1747,13 +1747,13 @@ llama-index-core = ">=0.12.0,<0.13.0" [[package]] name = "llama-index-llms-openai" -version = "0.3.4" +version = "0.3.10" description = "llama-index llms openai integration" optional = false python-versions = "<4.0,>=3.9" files = [ - {file = "llama_index_llms_openai-0.3.4-py3-none-any.whl", hash = "sha256:48fa34ce53330829bc4bd99a3da7137c0d4739a1598fe4029c0ea6cfbc10fda4"}, - {file = "llama_index_llms_openai-0.3.4.tar.gz", hash = "sha256:e0fd6faf43e4bcaf99ebbff1633eb6a51c9727698605118d81f0bf7881098a00"}, + {file = "llama_index_llms_openai-0.3.10-py3-none-any.whl", hash = "sha256:ceb344f4a1424970096652135d9c2663a50fa53eb6d24d146c7f989d8c098e2e"}, + {file = "llama_index_llms_openai-0.3.10.tar.gz", hash = "sha256:fda3345aa4ba41b25d92817ee8a5699931e79c7c879243f172d6e0e6cbc5c5b3"}, ] [package.dependencies] @@ -2166,13 +2166,13 @@ pygments = ">2.12.0" [[package]] name = "mkdocs-material" -version = "9.5.48" +version = "9.5.49" description = "Documentation that simply works" optional = false python-versions = ">=3.8" files = [ - {file = "mkdocs_material-9.5.48-py3-none-any.whl", hash = "sha256:b695c998f4b939ce748adbc0d3bff73fa886a670ece948cf27818fa115dc16f8"}, - {file = "mkdocs_material-9.5.48.tar.gz", hash = "sha256:a582531e8b34f4c7ed38c29d5c44763053832cf2a32f7409567e0c74749a47db"}, + {file = "mkdocs_material-9.5.49-py3-none-any.whl", hash = "sha256:c3c2d8176b18198435d3a3e119011922f3e11424074645c24019c2dcf08a360e"}, + {file = "mkdocs_material-9.5.49.tar.gz", hash = "sha256:3671bb282b4f53a1c72e08adbe04d2481a98f85fed392530051f80ff94a9621d"}, ] [package.dependencies] @@ -2627,13 +2627,13 @@ files = [ [[package]] name = "openai" -version = "1.57.2" +version = "1.58.1" description = "The official Python library for the openai API" optional = false python-versions = ">=3.8" files = [ - {file = "openai-1.57.2-py3-none-any.whl", hash = "sha256:f7326283c156fdee875746e7e54d36959fb198eadc683952ee05e3302fbd638d"}, - {file = "openai-1.57.2.tar.gz", hash = "sha256:5f49fd0f38e9f2131cda7deb45dafdd1aee4f52a637e190ce0ecf40147ce8cee"}, + {file = "openai-1.58.1-py3-none-any.whl", hash = "sha256:e2910b1170a6b7f88ef491ac3a42c387f08bd3db533411f7ee391d166571d63c"}, + {file = "openai-1.58.1.tar.gz", hash = "sha256:f5a035fd01e141fc743f4b0e02c41ca49be8fab0866d3b67f5f29b4f4d3c0973"}, ] [package.dependencies] @@ -2648,6 +2648,7 @@ typing-extensions = ">=4.11,<5" [package.extras] datalib = ["numpy (>=1)", "pandas (>=1.2.3)", "pandas-stubs (>=1.1.0.11)"] +realtime = ["websockets (>=13,<15)"] [[package]] name = "packaging" @@ -3363,20 +3364,20 @@ dev = ["argcomplete", "attrs (>=19.2)", "hypothesis (>=3.56)", "mock", "pygments [[package]] name = "pytest-asyncio" -version = "0.24.0" +version = "0.25.0" description = "Pytest support for asyncio" optional = false -python-versions = ">=3.8" +python-versions = ">=3.9" files = [ - {file = "pytest_asyncio-0.24.0-py3-none-any.whl", hash = "sha256:a811296ed596b69bf0b6f3dc40f83bcaf341b155a269052d82efa2b25ac7037b"}, - {file = "pytest_asyncio-0.24.0.tar.gz", hash = "sha256:d081d828e576d85f875399194281e92bf8a68d60d72d1a2faf2feddb6c46b276"}, + {file = "pytest_asyncio-0.25.0-py3-none-any.whl", hash = "sha256:db5432d18eac6b7e28b46dcd9b69921b55c3b1086e85febfe04e70b18d9e81b3"}, + {file = "pytest_asyncio-0.25.0.tar.gz", hash = "sha256:8c0610303c9e0442a5db8604505fc0f545456ba1528824842b37b4a626cbf609"}, ] [package.dependencies] pytest = ">=8.2,<9" [package.extras] -docs = ["sphinx (>=5.3)", "sphinx-rtd-theme (>=1.0)"] +docs = ["sphinx (>=5.3)", "sphinx-rtd-theme (>=1)"] testing = ["coverage (>=6.2)", "hypothesis (>=5.7.1)"] [[package]] @@ -4110,7 +4111,6 @@ files = [ {file = "tiktoken-0.8.0-cp310-cp310-win_amd64.whl", hash = "sha256:d8c2d0e5ba6453a290b86cd65fc51fedf247e1ba170191715b049dac1f628005"}, {file = "tiktoken-0.8.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:d622d8011e6d6f239297efa42a2657043aaed06c4f68833550cac9e9bc723ef1"}, {file = "tiktoken-0.8.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:2efaf6199717b4485031b4d6edb94075e4d79177a172f38dd934d911b588d54a"}, - {file = "tiktoken-0.8.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5637e425ce1fc49cf716d88df3092048359a4b3bbb7da762840426e937ada06d"}, {file = "tiktoken-0.8.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9fb0e352d1dbe15aba082883058b3cce9e48d33101bdaac1eccf66424feb5b47"}, {file = "tiktoken-0.8.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:56edfefe896c8f10aba372ab5706b9e3558e78db39dd497c940b47bf228bc419"}, {file = "tiktoken-0.8.0-cp311-cp311-win_amd64.whl", hash = "sha256:326624128590def898775b722ccc327e90b073714227175ea8febbc920ac0a99"}, @@ -4827,4 +4827,4 @@ type = ["pytest-mypy"] [metadata] lock-version = "2.0" python-versions = ">=3.9,<4.0" -content-hash = "f2d124721dcc72672b852b8ddff07b00f876df27758c64205e5d9ccd87ccd484" +content-hash = "ea0986450834a66f9deca7907e09a071322ab7f608cdea83c5d8af9ad78bd55f" diff --git a/pyproject.toml b/pyproject.toml index f39e3e8fbe085..27aff3fec993a 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -45,7 +45,7 @@ name = "llama-index" packages = [{from = "_llama-index", include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.12.5" +version = "0.12.6" [tool.poetry.dependencies] python = ">=3.9,<4.0" @@ -57,7 +57,7 @@ llama-index-agent-openai = "^0.4.0" llama-index-readers-file = "^0.4.0" llama-index-readers-llama-parse = ">=0.4.0" llama-index-indices-managed-llama-cloud = ">=0.4.0" -llama-index-core = "^0.12.5" +llama-index-core = "^0.12.6" llama-index-multi-modal-llms-openai = "^0.4.0" llama-index-cli = "^0.4.0" nltk = ">3.8.1" # avoids a CVE, temp until next release, should be in llama-index-core