# ChangeLog ## [2024-12-18] ### `llama-index-core` [0.12.7] - fix: add a timeout to langchain callback handler (#17296) - fix: make Document serialization event more backward compatible (#17312) ### `llama-index-embeddings-voyageai` [0.3.4] - Exposing additional keyword arguments for VoyageAI's embedding model (#17315) ### `llama-index-llms-keywordsai` [0.1.0] - Added KeywordsAI LLM (#16860) ### `llama-index-llms-oci-genai` [0.4.0] - Add OCI Generative AI tool calling support (#16888) ### `llama-index-llms-openai` [0.3.11] - support new o1 models (#17307) ### `llama-index-postprocessor-voyageai-rerank` [0.3.1] - VoyageAI Reranker optional API Key (#17310) ### `llama-index-vector-stores-azureaisearch` [0.3.1] - improve async search client handling (#17319) ### `llama-index-vector-stores-azurecosmosmongo` [0.4.0] - CosmosDB insertion timestamp bugfix (#17290) ### `llama-index-vector-stores-azurecosmosnosql` [1.3.0] - CosmosDB insertion timestamp bugfix (#17290) ## [2024-12-17] ### `llama-index-core` [0.12.6] - [bug fix] Ensure that StopEvent gets cleared from Context.\_in_progress["_done"] after a Workflow run (#17300) - fix: add a timeout to langchain callback handler (#17296) - tweak User vs tool in react prompts (#17273) - refact: Refactor Document to be natively multimodal (#17204) - fix: make ImageDocument derive from Document, backward compatible (#17259) - fix: accept already base64-encoded data in ImageBlock (#17244) - fix(metrics): fixed NDCG calculation and updated previous tests (#17236) - fix: remove llama-index-legacy dependency in llama-index-core (#17231) - Refined the default documentation generation for function tools (#17208) ### `llama-index-embeddings-voyageai` [0.3.3] - add support for voyageai >=0.3.0 (#17120) - Introducting VoyageAI's new multimodal embeddings model (#17261) - VoyageAI multimodal embedding, correction (#17284) ### `llama-index-experimental` [0.5.2] - Fixed import errors for experimental JSONalyzeQueryEngine (#17228) ### `llama-index-grapg-stores-neo4j` [0.4.4] - Add cypher corrector and allow graph schema filtering (#17223) - Add timeout config to neo4j graph (#17267) - Add text and embedding type to neo4j enhanced schema (#17289) ### `llama-index-llms-anthropic` [0.6.3] - add content blocks to anthropic (#17274) - Do not send blank content to anthropic (#17278) - Update anthropic type imports for v0.41.0 release (#17299) - Fix Anthropic tokenizer protocol (fix by Devin) (#17201) ### `llama-index-llms-bedrock` [0.3.3] - Add Amazon bedrock guardrails (#17281) ### `llama-index-llms-bedrock-converse` [0.4.2] - Add Amazon bedrock guardrails (#17281) ### `llama-index-llms-gemini` [0.4.1] - Gemini 2.0 support (#17249) ### `llama-index-llms-mistralai` [0.3.1] - add tool call id/name to mistral chat messages (#17280) ### `llama-index-llms-nvidia` [0.3.1] - Adding llama 3.3-70b as function-calling-capable (#17253) ### `llama-index-llms-openai` [0.3.10] - fix openai message dicts for tool calls (#17254) ### `llama-index-llms-text-generation-inference` [0.3.1] - Fix: TGI context window (#17252) ### `llama-index-multi-modal-llms-anthropic` [0.3.1] - handle more response types in anthropic multi modal llms (#17302) ### `llama-index-readers-confluence` [0.3.1] - Support Confluence cookies (#17276) ### `llama-index-vector-stores-milvus` [0.4.0] - Parse "milvus_search_config" out of "vector_store_kwargs" (#17221) - refactor and optimize milvus code (#17229) ### `llama-index-vector-stores-pinecone` [0.4.2] - Handle empty retrieved Pinecone index values (#17242) ### `llama-index-vector-stores-qdrant` [0.4.1] - feat: Add NOT filter condition to MetadataFilter and QdrantVectorStore (#17270) ### `llama-index-vector-stores-weaviate` [1.3.0] - Add async support to weaviate vector store integration (#17220) ## [2024-12-09] ### `llama-index-core` [0.12.5] - Refined the default description generation for function tools (#17208) ### `llama-index-multi-modal-llms-azure-openai` [0.3.2] - fix: relax pin on openai llm dependency (#17210) ### `llama-index-postprocessor-pinecone-native-rerank` [0.1.0] - feat: integration on pinecone hosted rerankers (#17192) ### `llama-index-tools-scrapegraph` [0.1.0] - Add Scrapegraph tool integration (#17238) ### `llama-index-vector-stores-postgres` [0.3.3] - Update pgvector dependency to version 0.3.6 (#17195) ## [2024-12-08] ### `llama-index-core` [0.12.4] - Fix sync and async structured streaming (#17194) - unpin pydantic to allow 2.8 or greater (#17193) - Update core structured predict streaming, add ollama structured predict (#17188) - bump tenacity dependency in llama-index-core (#17178) ### `llama-index-indices-managed-vectara` [0.3.1] - Add Verbose to Vectara `as_query_engine` (#17176) ### `llama-index-llms-ollama` [0.5.0] - Update core structured predict streaming, add ollama structured predict (#17188) ### `llama-index-llms-perplexity` [0.3.2] - Fix message format for perplexity (#17182) ### `llama-index-readers-web` [0.3.1] - Add possibility to use URI as doc id in WholeSiteReader (#17187) ### `llama-index-vector-stores-chroma` [0.4.1] - BUG FIX: llama-index-vectorstore-chromadb to work with chromadb v0.5.17 (#17184) ## [2024-12-06] ### `llama-index-core` [0.12.3] - cover SimpleDirectoryReader with unit tests (#17156) - docs: rewrite openai image reasoning example without multimodal LLM (#17148) - fix(metrics): fixed NDCG calculation and added comprehensive test cases (#17126) - feat: improve ImageBlock (#17111) - Remove forgotten print in ChatMemoryBuffer (#17114) - [FIX] Move JSONalyzeQueryEngine to experimental (#17110) ### `llama-index-embeddings-clip` [0.3.1] - Unrestrict clip models to use (#17162) ### `llama-index-embeddings-openai` [0.3.1] - fix/openai-embbeding-retry (#17072) ### `llama-index-embeddings-text-embeddings-inference` [0.3.1] - proper auth token in TEI (#17158) ### `llama-index-indices-managed-llama-cloud` [0.6.3] - chore: fix httpx_client typo in LlamaCloudRetriever (#17101) - fix: wrong project id variable in LlamaCloudRetriever (#17086) ### `llama-index-llms-bedrock-converse` [0.4.1] - Adding AWS Nova models to Bedrock Converse (#17139) ### `llama-index-llms-ollama` [0.4.2] - Ollama LLM: Added TypeError exception to `_get_response_token_counts` (#17150) ### `llama-index-llms-sambanovasystems` [0.4.3] - changes in openai identification in url (#17161) ### `llama-index-memory-mem0` [0.2.1] - Fix mem0 version check (#17159) ### `llama-index-multi-modal-llms-openai` [0.4.0] - fix: make OpenAIMultiModal work with new ChatMessage (#17138) ### `llama-index-postprocessor-bedrock-rerank` [0.3.0] - Add AWS Bedrock Reranker (#17134) ### `llama-index-readers-file` [0.4.1] - update doc id for unstructured reader (#17160) ### `llama-index-retrievers-duckdb-retriever` [0.4.0] - fix: use prepared statement in DuckDBRetriever (#17092) ### `llama-index-vector-stores-postgres` [0.3.2] - Create tables for pgvector regardless of schema status (#17100) ### `llama-index-vector-stores-weaviate` [1.2.4] - make alpha not none in weaviate (#17163) - Make Weaviate Vector Store integration work with complex properties (#17129) - Add support for `IS_EMPTY` metadata filters to Weaviate Vector Store integration (#17128) - Make Weaviate Vector Store integration support nested metadata filtering (#17107) ## [2024-11-26] ### `llama-index-core` [0.12.2] - improve traceback logging for workflows (#17040) - Initial version of checkpointing for Workflows (#17006) - Fix base component tests (#17062) - mark code splitter tests as optional (#17060) - update contributing guide and dev deps (#17051) - fix Handling of WorkflowDone exception (#17047) ### `llama-index-embeddings-vertex` [0.3.1] - remove pyarrow from vertex deps (#16997) ### `llama-index-llms-ibm` [0.3.1] - IBM watsonx.ai acomplete & achat integration (#17034) ### `llama-index-llms-openai` [0.3.2] - fixes tool_choice in certain scenarios for openai (#17058) ### `llama-index-llms-perplexity` [0.3.1] - Fix: update perplexity models, add unit tests and minor fixes (#17045) ### `llama-index-llms-vertex` [0.4.1] - remove pyarrow from vertex deps (#16997) ### `llama-index-multi-modal-llms-huggingface` [0.2.1] - Add stream_chat and conditionally set AutoModelClass to MllamaForConditionalGeneration (#17031) ### `llama-index-readers-box` [0.3.1] - Fix box api - mypy breakage (#17061) ### `llama-index-vector-stores-deeplake` [0.3.2] - deeplake v3 backward compatibility (#17057) ### `llama-index-vector-stores-mongodb` [0.6.0] - MongoDB Atlas: Adds search index commands and tests/examples of metadata filters (#15265) ### `llama-index-vector-stores-postgres` [0.3.1] - check if schema exists before doing rest of init during pgvector table creation (#17063) ## [2024-11-20] ### `llama-index-core` [0.12.1] - Pin pydantic to `<2.10` temporarily - feat[react-multimodal]: add gemini support for images in MultimodalReActAgentWorker (#16992) ### `llama-index-indices-managed-llama-cloud` [0.6.2] - api_key typo in llama-cloud-index (#17001) ### `llama-index-llms-openai` [0.3.1] - Add support for latest gpt-4o model (#17015) ### `llama-index-llms-zhipuai` [0.2.1] - zhipuai add stop param to api (#16996) ### `llama-index-multi-modal-llms-mistralai` [0.3.1] - Add support for mistral latest models (#16991) ### `llama-index-multi-modal-llms-replicate` [0.3.1] - Fix: Correct typo in replicate multi modal package (#16998) ### `llama-index-vector-stores-pinecone` [0.4.1] - add get_nodes to pinecone (#17007) ## [2024-11-17] **NOTE:** Updating to v0.12.0 will require bumping every other `llama-index-*` package! Every package has had a version bump. Only notable changes are below. ### `llama-index-core` [0.12.0] - Dropped python3.8 support, Unpinned numpy (#16973) - Kg/dynamic pg triplet retrieval limit (#16928) ### `llama-index-indices-managed-llama-cloud` [0.6.1] - Add ID support for LlamaCloudIndex & update from_documents logic, modernize apis (#16927) - allow skipping waiting for ingestion when uploading file (#16934) - add support for files endpoints (#16933) ### `llama-index-indices-managed-vectara` [0.3.0] - Add Custom Prompt Parameter (#16976) ### `llama-index-llms-bedrock` [0.3.0] - minor fix for messages/completion to prompt (#15729) ### `llama-index-llms-bedrock-converse` [0.4.0] - Fix async streaming with bedrock converse (#16942) ### `llama-index-multi-modal-llms-nvidia` [0.2.0] - add vlm support (#16751) ### `llama-index-readers-confluence` [0.3.0] - Permit passing params to Confluence client (#16961) ### `llama-index-readers-github` [0.5.0] - Add base URL extraction method to GithubRepositoryReader (#16926) ### `llama-index-vector-stores-weaviate` [1.2.0] - Allow passing in Weaviate vector store kwargs (#16954) ## [2024-11-11] ### `llama-index-core` [0.11.23] - Fix workflow timeout when streaming and exception happens (#16852) - New function for core.multi_modal_llms.generic_utils (#16896) - Added docstrings and unit tests for core.multimodal (#16872) - actually setting num_workers = num_cpus for ingestion pipeline (#16803) ### `llama-index-embeddings-nvidia` [0.2.5] - skip model validation for nvdev embedding models (#16883) ### `llama-index-embeddings-modelscope` [0.3.0] - Add ModelScope embedding support (#16873) ### `llama-index-graph-stores-kuzu` [0.4.0] - BREAKING: Update relationship table label nomenclature, existing graphs will need to be recreated (#16886) ### `llama-index-llms-anthropic` [0.4.0] - use new anthropic token counting api (#16909) ### `llama-index-llms-bedrock-converse` [0.3.8] - Add inference profile support to Bedrock Converse and add Llama 3.1/3.2 (#16827) ### `llama-index-llms-nvidia` [0.2.7] - skip model validation for nvdev llm models (#16882) ### `llama-index-llms-modelscope` [0.3.0] - Fix modelscope 'pipeline' referenced before assignment (#16864) ### `llama-index-llms-ollama` [0.3.6] - fix ignoring tool calls in additional kwargs of messages (#16764) ### `llama-index-llms-siliconflow` [0.1.0] - add siliconflow llm class (#16861) ### `llama-index-postprocessor-tei-rerank` [0.2.1] - fix top-n parameter in TEI reranker (#16884) ### `llama-index-readers-gitbook` [0.2.0] - add gitbook reader (#16862) ### `llama-index-readers-google` [0.4.3] - feat: add relative file path google drive (#16907) ### `llama-index-readers-file` [0.3.0] - Bump versions and update pypdf dependency to 5.1.0 (#16905) ### `llama-index-vector-stores-azureaisearch` [0.2.9] - Add UserAgent header "llamaindex-python" for azure search (#16895) ### `llama-index-vector-stores-clickhouse` [0.3.2] - fix clickhouse init in vector store (#16903) ## [2024-11-05] ### `llama-index-core` [0.11.22] - bring back support for prompt templates in context chat engines (#16821) - Fixed the JSON Format of Generated Sub-Question (double curly brackets) (#16820) - markdown splitter improve metadata (#16789) - fix empty index + generation synthesizer (#16785) ### `llama-index-embeddings-azure-inference` [0.2.4] - Support for api_version and Azure AI model inference service (#16802) ### `llama-index-embeddings-gemini` [0.2.2] - fix await-async-embeddings (#16790) ### `llama-index-embeddings-siliconflow` [0.1.0] - add siliconflow embedding class (#16753) ### `llama-index-indices-managed-vectara` [0.2.4] - Hotfix: Chain Query Configuration (#16818) ### `llama-index-llms-anthropic` [0.3.9] - Add Anthropic Claude Haiku 3.5 to the list of supported Claude models (#16823) ### `llama-index-llms-azure-inference` [0.2.4] - Support for api_version and Azure AI model inference service (#16802) ### `llama-index-llms-bedrock` [0.2.6] - Add Anthropic Claude Haiku 3.5 to the list of supported Claude models for bedrock and bedrock-converse integrations (#16825) ### `llama-index-llms-bedrock-converse` [0.3.7] - Add Anthropic Claude Haiku 3.5 to the list of supported Claude models for bedrock and bedrock-converse integrations (#16825) ### `llama-index-llms-dashscope` [0.2.5] - More tolerant definition of LLMMetadata information (#16830) - Fix abstract method signature error (#16809) ### `llama-index-llms-vllm` [0.3.0] - remove beam search param for latest vllm (#16817) ### `llama-index-postprocessor-colpali-rerank` [0.1.0] - Add ColPali as reranker (#16829) ### `llama-index-postprocessor-siliconflow-rerank` [0.1.0] - add siliconflow rerank class (#16737) ### `llama-index-readers-microsoft-onedrive` [0.2.2] - fix: add required_exts for one drive reader (#16822) ### `llama-index-vector-stores-chroma` [0.3.0] - Support breaking changes to filter syntax in latest chroma (#16806) ### `llama-index-vector-stores-pinecone` [0.3.0] - support sparse embedding models, fix delete for serverless for pinecone (#16819) ## [2024-10-31] ### `llama-index-core` [0.11.21] - Fixed issue with default value set as None for workflow `ctx.get()` (#16756) - fix various issues with react agent streaming (#16755) - add unit test for query pipeline (#16749) - Fix \_merge_ref_doc_kv_pairs duped for-loop (#16739) - bugfix: determine if nodes is none when creating index (#16703) - fixes LLMRerank default_parse_choice_select_answer_fn parsing issue (#16736) - fix return type check on workflows (#16724) - Fixing a verbose issue and making sql errors more informative (#16686) ### `llama-index-embeddings-siliconflow` [0.1.0] - add siliconflow embedding class (#16753) ### `llama-index-graph-stores-falkordb` [0.2.4] - Multi-Graph-Supported-FalkorDB (#16482) ### `llama-index-llms-anthropic` [0.3.8] - adding additional claude model name, for vertex AI (#16692) ### `llama-index-llms-bedrock-converse` [0.3.6] - Added mistral large2 model id in bedrock (#16742) - Improve Bedrock Tool Calling (#16723) - add new sonnet3.5 to function calling bedrock converse models (#16702) - update bedrock models (#16698) ### `llama-index-llms-bedrock` [0.2.5] - Added mistral large2 model id in bedrock (#16742) - add new sonnet3.5 to function calling bedrock converse models (#16702) - update bedrock models (#16698) ### `llama-index-llms-cohere` [0.3.2] - Adding support to the new Aya-Expanse models from Cohere (#16733) ### `llama-index-llms-dashscope` [0.2.3] - DashScope llm support async (#16711) ### `llama-index-llms-nvidia` [0.3.4] - add nvidia/llama-3.2-nv-embedqa-1b-v1 to set of supported models (#16694) ### `llama-index-llms-pipeshift` [0.1.0] - Pipeshift llama index integration (#16610) ### `llama-index-memory-mem0` [0.1.0] - add Mem0 as a memory (#16708) ### `llama-index-multi-modal-llms-anthropic` [0.2.4] - Fix anthropic multimodal deps conflict, update models (#16699) ### `llama-index-node-parser-docling` [0.2.0] - feat: update Docling reader & node parser to Docling v2 (#16677) ### `llama-index-postprocessor-nvidia-rerank` [0.3.3] - add nvidia/llama-3.2-nv-rerankqa-1b-v1 to set of supported models (#16695) ### `llama-index-postprocessor-siliconflow-rerank` [0.1.0] - add siliconflow rerank class (#16737) ### `llama-index-readers-docling` [0.2.0] - feat: update Docling reader & node parser to Docling v2 (#16677) ### `llama-index-readers-microsoft-onedrive` [0.2.1] - feat: add permissions to one drive metadata (#16646) ### `llama-index-storage-chat-store-azure` [0.2.4] - Add Managed Identity authentication support for Azure storage components (#16710) - Add missing awaits in azure chat store (#16645) ### `llama-index-storage-docstore-azure` [0.2.1] - Add Managed Identity authentication support for Azure storage components (#16710) ### `llama-index-storage-index-store-azure` [0.3.1] - Add Managed Identity authentication support for Azure storage components (#16710) ### `llama-index-storage-kvstore-azure` [0.2.1] - Add Managed Identity authentication support for Azure storage components (#16710) ### `llama-index-tools-openai-image-generation` [0.3.0] - Makes the tool more compatible with the options, also for the future (#16676) ### `llama-index-tools-vectara-query` [0.1.0] - Add Vectara Query Tool (#16722) ### `llama-index-vector-stores-azureaisearch` [0.2.6] - Allow defining retrievable fields in Azure Vector Store (#16766) - feat: add get_nodes azureai search (#16761) - Added get_nodes() function in AISearch vector store (#16653) - Fix querying for ID in AzureAISearchVectorStore (fixes delete_nodes by node_ids) (#16769) ### `llama-index-vector-stores-hnswlib` [0.2.0] - Fixed issue with persistence, rearranged and added new options to construction of HnswlibVectorStore (#16673) ### `llama-index-vector-stores-opensearch` [0.4.1] - Init OpensearchVectorClient with `os_async_client` (#16767) ### `llama-index-vector-stores-qdrant` [0.3.3] - chore: add embeddings on qdrant get_nodes return (#16760) ### `llama-index-vector-stores-weaviate` [1.1.3] - add default ID if node ID is not provided (#16671) ## [2024-10-24] ### `llama-index-core` [0.11.20] - [actually nothing!] ### `llama-index-embeddings-cohere` [0.3.0] - Add support for cohere multi-modal embeddings (#16667) ### `llama-index-embeddings-litellm` [0.2.2] - support timeout param in litellmembedding (#16532) ### `llama-index-graph-stores-neo4j` [0.3.5] - Make neo4j schema refresh configurable (#16651) - fix: receive warnings from dbms server in neo4j queries (#16598) ### `llama-index-indices-managed-vectara` [0.2.3] - add chain postprocessing for vectara (#16627) ### `llama-index-llms-anthropic` [0.3.7] - update anthropic model names (#16643) ### `llama-index-llms-openai` [0.2.16] - fix: skip processing of choice.delta when it is None (#16636) ### `llama-index-llms-reka` [0.1.0] - Reka llamaindex integration (llm and multi-modal-llm) (#15753) ### `llama-index-multi-modal-llms-reka` [0.1.0] - Reka llamaindex integration (llm and multi-modal-llm) (#15753) ### `llama-index-postprocessor-dashscope-rerank` [0.2.1] - Fix BUG where the api_key parameter is not set when calling DashScopeRerank (#16665) ### `llama-index-readers-microsoft-sharepoint` [0.3.4] - add retry logic to requests in cases where access token expires (#16662) ### `llama-index-storage-docstore-mongodb` [0.2.1] - fix missing dependency error for mongodb docstore (#16654) ### `llama-index-storage-docstore-couchbase` [0.1.0] - Add support for Couchbase for DocStore & IndexStore (#16509) ### `llama-index-storage-kvstore-couchbase` [0.1.0] - Add support for Couchbase for DocStore & IndexStore (#16509) ### `llama-index-storage-index-store-couchbase` [0.1.0] - Add support for Couchbase for IndexStore (#16509) ### `llama-index-storage-kvstore-mongodb` [0.2.1] - add missing dependency for mongodb kvstore (#16632) ### `llama-index-utils-workflow` [0.2.2] - Fix drawing for HITL (#16624) ## [2024-10-18] ### `llama-index-core` [0.11.19] - Raise errors in instrumentation properly when handling asyncio futures (#16603) - fix: pass params to VectorStoreQuery properly in PGRetriever (#16586) - Fix structured predict type hints (#16585) - Add async version of retry_on_exceptions_with_backoff utility (#16374) - Refine CONTRIBUTING.md Documentation (#16548) ### `llama-index-embeddings-gaudi` [0.1.0] - Add embedding integration with Intel Gaudi in llama-index-embeddings-gaudi (#16521) ### `llama-index-embeddings-openvino` [0.4.1] - Add OpenClip support through OpenVINO embedding (#16554) ### `llama-index-graph-stores-neo4j` [0.3.4] - fix: remove warnings from dbms server in neo4j queries (#16598) ### `llama-index-llms-ibm` [0.2.2] - IBM watsonx.ai Chat integration + function calling support (#16589) ### `llama-index-llms-mistralai` [0.2.7] - Add support for mistral latest models (#16571) ### `llama-index-llms-openai` [0.2.15] - Added check for fine-tuned models in function_calling method (#16568) - allow passing in openai clients directly (#16560) ### `llama-index-llms-replicate` [0.3.0] - Use Replicate streaming API (#16597) ### `llama-index-multi-modal-llms-openvino` [0.1.0] - Add OpenVINO multimodal support (#16567) ### `llama-index-multi-modal-llms-zhipuai` [0.1.0] - add zhipuai multi modal llm class (#16551) ### `llama-index-readers-google` [0.4.2] - Use port of redirect uri in credential file to run local server in GoogleDocsReader (#16327) ### `llama-index-storage-chat-store-postgres` [0.1.0] - feat: postgres chat store Integration (#16557) ### `llama-index-storage-chat-store-upstash` [0.1.2] - Fix pydantic errors in upstash chat store (#16559) ### `llama-index-vector-stores-azurecosmosmongo` [0.2.2] - Add DiskANN for Azure Cosmos DB Mongo vector store (#16581) ### `llama-index-vector-stores-hnswlib` [0.1.0] - Hnswlib Vector Store integration (#16443) ### `llama-index-vector-stores-oceanbase` [0.1.0] - Add vector store integration of OceanBase (#16550) ### `llama-index-vector-stores-qdrant` [0.3.2] - Added optional name of text field in Qdrant vector database. (#16576) ## [2024-10-14] ### `llama-index-core` [0.11.18] - Handle Empty Nodes List in PG Retrieval while adding text (#16447) - Improved text2sql parsing (#16445) ### `llama-index-embeddings-litellm` [0.2.2] - support timeout param in litellmembedding (#16532) ### `llama-index-embeddings-zhipuai` [0.1.0] - add zhipuai embedding class (#16505) ### `llama-index-graph-stores-memgraph` [0.1.0] - Add Memgraph Graph Store Integration (#16345) ### `llama-index-llms-anthropic` [0.3.6] - Add anthropic bedrock support (#16478) ### `llama-index-llms-databricks` [0.2.1] - Fix Databricks structured_predict (#16527) ### `llama-index-multi-modal-llms-anthropic` [0.2.3] - Propagate default_headers in AnthropicMultiModal (#16496) ### `llama-index-readers-document360` [0.1.0] - Added Document360Reader. Contributed by the PLACE team. (#16305) ### `llama-index-readers-zyte-serp` [0.1.0] - Add Zyte serp integration (#16417) ### `llama-index-readers-upstage` [0.2.1] - Added UpstageDocumentParseReader (#16099) ### `llama-index-storage-chat-store-azure` [0.2.2] - Fix async methods in azure chat store (#16531) ### `llama-index-tools-weather` [0.1.0] - Fix format temp function (#16487) ### `llama-index-vector-stores-elasticsearch` [0.3.3] - Add Support for Custom Metadata Keyword Suffix in Elasticsearch Integration (#16519) ### `llama-index-vector-stores-nile` [0.1.0] - Add vector store integration for Nile (multi-tenant postgres) (#16437) ### `llama-index-vector-stores-opensearch` [0.4.0] - Use efficient kNN filtering, fix filtering when input value is array of string (#16393) ### `llama-index-vector-stores-oracledb` [0.1.4] - Various fixes for Oracle vector store integration (#16536) ### `llama-index-vector-stores-qdrant` [0.3.1] - Update model name typo in Qdrant utils.py (#16494) ### `llama-index-vector-stores-timescalevector` [0.2.2] - fix timescale vector store class (#16539) ### `llama-index-vector-stores-weaviate` [1.1.2] - Fixed issue for similarity score from Weaviate (#16489) ## [2024-10-08] ### `llama-index-core` [0.11.17] - Fix ChatMessage serialization with janky openai types (#16410) ### `llama-index-embeddings-gemini` [0.2.1] - fix gemini embedding async method (#16369) ### `llama-index-llms-gaudi` [0.1.0] - Add llm integration with Intel Gaudi in llama-index-llms-gaudi (#16308) ### `llama-index-llms-openai` [0.2.12] - don't include tool calls if there was none (#16408) ### `llama-index-multi-modal-llms-huggingface` [0.1.1] - LlamaMultiModal class bug fix (#16413) - MultiModal.HuggingFaceMultiModal: fix errors and README, add stream_complete (#16376) ### `llama-index-node-parser-docling` [0.1.0] - feat: add Docling reader and node parser (#16406) ### `llama-index-readers-docling` [0.1.0] - feat: add Docling reader and node parser (#16406) ### `llama-index-readers-zyte-serp` [0.1.0] - Add Zyte serp integration (#16417) ### `llama-index-vector-stores-azureaisearch` [0.2.2] - fix(bug): fixed bug with ensuring the metadata_mapping value (#16431) ### `llama-index-vector-stores-objectbox` [0.1.0] - Add ObjectBox Vector Store Integration (#16314) ## [2024-10-03] ### `llama-index-core` [0.11.16] - Treat non-dict tool JSON function arguments as empty (instead of unexpected raise) (#16316) - Fixing instrumentation for workflows (#16290) - (workaround) Suppress token detaching exception during workflow tracing (#16364) - Raise warning instead of error when nodes have no content (#16354) - Fix typo in BasePGRetriever causing graph context to not be added (#16357) ### `llama-index-embeddings-vertex-endpoint` [0.1.0] - adding vertex endpoint embedding (#16351) ### `llama-index-llms-fireworks` [0.2.1] - Adding support in FireworksAI for Meta 3.2 Models: 1b-instruct; 3b-instruct; 11b-vision; 90b-vision (#16349) ### `llama-index-multi-modal-llms-openai` [0.2.1] - Refactor OpenAI `update_tool_calls` (#16309) ### `llama-index-vector-stores-milvus` [0.2.7] - Add support for nested MetadataFilters and FilterOperator.IS_EMPTY (#16329) ## [2024-10-02] ### `llama-index-core` [0.11.15] - added `to_dict()`, `from_dict()` and serializers for workflow context (#16250) - Ability to cancel workflow execution with `handler.cancel_run()` (#16320) - (breaking) Refactor `WorkflowHandler.run_step()` so user manually emits Event to start next step in workflow (#16277) ### `llama-index-embeddings-oracleai` [0.1.0] - Oraclevs integration (#16161) ### `llama-index-experimental` [0.4.0] - nudge-ft package and add an example for expanding your dataset (#16269) ### `llama-index-llms-anthropic` [0.3.3] - Add support for prompt caching for Anthropic LLM (#16270) ### `llama-index-llms-gemini` [0.3.6] - Output token usage in raw data for Google Gemini LLMs (#16313) ### `llama-index-llms-openai` [0.2.10] - add 4o mini to azure openai models (#16335) ### `llama-index-llms-vertex` [0.3.7] - Rremoving safety settings from generation config for Vertex AI models (#16337) ### `llama-index-multi-modal-llms-huggingface` [0.1.0] - LlamaIndex Multi_Modal_Llms Integration: Huggingface (#16133) ### `llama-index-readers-minio` [0.2.1] - rm extra print statements & replace create tmpfile function (#16291) ### `llama-index-readers-oracleai` [0.1.0] - Oraclevs integration (#16161) ### `llama-index-readers-web` [0.2.3] - Add Zyte Web Reader (#16197) ### `llama-index-retrievers-bm25` [0.4.0] - bump deps for latest bm25s version (#16339) - Update BM25 retriever to use metadata (#16267) ### `llama-index-storage-chat-store-redis` [0.3.2] - fix check for async client in redis chat store (#16321) ### `llama-index-storage-chat-store-upstash` [0.1.0] - Upstash Storage Chat Store Integration (#16237) ### `llama-index-vector-stores-milvus` [0.2.6] - milvus: always set self.\_collection (#16306) - Fix milvus collection creation with index_config (#16165) ### `llama-index-vector-stores-oracledb` [0.1.0] - Oracledb integration (#16161) ## `llama-index-vector-stores-postgres` [0.2.6] - Support TEXT_MATCH FilterOperator in Postgres Vector Store (#16304) ## [2024-09-26] ### `llama-index-core` [0.11.14] - Enhance insert Method in BaseIndex to Support Customizable Transformations (#16206) - Ensure ChatMemoryBuffer's chat history never begins with a TOOL message (#16214) - safe prompt helper string formatting (#16219) - [Feature Request] Support max concurrent workflow_instance.run() executions (#16215) - Workflows + Human In The Loop Dedicated Support (#16220) ### `llama-index-graph-stores-neptune` [0.2.2] - fix NoneType object error when passing in provided client (#16174) ### `llama-index-llms-ollama` [0.3.3] - fix ollama chat missing `keep_alive` (#16182) ### `llama-index-llms-vertex` [0.3.6] - Fix vertex init function (#16216) ### `llama-index-multi-modal-llms-mistral` [0.1.0] - Add support for Mistral Multi modal LLM (#16191) ### `llama-index-readers-jira` [0.3.0] - Add pagination support for Jira Reader (#16226) ### `llama-index-vector-stores-azurecosmosmongo` [0.2.1] - Azure Cosmos DB Filtered Vector Search (#16175) ### `llama-index-vector-stores-azurecosmosnosql` [1.1.0] - Azure Cosmos DB Filtered Vector Search (#16175) ### `llama-index-vector-stores-deeplake` [0.2.1] - Add missing JWT dependency (#16236) ## [2024-09-24] ### `llama-index-core` [0.11.13] - add option for string node representation during retireval in property graphs (#16100) - improve markdown element node parser and structured prediction reliability (#16172) ### `llama-index-graph-stores-neptune` [0.2.1] - Fixed issue where Neptune was adding additional labels (#16137) ### `llama-index-llms-vertext` [0.3.5] - Pass safety_settings to send_message methods to fix settings not being sent to API (#16153) ### `llama-index-readers-box` [0.2.3] - upgrading box sdk to >= 1.5.0 #16169 ### `llama-index-storage-chat-store-dynamodb` [0.2.0] - Async support for dynamodb (#16139) ### `llama-index-storage-chat-store-redis` [0.3.1] - Async support for redis (#16139) ### `llama-index-vector-stores-astra-db` [0.3.0] - Depend on AstraPy 1.5 and above for AstraDBVectorStore (#16164) ## [2024-09-22] ### `llama-index-core` [0.11.12] - Correct Pydantic warning(s) issed for llm base class (#16141) - globally safe format prompt variables in strings with JSON (#15734) - account for tools in prompt helper and response synthesizers (#16157) ### `llama-index-readers-google` [0.4.1] - feat: add drive link to google drive reader metadata (#16156) ### `llama-index-readers-microsoft-sharepoint` [0.3.2] - Add required_exts option to SharePoint reader (#16152) ### `llama-index-vector-stores-milvus` [0.2.4] - Support user-defined schema in MilvusVectorStore (#16151) ## [2024-09-20] ### `llama-index-core` [0.11.11] - Use response synthesizer in context chat engines (#16017) - Async chat memory operation (#16127) - Sql query add option for markdown response (#16103) - Add support for Path for SimpleDirectoryReader (#16108) - Update chat message class for multi-modal (#15969) - fix: `handler.stream_events()` doesn't yield StopEvent (#16115) - pass `hybrid_top_k` in vector retriever (#16105) ### `llama-index-embeddings-elasticsearch` [0.2.1] - fix elasticsearch embedding async function (#16083) ### `llama-index-embeddings-jinaai` [0.3.1] - feat: update JinaEmbedding for v3 release (#15971) ### `llama-index-experimental` [0.3.3] - Enhance Pandas Query Engine Output Processor (#16052) ### `llama-index-indices-managed-vertexai` [0.1.1] - fix incorrect parameters in VertexAIIndex client (#16080) ### `llama-index-node-parser-topic` [0.1.0] - Add TopicNodeParser based on MedGraphRAG paper (#16131) ### `llama-index-multi-modal-llms-ollama` [0.3.2] - Implement async for multi modal ollama (#16091) ### `llama-index-postprocessor-cohere-rerank` [0.2.1] - feat: add configurable base_url field in rerank (#16050) ### `llama-index-readers-file` [0.2.2] - fix bug missing import for bytesio (#16096) ### `llama-index-readers-wordpress` [0.2.2] - Wordpress: Allow control of whether Pages and/or Posts are retrieved (#16128) - Fix Issue 16071: wordpress requires username, password (#16072) ### `llama-index-vector-stores-lancedb` [0.2.1] - fix hybrid search with latest lancedb client (#16057) ### `llama-index-vector-stores-mongodb` [0.3.0] - Fix mongodb hybrid search top-k specs (#16105) ## [2024-09-16] ### `llama-index-core` [0.11.10] - context/result refactor for workflows (#16036) - add sparse embedding abstraction (#16018) - Fix Pydantic models numeric validation (#16008) - Human in loop workflow example (#16011) ### `llama-index-callbacks-opik` [0.1.0] - opik integration (#16007) ### `llama-index-indices-managed-llama-cloud` [0.3.1] - update llamacloud index with image nodes (#15996) ### `llama-index-indices-managed-vectara` [0.2.2] - Hotfix: Fix Citations Text (#16015) ### `llama-index-llms-huggingface` [0.3.4] - Fix: unnecessary warning issue in HuggingFace LLM when tokenizer is provided as argument (#16037) ### `llama-index-readers-dashvector` [0.3.0] - fix: new Data Connector adaption for DashVector (#16028) ### `llama-index-readers-quip` [0.1.0] - add quip reader (#16000) ### `llama-index-sparse-embeddings-fastembed` [0.1.0] - add fastembed sparse embeddings (#16018) ### `llama-index-vector-stores-elasticsearch` [0.2.1] - Fix: get all documents from Elasticsearch KVStore (#16006) ### `llama-index-vector-stores-lancedb` [0.2.3] - temporarily limit lancedb version (#16045) ### `llama-index-vector-stores-postgres` [0.2.5] - Implement `get_nodes()` on PGVectorStore (#16026) ## [2024-09-12] ### `llama-index-core` [0.11.9] - Add callback manager to retriever query engine from args (#15990) - Do not pass system prompt from fn calling runner to fn calling worker (#15986) - fix: Error when parsing react output if tool name contains non-English characters (#15956) ### `llama-index-embeddings-alibabacloud-aisearch` [0.1.0] - Add four alibabacloud-aisearch llama-index integrations: rerank, node_parser, readers, embeddings (#15934) ### `llama-index-experimental` [0.3.1] - Add NUDGE Finetuning (#15954) ### `llama-index-graph-stores-falkordb` [0.2.2] - update falkordb client (#15940) ### `llama-index-llms-openai` [0.2.5] - Add support for o1 openai models (#15979) - force temp to 1.0 for o1 (#15983) ### `llama-index-node-parser-alibabacloud-aisearch` [0.1.0] - Add four alibabacloud-aisearch llama-index integrations: rerank, node_parser, readers, embeddings (#15934) ### `llama-index-postprocessor-alibabacloud-aisearch-rerank` [0.1.0] - Add four alibabacloud-aisearch llama-index integrations: rerank, node_parser, readers, embeddings (#15934) ### `llama-index-readers-alibabacloud-aisearch` [0.1.0] - Add four alibabacloud-aisearch llama-index integrations: rerank, node_parser, readers, embeddings (#15934) ### `llama-index-vector-stores-opensearch` [0.3.0] - Differentiate sync and async calls in OpenSearchVectorClient (#15945) ### `llama-index-vector-stores-postgres` [0.2.4] - fix attribute error in PGVectorStore (#15961) - add support for engine parameters (#15951) ### `llama-index-vector-stores-wordlift` [0.4.5] - Catch nest_asyncio errors (#15975) ## [2024-09-09] ### `llama-index-core` [0.11.8] - feat: Add a retry policy config to workflow steps (#15757) - Add doc id to Langchain format conversions (#15928) ### `llama-index-chat-store-dynamodb` [0.1.0] - Add DynamoDBChatStore (#15917) ### `llama-index-cli` [0.3.1] - Fix RagCLI pydantic error (#15931) ### `llama-index-llms-alibabacloud-aisearch` [0.1.0] - add llama-index llms alibabacloud_aisearch integration (#15850) ### `llama-index-llms-mistralai` [0.2.3] - Make default mistral model support function calling with `large-latest` (#15906) ### `llama-index-llms-vertex` [0.3.4] - Add InternalServerError to retry decorator (#15921) ### `llama-index-postprocessor-rankllm-rerank` [0.3.0] - Update RankLLM with new rerankers (#15892) ### `llama-index-vector-stores-azurecosmosnosql` [1.0.0] - Adding vector store for Azure Cosmos DB NoSql (#14158) ### `llama-index-readers-microsoft-sharepoint` [0.3.1] - Fix error handling in sharepoint reader, fix error with download file (#15868) ### `llama-index-vector-stores-wordlift` [0.4.4] - Adding support for MetadataFilters to WordLift Vector Store (#15905) ### `llama-index-vector-stores-opensearch` [0.2.2] - Opensearch Serverless filtered query support using knn_score script (#15899) ## [2024-09-06] ### `llama-index-core` [0.11.7] - Make SentenceSplitter's secondary_chunking_regex optional (#15882) - force openai structured output (#15706) - fix assert error, add type ignore for streaming agents (#15887) - Fix image document deserialization issue (#15857) ### `llama-index-graph-stores-kuzu` [0.3.2] - Bug fix for KuzuPropertyGraphStore: Allow upserting relations even when chunks are absent (#15889) ### `llama-index-llms-bedrock-converse` [0.3.0] - Removed unused llama-index-llms-anthropic dependency from Bedrock Converse (#15869) ### `llama-index-vector-stores-postgres` [0.2.2] - Fix PGVectorStore with latest pydantic, update pydantic imports (#15886) ### `llama-index-vector-stores-tablestore` [0.1.0] - Add TablestoreVectorStore (#15657) ## [2024-09-05] ### `llama-index-core` [0.11.6] - add llama-deploy docs to docs builds (#15794) - Add oreilly course cookbooks (#15845) ### `llama-index-readers-box` [0.2.1] - Various bug fixes (#15836) ### `llama-index-readers-file` [0.2.1] - Update ImageReader file loading logic (#15848) ### `llama-index-tools-box` [0.2.1] - Various bug fixes (#15836) ### `llama-index-vector-stores-opensearch` [0.2.1] - Refresh Opensearch index after delete operation (#15854) ## [2024-09-04] ### `llama-index-core` [0.11.5] - remove unneeded assert in property graph retriever (#15832) - make simple property graphs serialize again (#15833) - fix json schema for fastapi return types on core components (#15816) ### `llama-index-llms-nvidia` [0.2.2] - NVIDIA llm: Add Completion for starcoder models (#15802) ### `llama-index-llms-ollama` [0.3.1] - add ollama response usage (#15773) ### `llama-index-readers-dashscope` [0.2.1] - fix pydantic v2 validation errors (#15800) ### `llama-index-readers-discord` [0.2.1] - fix: convert Document id from int to string in DiscordReader (#15806) ### `llama-index-vector-stores-mariadb` [0.1.0] - Add MariaDB vector store integration package (#15564) ## [2024-09-02] ### `llama-index-core` [0.11.4] - Add mypy to core (#14883) - Fix incorrect instrumentation fields/types (#15752) - FunctionCallingAgent sources bug + light wrapper to create agent (#15783) - Add text to sql advanced workflow nb (#15775) - fix: remove context after streaming workflow to enable streaming again (#15776) - Fix chat memory persisting and loading methods to use correct JSON format (#15545) - Fix `_example_type` class var being read as private attr with Pydantic V2 (#15758) ### `llama-index-embeddings-litellm` [0.2.1] - add dimensions param to LiteLLMEmbedding, fix a bug that prevents reading vars from env (#15770) ### `llama-index-embeddings-upstage` [0.2.1] - Bugfix upstage embedding when initializing the UpstageEmbedding class (#15767) ### `llama-index-embeddings-sagemaker-endpoint` [0.2.2] - Fix Sagemaker Field required issue (#15778) ### `llama-index-graph-stores-falkordb` [0.2.1] - fix relations upsert with special chars (#15769) ### `llama-index-graph-stores-neo4j` [0.3.1] - Add native vector index support for neo4j lpg and fix vector filters (#15759) ### `llama-index-llms-azure-inference` [0.2.2] - fix: GitHub Models metadata retrieval (#15747) ### `llama-index-llms-bedrock` [0.2.1] - Update `base.py` to fix `self` issues (#15729) ### `llama-index-llms-ollama` [0.3.1] - add ollama response usage (#15773) ### `llama-index-llms-sagemaker-endpoint` [0.2.2] - Fix Sagemaker Field required issue (#15778) ### `llama-index-multi-modal-llms-anthropic` [0.2.1] - Support image type detection without knowing the file name (#15763) ### `llama-index-vector-stores-milvus` [0.2.2] - feat: implement get_nodes for MilvusVectorStore (#15696) ### `llama-index-vector-stores-tencentvectordb` [0.2.1] - fix: tencentvectordb inconsistent attribute name (#15733) ## [2024-08-29] ### `llama-index-core` [0.11.3] - refact: merge Context and Session to simplify the workflows api (#15709) - chore: stop using deprecated `ctx.data` in workflows docs (#15716) - fix: stop streaming workflow events when a step raises (#15714) - Fix llm_chat_callback for multimodal llms (#15700) - chore: Increase unit tests coverage for the workflow package (#15691) - fix SimpleVectorStore.from_persist_dir() behaviour (#15534) ### `llama-index-embeddings-azure-openai` [0.2.5] - fix json serialization for azure embeddings (#15724) ### `llama-index-graph-stores-kuzu` [0.3.0] - Add KuzuPropertyGraphStore (#15678) ### `llama-index-indices-managed-vectara` [0.2.1] - added new User Defined Function reranker (#15546) ### `llama-index-llms-mistralai` [0.2.2] - Fix `random_seed` type in mistral llm (#15701) ### `llama-index-llms-nvidia` [0.2.1] - Add function/tool calling support to nvidia llm (#15359) ### `llama-index-multi-modal-llms-ollama` [0.3.0] - bump ollama client deps for multimodal llm (#15702) ### `llama-index-readers-web` [0.2.1] - Fix: Firecrawl scraping url response (#15720) ### `llama-index-selectors-notdiamond` [0.1.0] - Adding Not Diamond to llama_index (#15703) ### `llama-index-vector-stores-milvus` [0.2.3] - MMR in Milvus vector stores (#15634) - feat: implement get_nodes for MilvusVectorStore (#15696) ## [2024-08-27] ### `llama-index-core` [0.11.2] - fix tool schemas generation for pydantic v2 to handle nested models (#15679) - feat: support default values for nested workflows (#15660) - feat: allow FunctionTool with just an async fn (#15638) - feat: Allow streaming events from steps (#15488) - fix auto-retriever pydantic indent error (#15648) - Implement Router Query Engine example using workflows (#15635) - Add multi step query engine example using workflows (#15438) - start traces for llm-level operations (#15542) - Pass callback_manager to init in CodeSplitter from_defaults (#15585) ### `llama-index-embeddings-xinference` [0.1.0] - Add Xinference Embedding Class (#15579) ### `llama-index-llms-ai21` [0.3.3] - Integrations: AI21 function calling Support (#15622) ### `llama-index-llms-anthropic` [0.3.0] - Added support for anthropic models through GCP Vertex AI (#15661) ### `llama-index-llms-cerebras` [0.1.0] - Implement Cerebras Integration (#15665) ### `llama-index-postprocessor-nvidia-rerank` [0.3.1] - fix downloaded nim endpoint path (#15645) - fix llama-index-postprocessor-nvidia-rerank tests (#15643) ### `llama-index-postprocessor-xinference-rerank` [0.1.0] - add xinference rerank class (#15639) ### `llama-index-vector-stores-alibabacloud-opensearch` [0.2.1] - fix set output fields in AlibabaCloudOpenSearchConfig (#15562) ### `llama-index-vector-stores-azureaisearch` [0.2.1] - Upgrade azure-search-documents to 2024-07-01 GA API and Add Support for Scalar and Binary Quantization in Index Creation (#15650) ### `llama-index-vector-stores-neo4j` [0.2.1] - Neo4j Vector Store: Make Embedding Dimension Check Optional (#15628) ### `llama-inde-vector-stores-milvus` [0.2.1] - Change the default consistency level of Milvus (#15577) ### `llama-index-vector-stores-elasticsearch` [0.3.2] - Fix the ElasticsearchStore key error (#15631) ## [2024-08-23] ### `llama-index-core` [0.11.1] - Replacing client-side docs search with algolia (#15574) - Add docs on extending workflows (#15573) - rename method for nested workflows to add_workflows (#15596) - chore: fix @step usage in the core codebase (#15588) - Modify the validate function in ReflectionWorkflow example notebook to use pydantic model_validate_json method (#15567) - feature: allow concurrent runs of the same workflow instance (#15568) - docs: remove redundant pass_context=True from docs and examples (#15571) ### `llama-index-embeddings-openai` [0.2.3] - fix openai embeddings with pydantic v2 (#15576) ### `llama-index-embeddings-voyageai` [0.2.1] - bump voyage ai embedding client dep (#15595) ### `llama-index-llms-vertex` [0.3.3] - Vertex LLM: Correctly add function calling part to prompt (#15569) - Vertex LLM: Remove manual setting of message content to Function Calling (#15586) ## [2024-08-22] ### `llama-index-core` [0.11.0] - removed deprecated `ServiceContext` -- using this now will print an error with a link to the migration guide - removed deprecated `LLMPredictor` -- using this now will print an error, any existing LLM is a drop-in replacement - made `pandas` an optional dependency ### `Everything Else` - bumped the minor version of every package to account for the new version of `llama-index-core` ## [2024-08-21] ### `llama-index-core` [0.10.68] - remove nested progress bars in base element node parser (#15550) - Adding exhaustive docs for workflows (#15556) - Adding multi-strategy workflow with reflection notebook example (#15445) - remove openai dep from core (#15527) - Improve token counter to handle more response types (#15501) - feat: Allow using step decorator without parentheses (#15540) - feat: workflow services (aka nested workflows) (#15325) - Remove requirement to specify "allowed_query_fields" parameter when using "cypher_validator" in TextToCypher retriever (#15506) ### `llama-index-embeddings-mistralai` [0.1.6] - fix mistral embeddings usage (#15508) ### `llama-index-embeddings-ollama` [0.2.0] - use ollama client for embeddings (#15478) ### `llama-index-embeddings-openvino` [0.2.1] - support static input shape for openvino embedding and reranker (#15521) ### `llama-index-graph-stores-neptune` [0.1.8] - Added code to expose structured schema for Neptune (#15507) ### `llama-index-llms-ai21` [0.3.2] - Integration: AI21 Tools support (#15518) ### `llama-index-llms-bedrock` [0.1.13] - Support token counting for llama-index integration with bedrock (#15491) ### `llama-index-llms-cohere` [0.2.2] - feat: add tool calling support for achat cohere (#15539) ### `llama-index-llms-gigachat` [0.1.0] - Adding gigachat LLM support (#15313) ### `llama-index-llms-openai` [0.1.31] - Fix incorrect type in OpenAI token usage report (#15524) - allow streaming token counts for openai (#15548) ### `llama-index-postprocessor-nvidia-rerank` [0.2.1] - add truncate support (#15490) - Update to 0.2.0, remove old code (#15533) - update default model to nvidia/nv-rerankqa-mistral-4b-v3 (#15543) ### `llama-index-readers-bitbucket` [0.1.4] - Fixing the issues in loading file paths from bitbucket (#15311) ### `llama-index-readers-google` [0.3.1] - enhance google drive reader for improved functionality and usability (#15512) ### `llama-index-readers-remote` [0.1.6] - check and sanitize remote reader urls (#15494) ### `llama-index-vector-stores-qdrant` [0.2.17] - fix: setting IDF modifier in QdrantVectorStore for sparse vectors (#15538) ## [2024-08-18] ### `llama-index-core` [0.10.67] - avoid nltk 3.9 since its broken (#15473) - docs: openllmetry now uses instrumentation (#15443) - Fix LangChainDeprecationWarning (#15397) - Add get/set API to the Context and make it coroutine-safe (#15152) - docs: Cleanlab's cookbook (#15352) - pass kwargs in `async_add()` for vector stores (#15333) - escape json in structured llm (#15404) - docs: Add JSONAlyze Query Engine using workflows cookbook (#15408) ### `llama-index-embeddings-gigachat` [0.1.0] - Add GigaChat embedding (#15278) ### `llama-index-finetuning` [0.1.12] - feat: Integrating Azure OpenAI Finetuning (#15297) ### `llama-index-graph-stores-neptune` [0.1.7] - Exposed NeptuneQueryException and added additional debug information (#15448) - Fixed issue #15414 and added ability to do partial matchfor Neptune Analytics (#15415) - Use backticks to escape label (#15324) ### `llama-index-llms-cohere` [0.2.1] - feat: add tool calling for cohere (#15144) ### `llama-index-packs-corrective-rag` [0.1.2] - Ports over LongRAGPack, Corrective RAG Pack, and Self-Discover Pack to Workflows (#15160) ### `llama-index-packs-longrag` [0.1.1] - Ports over LongRAGPack, Corrective RAG Pack, and Self-Discover Pack to Workflows (#15160) ### `llama-index-packs-self-discover` [0.1.2] - Ports over LongRAGPack, Corrective RAG Pack, and Self-Discover Pack to Workflows (#15160) ### `llama-index-readers-preprocess` [0.1.4] - Enhance PreprocessReader (#15302) ## [2024-08-15] ### `llama-index-core` [0.10.66] - Temporarily revert nltk dependency due to latest version being removed from pypi - Add citation query engine with workflows example (#15372) - bug: Semantic double merging splitter creates chunks larger thank chunk size (#15188) - feat: make `send_event()` in workflows assign the target step (#15259) - make all workflow events accessible like mappings (#15310) ### `llama-index-indices-managed-bge-m3` [0.1.0] - Add BGEM3Index (#15197) ### `llama-index-llms-huggingface` [0.2.7] - update HF's completion_to_prompt (#15354) ### `llama-index-llms-sambanova` [0.1.0] - Wrapper for SambaNova (Sambaverse and SambaStudio) with Llama-index (#15220) ### `llama-index-packs-code-hierarchy` [0.1.7] - Update code_hierarchy.py adding php support (#15145) ### `llama-index-postprocessor-dashscope-rerank` [0.1.4] - fix bug when calling llama-index-postprocessor-dashscope-rerank (#15358) ### `llama-index-readers-box` [0.1.2] - Box refactor: Box File to Llama-Index Document adaptor (#15314) ### `llama-index-readers-gcs` [0.1.8] - GCSReader: Implementing ResourcesReaderMixin and FileSystemReaderMixin (#15365) ### `llama-index-tools-box` [0.1.1] - Box refactor: Box File to Llama-Index Document adaptor (#15314) - Box tools for AI Agents (#15236) ### `llama-index-vector-stores-postgres` [0.1.14] - Check if hnsw index exists (#15287) ## [2024-08-12] ### `llama-index-core` [0.10.65] - chore: bump nltk version (#15277) ### `llama-index-tools-box` [0.1.0] - Box tools for AI Agents (#15236) ### `llama-index-multi-modal-llms-gemini` [0.1.8] - feat: add default_headers to Gemini multi-model (#15296) ### `llama-index-vector-stores-clickhouse` [0.2.0] - chore: stop using ServiceContext from the clickhouse integration (#15300) ### `llama-index-experimental` [0.2.0] - chore: remove ServiceContext usage from experimental package (#15301) ### `llama-index-extractors-marvin` [0.1.4] - fix: MarvinMetadataExtractor functionality and apply async support (#15247) ### `llama-index-utils-workflow` [0.1.1] - chore: bump black version (#15288) - chore: bump nltk version (#15277) ### `llama-index-readers-microsoft-onedrive` [0.1.9] - chore: bump nltk version (#15277) ### `llama-index-embeddings-upstage` [0.1.3] - chore: bump nltk version (#15277) ### `llama-index-embeddings-nvidia` [0.1.5] - chore: bump nltk version (#15277) ### `llama-index-embeddings-litellm` [0.1.1] - chore: bump nltk version (#15277) ### `llama-index-legacy` [0.9.48post1] - chore: bump nltk version (#15277) ### `llama-index-packs-streamlit-chatbot` [0.1.5] - chore: bump nltk version (#15277) ### `llama-index-embeddings-huggingface` [0.2.3] - Feature: added multiprocessing for creating hf embedddings (#15260) ## [2024-08-09] ### `llama-index-core` [0.10.64] - fix: children nodes not carrying metadata from source nodes (#15254) - Workflows: fix the validation error in the decorator (#15252) - fix: strip '''sql (Markdown SQL code snippet) in SQL Retriever (#15235) ### `llama-index-indices-managed-colbert` [0.2.0] - Remove usage of ServiceContext in Colbert integration (#15249) ### `llama-index-vector-stores-milvus` [0.1.23] - feat: Support Milvus collection properties (#15241) ### `llama-index-llms-cleanlab` [0.1.2] - Update models supported by Cleanlab TLM (#15240) ### `llama-index-llms-huggingface` [0.2.6] - add generation prompt to HF chat template (#15239) ### `llama-index-llms-openvino` [0.2.1] - add generation prompt to HF chat template (#15239) ### `llama-index-graph-stores-neo4j` [0.2.14] - Neo4jPropertyGraphStore.get() check for id prop (#15228) ### `llama-index-readers-file` [0.1.33] - Fix fs.open path type (#15226) ## [2024-08-08] ### `llama-index-core` [0.10.63] - add num_workers in workflow decorator to resolve step concurrancy issue (#15210) - Sub Question Query Engine as workflow notebook example (#15209) - Add Llamatrace to workflow notebooks (#15186) - Use node hash instead of node text to match nodes in fusion retriever (#15172) ### `llama-index-embeddings-mistralai` [0.1.5] - handle mistral v1.0 client (#15229) ### `llama-index-extractors-relik` [0.1.1] - Fix relik extractor skip error (#15225) ### `llama-index-finetuning` [0.1.11] - handle mistral v1.0 client (#15229) ### `llama-index-graph-stores-neo4j` [0.2.14] - Add neo4j generic node label (#15191) ### `llama-index-llms-anthropic` [0.1.17] - Allow for images in Anthropic messages (#15227) ### `llama-index-llms-mistralai` [0.1.20] - handle mistral v1.0 client (#15229) ### `llama-index-packs-mixture-of-agents` [0.1.2] - Update Mixture Of Agents llamapack with workflows (#15232) ### `llama-index-tools-slack` [0.1.4] - Fixed slack client ref in ToolSpec (#15202) ## [2024-08-06] ### `llama-index-core` [0.10.62] - feat: Allow None metadata filter by using IS_EMPTY operator (#15167) - fix: use parent source node to node relationships if possible during node parsing (#15182) - Use node hash instead of node text to match nodes in fusion retriever (#15172) ### `llama-index-graph-stores-neo4j` [0.2.13] - Neo4j property graph client side batching (#15179) ### `llama-index-graph-stores-neptune` [0.1.4] - PropertyGraphStore support for Amazon Neptune (#15126) ### `llama-index-llms-gemini` [0.2.0] - feat: add default_headers to Gemini model (#15141) ### `llama-index-llms-openai` [0.1.28] - OpenAI: Support new strict functionality in tool param (#15177) ### `llama-index-vector-stores-opensearch` [0.1.14] - Add support for full MetadataFilters in Opensearch (#15176) ### `llama-index-vector-stores-qdrant` [0.2.15] - feat: Allow None metadata filter by using IS_EMPTY operator (#15167) ### `llama-index-vector-stores-wordlift` [0.3.0] - Add support for fields projection and update sample Notebook (#15140) ## [2024-08-05] ### `llama-index-core` [0.10.61] - Tweaks to workflow docs (document `.send_event()`, expand examples) (#15154) - Create context manager to instrument event and span tags (#15116) - keyval index store index store updated to accept custom collection suffix (#15134) - make workflow context able to collect multiples of the same event (#15153) - Fix `__str__` method for AsyncStreamingResponse (#15131) ### `llama-index-callbacks-literalai` [1.0.0] - feat(integration): add a global handler for Literal AI (#15064) ### `llama-index-extractors-relik` [0.1.0] - Add relik kg constructor (#15123) ### `llama-index-graph-stores-neo4j` [0.1.12] - fix neo4j property graph relation properties when querying (#15068) ### `llama-index-llms-fireworks` [0.1.9] - feat: add default_headers to Fireworks llm (#15150) ### `llama-index-llms-gemini` [0.1.12] - Fix: Gemini 1.0 Pro Vision has been official deprecated, switch default model to gemini-1.5-flash (#15000) ### `llama-index-llms-paieas` [0.1.0] - Add LLM for AlibabaCloud PaiEas (#14983) ### `llama-index-llms-predibase` [0.1.7] - Fix Predibase Integration for HuggingFace-hosted fine-tuned adapters (#15130) ## [2024-02-02] ### `llama-index-core` [0.10.60] - update `StartEvent` usage to allow for dot notation attribute access (#15124) - Add GraphRAGV2 notebook (#15119) - Fixed minor bug in DynamicLLMPathExtractor as well as default output parsers not working (#15085) - update typing for workflow timeouts (#15102) - fix(sql_wrapper): dont mention foreign keys when there is none (#14998) ### `llama-index-graph-stores-neo4j` [0.2.11] - fix neo4j retrieving relation properties (#15111) (#15108) ### `llama-index-llms-vllm` [0.1.9] - Update base.py to use @atexit for cleanup (#15047) ### `llama-index-vector-stores-pinecone` [0.1.9] - bump pinecone client version deps (#15121) ### `llama-index-vector-stores-redis` [0.2.1] - Handle nested MetadataFilters for Redis vector store (#15093) ### `llama-index-vector-stores-wordlift` [0.2.0] - Update WordLift Vector Store to use new client package (#15045) ## [2024-07-31] ### `llama-index-core` [0.10.59] - Introduce `Workflow`s for event-driven orchestration (#15067) - Added feature to context chat engine allowing previous chunks to be inserted into the current context window (#14889) - MLflow Integration added to docs (#14977) - docs(literalai): add Literal AI integration to documentation (#15023) - expand span coverage for query pipeline (#14997) - make re-raising error skip constructor during `asyncio_run()` (#14970) ### `llama-index-embeddings-ollama` [0.1.3] - Add proper async embedding support ### `llama-index-embeddings-textembed` [0.0.1] - add support for textembed embedding (#14968) ### `llama-index-graph-stores-falkordb` [0.1.5] - initial implementation FalkorDBPropertyGraphStore (#14936) ### `llama-index-llms-azure-inference` [0.1.1] - Fix: Azure AI inference integration support for tools (#15044) ### `llama-index-llms-fireworks` [0.1.7] - Updates to Default model for support for function calling (#15046) ### `llama-index-llms-ollama` [0.2.2] - toggle for ollama function calling (#14972) - Add function calling for Ollama (#14948) ### `llama-index-llms-openllm` [0.2.0] - update to OpenLLM 0.6 (#14935) ### `llama-index-packs-longrag` [0.1.0] - Adds a LlamaPack that implements LongRAG (#14916) ### `llama-index-postprocessor-tei-rerank` [0.1.0] - Support for Re-Ranker via Text Embedding Interface (#15063) ### `llama-index-readers-confluence` [0.1.7] - confluence reader sort auth parameters priority (#14905) ### `llama-index-readers-file` [0.1.31] - UnstructuredReader use filename as ID (#14946) ### `llama-index-readers-gitlab` [0.1.0] - Add GitLab reader integration (#15030) ### `llama-index-readers-google` [0.2.11] - Fix issue with average ratings being a float vs an int (#15070) ### `llama-index-retrievers-bm25` [0.2.2] - use proper stemmer in bm25 tokenize (#14965) ### `llama-index-vector-stores-azureaisearch` [0.1.13] - Fix issue with deleting non-existent index (#14949) ### `llama-index-vector-stores-elasticsearch` [0.2.5] - disable embeddings for sparse strategy (#15032) ### `llama-index-vector-stores-kdbai` [0.2.0] - Update default sparse encoder for Hybrid search (#15019) ### `llama-index-vector-stores-milvus` [0.1.22] - Enhance MilvusVectorStore with flexible index management for overwriting (#15058) ### `llama-index-vector-stores-postgres` [0.1.13] - Adds option to construct PGVectorStore with a HNSW index (#15024) ## [2024-07-24] ### `llama-index-core` [0.10.58] - Fix: Token counter expecting response.raw as dict, got ChatCompletionChunk (#14937) - Return proper tool outputs per agent step instead of all (#14885) - Minor bug fixes to async structured streaming (#14925) ### `llama-index-llms-fireworks` [0.1.6] - fireworks ai llama3.1 support (#14914) ### `llama-index-multi-modal-llms-anthropic` [0.1.6] - Add claude 3.5 sonnet to multi modal llms (#14932) ### `llama-index-retrievers-bm25` [0.2.1] - 🐞 fix(integrations): BM25Retriever persist missing arg similarity_top_k (#14933) ### `llama-index-retrievers-vertexai-search` [0.1.0] - Llamaindex retriever for Vertex AI Search (#14913) ### `llama-index-vector-stores-deeplake` [0.1.5] - Improved `deeplake.get_nodes()` performance (#14920) ### `llama-index-vector-stores-elasticsearch` [0.2.3] - Bugfix: Don't pass empty list of embeddings to elasticsearch store when using sparse strategy (#14918) ### `llama-index-vector-stores-lindorm` [0.1.0] - Add vector store integration of lindorm (#14623) ### `llama-index-vector-stores-qdrant` [0.2.14] - feat: allow to limit how many elements retrieve (qdrant) (#14904) ## [2024-07-22] ### `llama-index-core` [0.10.57] - Add an optional parameter similarity_score to VectorContextRetrieve… (#14831) - add property extraction (using property names and optional descriptions) for KGs (#14707) - able to attach output classes to LLMs (#14747) - Add streaming for tool calling / structured extraction (#14759) - fix from removing private variables when copying/pickling (#14860) - Fix empty array being send to vector store in ingestion pipeline (#14859) - optimize ingestion pipeline deduping (#14858) - Add an optional parameter similarity_score to VectorContextRetriever (#14831) ### `llama-index-llms-azure-openai` [0.1.10] - Bugfix: AzureOpenAI may fail with custom azure_ad_token_provider (#14869) ### `llama-index-llms-bedrock-converse` [0.1.5] - feat: ✨ Implement async functionality in BedrockConverse (#14326) ### `llama-index-llms-langchain` [0.3.0] - make some dependencies optional - bump langchain version in integration (#14879) ### `llama-index-llms-ollama` [0.1.6] - Bugfix: ollama streaming response (#14830) ### `llama-index-multi-modal-llms-anthropic` [0.1.5] - align deps (#14850) ### `llama-index-readers-notion` [0.1.10] - update notion reader to handle duplicate pages, database+page ids (#14861) ### `llama-index-vector-stores-milvus` [0.1.21] - Implements delete_nodes() and clear() for Weviate, Opensearch, Milvus, Postgres, and Pinecone Vector Stores (#14800) ### `llama-index-vector-stores-mongodb` [0.1.8] - MongoDB Atlas Vector Search: Enhanced Metadata Filtering (#14856) ### `llama-index-vector-stores-opensearch` [0.1.13] - Implements delete_nodes() and clear() for Weviate, Opensearch, Milvus, Postgres, and Pinecone Vector Stores (#14800) ### `llama-index-vector-stores-pinecone` [0.1.8] - Implements delete_nodes() and clear() for Weviate, Opensearch, Milvus, Postgres, and Pinecone Vector Stores (#14800) ### `llama-index-vector-stores-postgres` [0.1.12] - Implements delete_nodes() and clear() for Weviate, Opensearch, Milvus, Postgres, and Pinecone Vector Stores (#14800) ### `llama-index-vector-stores-weaviate` [1.0.2] - Implements delete_nodes() and clear() for Weviate, Opensearch, Milvus, Postgres, and Pinecone Vector Stores (#14800) ## [2024-07-19] ### `llama-index-core` [0.10.56] - Fixing the issue where the \_apply_node_postprocessors function needs QueryBundle (#14839) - Add Context-Only Response Synthesizer (#14439) - Fix AgentRunner AgentRunStepStartEvent dispatch (#14828) - Improve output format system prompt in ReAct agent (#14814) - Remove double curly replacing from output parser utils (#14735) - Update simple_summarize.py (#14714) ### `llama-index-tools-azure-code-interpreter` [0.2.0] - chore: read AZURE_POOL_MANAGEMENT_ENDPOINT from env vars (#14732) ### `llama-index-llms-azure-inference` [0.1.0] - Azure AI Inference integration (#14672) ### `llama-index-embeddings-azure-inference` [0.1.0] - Azure AI Inference integration (#14672) ### `llama-index-llms-bedrock-converse` [0.1.5] - feat: ✨ Implement async functionality in BedrockConverse (#14326) ### `llama-index-embeddings-yandexgpt` [0.1.5] - Add new integration for YandexGPT Embedding Model (#14313) ### `llama-index-tools-google` [0.1.6] - Update docstring for gmailtoolspec's search_messages tool (#14840) ### `llama-index-postprocessor-nvidia-rerank` [0.1.5] - add support for nvidia/nv-rerankqa-mistral-4b-v3 (#14844) ### `llama-index-embeddings-openai` [0.1.11] - Fix OpenAI Embedding async client bug (#14835) ### `llama-index-embeddings-azure-openai` [0.1.11] - Fix Azure OpenAI LLM and Embedding async client bug (#14833) ### `llama-index-llms-azure-openai` [0.1.9] - Fix Azure OpenAI LLM and Embedding async client bug (#14833) ### `llama-index-multi-modal-llms-openai` [0.1.8] - Add support for gpt-4o-mini (#14820) ### `llama-index-llms-openai` [0.1.26] - Add support for gpt-4o-mini (#14820) ### `llama-index-llms-mistralai` [0.1.18] - Add support for mistralai nemo model (#14819) ### `llama-index-graph-stores-neo4j` [0.2.8] - Fix bug when sanitize is used in neo4j property graph (#14812) - Add filter to get_triples in neo4j (#14811) ### `llama-index-vector-stores-azureaisearch` [0.1.12] - feat: add nested filters for azureaisearch (#14795) ### `llama-index-vector-stores-qdrant` [0.2.13] - feat: Add NOT IN filter for Qdrant vector store (#14791) ### `llama-index-vector-stores-azureaisearch` [0.1.11] - feat: add azureaisearch supported conditions (#14787) - feat: azureaisearch support collection string (#14712) ### `llama-index-tools-weather` [0.1.4] - Fix OpenWeatherMapToolSpec.forecast_tommorrow_at_location (#14745) ### `llama-index-readers-microsoft-sharepoint` [0.2.6] - follow odata.nextLink (#14708) ### `llama-index-vector-stores-qdrant` [0.2.12] - Adds Quantization option to QdrantVectorStore (#14740) ### `llama-index-vector-stores-azureaisearch` [0.1.10] - feat: improve azureai search deleting (#14693) ### `llama-index-agent-openai` [0.2.9] - fix: tools are required for attachments in openai api (#14609) ### `llama-index-readers-box` [0.1.0] - new integration ### `llama-index-embeddings-fastembed` [0.1.6] - fix fastembed python version (#14710) ## [2024-07-11] ### `llama-index-core` [0.10.55] - Various docs updates ### `llama-index-llms-cleanlab` [0.1.1] - Add user configurations for Cleanlab LLM integration (#14676) ### `llama-index-readers-file` [0.1.30] - race between concurrent pptx readers over a single temp filename (#14686) ### `llama-index-tools-exa` [0.1.4] - changes to Exa search tool getting started and example notebook (#14690) ## [2024-07-10] ### `llama-index-core` [0.10.54] - fix: update operator logic for simple vector store filter (#14674) - Add AgentOps integration (#13935) ### `llama-index-embeddings-fastembed` [0.1.5] - chore: update required python version in Qdrant fastembed package (#14677) ### `llama-index-embeddings-huggingface-optimum-intel` [0.1.6] - Bump version llama-index-embeddings-huggingface-optimum-intel (#14670) ### `llama-index-vector-stores-elasticsearch` [0.2.2] - Added support for custom index settings (#14655) ### `llama-index-callbacks-agentops` [0.1.0] - Initial release ### `llama-index-indices-managed-vertexai` [0.0.2] - Fix #14637 Llamaindex managed Vertex AI index needs to be updated. (#14641) ### `llama-index-readers-file` [0.1.29] - fix unstructured import in simple file reader (#14642) ## [2024-07-08] ### `llama-index-core` [0.10.53] - fix handling react usage in `llm.predict_and_call` for llama-agents (#14556) - add the missing arg verbose when `ReActAgent` calling `super().__init__` (#14565) - fix `llama-index-core\llama_index\core\node_parser\text\utils.py` error when use IngestionPipeline parallel (#14560) - deprecate `KnowledgeGraphIndex`, tweak docs (#14575) - Fix `ChatSummaryMemoryBuffer` fails to summary chat history with tool callings (#14563) - Added `DynamicLLMPathExtractor` for Entity Detection With a Schema inferred by LLMs on the fly (#14566) - add cloud document converter (#14608) - fix KnowledgeGraphIndex arg 'kg_triple_extract_template' typo error (#14619) - Fix: Update `UnstructuredElementNodeParser` due to change in unstructured (#14606) - Update ReAct Step to solve issue with incomplete generation (#14587) ### `llama-index-callbacks-promptlayer` [0.1.3] - Conditions logging to promptlayer on successful request (#14632) ### `llama-index-embeddings-databricks` [0.1.0] - Add integration embeddings databricks (#14590) ### `llama-index-llms-ai21` [0.3.1] - Fix MessageRole import from the wrong package in AI21 Package (#14596) ### `llama-index-llms-bedrock` [0.1.12] - handle empty response in Bedrock AnthropicProvider (#14479) - add claude 3.5 sonnet support to Bedrock InvokeAPI (#14594) ### `llama-index-llms-bedrock-converse` [0.1.4] - Fix Bedrock Converse's tool use blocks, when there are multiple consecutive function calls (#14386) ### `llama-index-llms-optimum-intel` [0.1.0] - add optimum intel with ipex backend to llama-index-integration (#14553) ### `llama-index-llms-qianfan` [0.1.0] - add baidu-qianfan llm (#14414) ### `llama-index-llms-text-generation-inference` [0.1.4] - fix: crash LLMMetadata in model name lookup (#14569) - Remove hf embeddings dep from text-embeddings-inference (#14592) ### `llama-index-llms-yi` [0.1.1] - update yi llm context_window (#14578) ### `llama-index-readers-file` [0.1.28] - add fs arg to PandasExcelReader.load_data (#14554) - UnstructuredReader enhancements (#14390) ### `llama-index-readers-web` [0.1.22] - nit: firecrawl fixes for creating documents (#14579) ### `llama-index-retrievers-bm25` [0.2.0] - Update BM25Retriever to use newer (and faster) bm25s library #(14581) ### `llama-index-vector-stores-qdrant` [0.2.11] - refactor: Don't swallow exceptions from Qdrant collection_exists (#14564) - add support for qdrant bm42, setting sparse + dense configs (#14577) ## [2024-07-03] ### `llama-index-core` [0.10.52] - fix file reader path bug on windows (#14537) - follow up with kwargs propagation in colbert index due to change in parent class (#14522) - deprecate query pipeline agent in favor of FnAgentWorker (#14525O) ### `llama-index-callbacks-arize-phoenix` [0.1.6] - support latest version of arize #14526 ### `llama-index-embeddings-litellm` [0.1.0] - Add support for LiteLLM Proxy Server for embeddings (#14523) ### `llama-index-finetuning` [0.1.10] - Adding device choice from sentence_transformers (#14546) ### `llama-index-graph-stores-neo4` [0.2.7] - Fixed ordering of returned nodes on vector queries (#14461) ### `llama-index-llms-bedrock` [0.1.10] - handle empty response in Bedrock AnthropicProvider (#14479) ### `llama-index-llms-bedrock-converse` [0.1.4] - Fix Bedrock Converse's join_two_dicts function when a new string kwarg is added (#14548) ### `llama-index-llms-upstage` [0.1.4] - Add upstage tokenizer and token counting method (#14502) ### `llama-index-readers-azstorage-blob` [0.1.7] - Fix bug with getting object name for blobs (#14547) ### `llama-index-readers-file` [0.1.26] - Pandas excel reader load data fix for appending documents (#14501) ### `llama-index-readers-iceberg` [0.1.0] - Add Iceberg Reader integration to LLamaIndex (#14477) ### `llama-index-readers-notion` [0.1.8] - Added retries (#14488) - add `list_databases` method (#14488) ### `llama-index-readers-slack` [0.1.5] - Enhance SlackReader to fetch Channel IDs from Channel Names/Patterns (#14429) ### `llama-index-readers-web` [0.1.21] - Add API url to firecrawl reader (#14452) ### `llama-index-retrievers-bm25` [0.1.5] - fix score in nodes returned by the BM25 retriever (#14495) ### `llama-index-vector-stores-azureaisearch` [0.1.9] - add async methods to azure ai search (#14496) ### `llama-index-vector-stores-kdbai` [0.1.8] - Kdbai rest compatible (#14511) ### `llama-index-vector-stores-mongodb` [0.1.6] - Adds Hybrid and Full-Text Search to MongoDBAtlasVectorSearch (#14490) ## [2024-06-28] ### `llama-index-core` [0.10.51] - fixed issue with function calling llms and empty tool calls (#14453) - Fix ChatMessage not considered as stringable in query pipeline (#14378) - Update schema llm path extractor to also take a list of valid triples (#14357) - Pass the kwargs on when `build_index_from_nodes` (#14341) ### `llama-index-agent-dashscope` [0.1.0] - Add Alibaba Cloud dashscope agent (#14318) ### `llama-index-graph-stores-neo4j` [0.2.6] - Add MetadataFilters to neo4j_property_graph (#14362) ### `llama-index-llms-nvidia` [0.1.4] - add known context lengths for hosted models (#14436) ### `llama-index-llms-perplexity` [0.1.4] - update available models (#14409) ### `llama-index-llms-predibase` [0.1.6] - Better error handling for invalid API token (#14440) ### `llama-index-llms-yi` [0.1.0] - Integrate Yi model (#14353) ### `llama-index-readers-google` [0.2.9] - Creates Data Loader for Google Chat (#14397) ### `llama-index-readers-s3` [0.1.10] - Invalidate s3fs cache in S3Reader (#14441) ### `llama-index-readers-structured-data` [0.1.0] - Add StructuredDataReader support for xlsx, csv, json and jsonl (#14369) ### `llama-index-tools-jina` [0.1.0] - Integrating a new tool called jina search (#14317) ### `llama-index-vector-stores-astradb` [0.1.8] - Update Astra DB vector store to use modern astrapy library (#14407) ### `llama-index-vector-stores-chromadb` [0.1.10] - Fix the index accessing of ids of chroma get (#14434) ### `llama-index-vector-stores-deeplake` [0.1.4] - Implemented delete_nodes() and clear() in deeplake vector store (#14457) - Implemented get_nodes() in deeplake vector store (#14388) ### `llama-index-vector-stores-elasticsearch` [0.2.1] - Add support for dynamic metadata fields in Elasticsearch index creation (#14431) ### `llama-index-vector-stores-kdbai` [0.1.7] - Kdbai version compatible (#14402) ## [2024-06-24] ### `llama-index-core` [0.10.50] - added dead simple `FnAgentWorker` for custom agents (#14329) - Pass the kwargs on when build_index_from_nodes (#14341) - make async utils a bit more robust to nested async (#14356) ### `llama-index-llms-upstage` [0.1.3] - every llm is a chat model (#14334) ### `llama-index-packs-rag-evaluator` [0.1.5] - added possibility to run local embedding model in RAG evaluation packages (#14352) ## [2024-06-23] ### `llama-index-core` [0.10.49] - Improvements to `llama-cloud` and client dependencies (#14254) ### `llama-index-indices-managed-llama-cloud` [0.2.1] - Improve the interface and client interactions in `LlamaCloudIndex` (#14254) ### `llama-index-llms-bedrock-converse` [0.1.3] - add claude sonnet 3.5 to bedrock converse (#14306) ### `llama-index-llms-upstage` [0.1.2] - set default context size (#14293) - add api_key alias on upstage llm and embeddings (#14233) ### `llama-index-storage-kvstore-azure` [0.1.2] - Optimized inserts (#14321) ### `llama-index-utils-azure` [0.1.1] - azure_table_storage params bug (#14182) ### `llama-index-vector-stores-neo4jvector` [0.1.6] - Add neo4j client method (#14314) ## [2024-06-21] ### `llama-index-core` [0.10.48] - Improve efficiency of average precision (#14260) - add crewai + llamaindex cookbook (#14266) - Add mimetype field to TextNode (#14279) - Improve IBM watsonx.ai docs (#14271) - Updated frontpage of docs, added agents guide, and more (#14089) ### `llama-index-llms-anthropic` [0.1.14] - Add support for claude 3.5 (#14277) ### `llama-index-llms-bedrock-converse` [0.1.4] - Implement Bedrock Converse API for function calling (#14055) ## [2024-06-19] ### `llama-index-core` [0.10.47] - added average precision as a retrieval metric (#14189) - added `.show_jupyter_graph()` method visualizing default simple graph_store in jupyter notebooks (#14104) - corrected the behaviour of nltk file lookup (#14040) - Added helper args to generate_qa_pairs (#14054) - Add new chunking semantic chunking method: double-pass merging (#13629) - enable stepwise execution of query pipelines (#14117) - Replace tenacity upper limit by only rejecting 8.4.0 (#14218) - propagate error_on_no_tool_call kwarg in `llm.predict_and_call()` (#14253) - in query pipeline, avoid casting nodes as strings and use `get_content()` instead (#14242) - Fix NLSQLTableQueryEngine response metadata (#14169) - do not overwrite relations in default simple property graph (#14244) ### `llama-index-embeddings-ipex-llm` [0.1.5] - Enable selecting Intel GPU for ipex embedding integrations (#14214) ### `llama-index-embeddings-mixedbreadai` [0.1.0] - add mixedbread ai integration (#14161) ### `llama-index-graph-stores-neo4j` [0.2.5] - Add default node property to neo4j upsert relations (#14095) ### `llama-index-indices-managed-postgresml` [0.3.0] - Added re-ranking into the PostgresML Managed Index (#14134) ### `llama-index-llms-ai21` [0.3.0] - use async AI21 client for async methods (#14193) ### `llama-index-llms-bedrock-converse` [0.1.2] - Added (fake) async calls to avoid errors (#14241) ### `llama-index-llms-deepinfra` [0.1.3] - Add function calling to deep infra llm (#14127) ### `llama-index-llms-ipex-llm` [0.1.8] - Enable selecting Intel GPU for ipex embedding integrations (#14214) ### `llama-index-llms-oci-genai` [0.1.1] - add command r support oci genai (#14080) ### `llama-index-llms-premai` [0.1.7] - Prem AI Templates Llama Index support (#14105) ### `llama-index-llms-you` [0.1.0] - Integrate You.com conversational APIs (#14207) ### `llama-index-readers-mongodb` [0.1.8] - Add metadata field "collection_name" to SimpleMongoReader (#14245) ### `llama-index-readers-pdf-marker` [0.1.0] - add marker-pdf reader (#14099) ### `llama-index-readers-upstage` [0.1.0] - Added upstage as a reader (#13415) ### `llama-index-postprocessor-mixedbreadai-rerank` [0.1.0] - add mixedbread ai integration (#14161) ### `llama-index-vector-stores-lancedb` [0.1.6] - LanceDB: code cleanup, minor updates (#14077) ### `llama-index-vector-stores-opensearch` [0.1.12] - add option to customize default OpenSearch Client and Engine (#14249) ## [2024-06-17] ### `llama-index-core`[0.10.46] - Fix Pin tenacity and numpy in core (#14203) - Add precision and recall metrics (#14170) - Enable Function calling and agent runner for Vertex AI (#14088) - Fix for batch_gather (#14162) ### `llama-index-utils-huggingface` [0.1.1] - Remove sentence-transformers dependency from HuggingFace utils package (#14204) ### `llama-index-finetuning` [0.1.8] - Add MistralAI Finetuning API support (#14101) ### `llama-index-llms-mistralai` [0.1.16] - Update MistralAI (#14199) ### `llama-index-llms-bedrock-converse` [0.1.0] - fix: 🐛 Fix Bedrock Converse' pyproject.toml for the PyPI release (#14197) ### `llama-index-utils-azure` [0.1.1] - Use typical include llama_index/ (#14196) - Feature/azure_table_storage (#14182) ### `llama-index-embeddings-nvidia` [0.1.4] - add support for nvidia/nv-embed-v1 (https://huggingface.co/nvidia/NV-Embed-v1) (#14194) ### `llama-index-retrievers-you` [0.1.3] - add news retriever (#13934) ### `llama-index-storage-kvstore-azure` [0.1.1] - Fixes a bug where there is a missing await. (#14177) ### `llama-index-embeddings-nomic` [0.4.0post1] - Restore Nomic Embed einops dependency (#14176) ### `llama-index-retrievers-bm25` [0.1.4] - Changing BM25Retriever \_retrieve to use numpy methods (#14015) ### `llama-index-llms-gemini` [0.1.11] - Add missing @llm_chat_callback() to Gemini.stream_chat (#14166) ### `llama-index-llms-vertex` [0.2.0] - Enable Function calling and agent runner for Vertex AI (#14088) ### `llama-index-vector-stores-opensearch` [0.1.11] - feat: support VectorStoreQueryMode.TEXT_SEARCH on OpenSearch VectorStore (#14153) ## [2024-06-14] ### `llama-index-core` [0.10.45] - Fix parsing sql query.py (#14109) - Implement NDCG metric (#14100) - Fixed System Prompts for Structured Generation (#14026) - Split HuggingFace embeddings in HuggingFace API and TextGenerationInference packages (#14013) - Add PandasExcelReader class for parsing excel files (#13991) - feat: add spans to ingestion pipeline (#14062) ### `llama-index-vector-stores-qdrant` [0.2.10] - Fix Qdrant nodes (#14149) ### `llama-index-readers-mongodb` [0.1.7] - Fixes TypeError: sequence item : expected str instance, int found ### `llama-index-indices-managed-vertexai` [0.0.1] - feat: Add Managed Index for LlamaIndex on Vertex AI for RAG (#13626) ### `llama-index-llms-oci-genai` [0.1.1] - Feature/add command r support oci genai (#14080) ### `llama-index-vector-stores-milvus` [0.1.20] - MilvusVectorStore: always include text_key in output_fields (#14076) ### `llama-index-packs-mixture-of-agents` [0.1.0] - Add Mixture Of Agents paper implementation (#14112) ### `llama-index-llms-text-generation-inference` [0.1.0] - Split HuggingFace embeddings in HuggingFace API and TextGenerationInference packages (#14013) ### `llama-index-llms-huggingface-api` [0.1.0] - Split HuggingFace embeddings in HuggingFace API and TextGenerationInference packages (#14013) ### `llama-index-embeddings-huggingface-api` [0.1.0] - Split HuggingFace embeddings in HuggingFace API and TextGenerationInference packages (#14013) ### `llama-index-utils-huggingface` [0.1.0] - Split HuggingFace embeddings in HuggingFace API and TextGenerationInference packages (#14013) ### `llama-index-llms-watsonx` [0.1.8] - Feat: IBM watsonx.ai llm and embeddings integration (#13600) ### `llama-index-llms-ibm` [0.1.0] - Feat: IBM watsonx.ai llm and embeddings integration (#13600) ### `llama-index-embeddings-ibm` [0.1.0] - Feat: IBM watsonx.ai llm and embeddings integration (#13600) ### `llama-index-vector-stores-milvus` [0.1.19] - Fix to milvus filter enum parsing (#14111) ### `llama-index-llms-anthropic` [0.1.13] - fix anthropic llm calls (#14108) ### `llama-index-storage-index-store-postgres` [0.1.4] - Wrong mongo name was used instead of Postgres (#14107) ### `llama-index-embeddings-bedrock` [0.2.1] - Remove unnecessary excluded from fields in Bedrock embedding (#14085) ### `llama-index-finetuning` [0.1.7] - Feature/added trust remote code (#14102) ### `llama-index-readers-file` [0.1.25] - nit: fix for pandas excel reader (#14086) ### `llama-index-llms-anthropic` [0.1.12] - Update anthropic dependency to 0.26.2 minimum version (#14091) ### `llama-index-llms-llama-cpp` [0.1.4] - Add support for Llama 3 Instruct prompt format (#14072) ### `llama-index-llms-bedrock-converse` [0.1.8] - Implement Bedrock Converse API for function calling (#14055) ### `llama-index-vector-stores-postgres` [0.1.11] - fix/postgres-metadata-in-filter-single-elem (#14035) ### `llama-index-readers-file` [0.1.24] - Add PandasExcelReader class for parsing excel files (#13991) ### `llama-index-embeddings-ipex-llm` [0.1.4] - Update dependency of llama-index-embeddings-ipex-llm ### `llama-index-embeddings-gemini` [0.1.8] - Add api key as field in Gemini Embedding (#14061) ### `llama-index-vector-stores-milvus` [0.1.18] - Expand milvus vector store filter options (#13961) ## [2024-06-10] ### `llama-index-core` [0.10.44] - Add WEBP and GIF to supported image types for SimpleDirectoryReader (#14038) - refactor: add spans to abstractmethods via mixin (#14003) - Adding streaming support for SQLAutoVectorQueryEngine (#13947) - add option to specify embed_model to NLSQLTableQueryEngine (#14006) - add spans for multimodal LLMs (#13966) - change to compact in auto prev next (#13940) - feat: add exception events for streaming errors (#13917) - feat: add spans for tools (#13916) ### `llama-index-embeddings-azure-openai` [0.1.10] - Fix error when using azure_ad without setting the API key (#13970) ### `llama-index-embeddings-jinaai` [0.2.0] - add Jina Embeddings MultiModal (#13861) ### `llama-index-embeddings-nomic` [0.3.0] - Add Nomic multi modal embeddings (#13920) ### `llama-index-graph-stores-neo4j` [0.2.3] - ensure cypher returns list before iterating (#13938) ### `llama-index-llms-ai21` [0.2.0] - Add AI21 Labs Jamba-Instruct Support (#14030) ### `llama-index-llms-deepinfra` [0.1.2] - fix(deepinfrallm): default max_tokens (#13998) ### `llama-index-llms-vllm` [0.1.8] - correct `__del__()` Vllm (#14053) ### `llama-index-packs-zenguard` [0.1.0] - Add ZenGuard llamapack (#13959) ### `llama-index-readers-google` [0.2.7] - fix how class attributes are set in google drive reader (#14022) - Add Google Maps Text Search Reader (#13884) ### `llama-index-readers-jira` [0.1.4] - Jira personal access token with hosted instances (#13890) ### `llama-index-readers-mongodb` [0.1.6] - set document ids when loading (#14000) ### `llama-index-retrievers-duckdb-retriever` [0.1.0] - Add DuckDBRetriever (#13929) ### `llama-index-vector-stores-chroma` [0.1.9] - Add inclusion filter to chromadb (#14010) ### `llama-index-vector-stores-lancedb` [0.1.5] - Fix LanceDBVectorStore `add()` logic (#13993) ### `llama-index-vector-stores-milvus` [0.1.17] - Support all filter operators for Milvus vector store (#13745) ### `llama-index-vector-stores-postgres` [0.1.10] - Broaden SQLAlchemy support in llama-index-vector-stores-postgres to 1.4+ (#13936) ### `llama-index-vector-stores-qdrant` [0.2.9] - Qdrant: Create payload index for `doc_id` (#14001) ## [2024-06-02] ### `llama-index-core` [0.10.43] - use default UUIDs when possible for property graph index vector stores (#13886) - avoid empty or duplicate inserts in property graph index (#13891) - Fix cur depth for `get_rel_map` in simple property graph store (#13888) - (bandaid) disable instrumentation from logging generators (#13901) - Add backwards compatibility to Dispatcher.get_dispatch_event() method (#13895) - Fix: Incorrect naming of acreate_plan in StructuredPlannerAgent (#13879) ### `llama-index-graph-stores-neo4j` [0.2.2] - Handle cases where type is missing (neo4j property graph) (#13875) - Rename `Neo4jPGStore` to `Neo4jPropertyGraphStore` (with backward compat) (#13891) ### `llama-index-llms-openai` [0.1.22] - Improve the retry mechanism of OpenAI (#13878) ### `llama-index-readers-web` [0.1.18] - AsyncWebPageReader: made it actually async; it was exhibiting blocking behavior (#13897) ### `llama-index-vector-stores-opensearch` [0.1.10] - Fix/OpenSearch filter logic (#13804) ## [2024-05-31] ### `llama-index-core` [0.10.42] - Allow proper setting of the vector store in property graph index (#13816) - fix imports in langchain bridge (#13871) ### `llama-index-graph-stores-nebula` [0.2.0] - NebulaGraph support for PropertyGraphStore (#13816) ### `llama-index-llms-langchain` [0.1.5] - fix fireworks imports in langchain llm (#13871) ### `llama-index-llms-openllm` [0.1.5] - feat(openllm): 0.5 sdk integrations update (#13848) ### `llama-index-llms-premai` [0.1.5] - Update SDK compatibility (#13836) ### `llama-index-readers-google` [0.2.6] - Fixed a bug with tokens causing an infinite loop in GoogleDriveReader (#13863) ## [2024-05-30] ### `llama-index-core` [0.10.41] - pass embeddings from index to property graph retriever (#13843) - protect instrumentation event/span handlers from each other (#13823) - add missing events for completion streaming (#13824) - missing callback_manager.on_event_end when there is exception (#13825) ### `llama-index-llms-gemini` [0.1.10] - use `model` kwarg for model name for gemini (#13791) ### `llama-index-llms-mistralai` [0.1.15] - Add mistral code model (#13807) - update mistral codestral with fill in middle endpoint (#13810) ### `llama-index-llms-openllm` [0.1.5] - 0.5 integrations update (#13848) ### `llama-index-llms-vertex` [0.1.8] - Safety setting for Pydantic Error for Vertex Integration (#13817) ### `llama-index-readers-smart-pdf-loader` [0.1.5] - handle path objects in smart pdf reader (#13847) ## [2024-05-28] ### `llama-index-core` [0.10.40] - Added `PropertyGraphIndex` and other supporting abstractions. See the [full guide](https://docs.llamaindex.ai/en/latest/module_guides/indexing/lpg_index_guide/) for more details (#13747) - Updated `AutoPrevNextNodePostprocessor` to allow passing in response mode and LLM (#13771) - fix type handling with return direct (#13776) - Correct the method name to `_aget_retrieved_ids_and_texts` in retrievval evaluator (#13765) - fix: QueryTransformComponent incorrect call `self._query_transform` (#13756) - implement more filters for `SimpleVectorStoreIndex` (#13365) ### `llama-index-embeddings-bedrock` [0.2.0] - Added support for Bedrock Titan Embeddings v2 (#13580) ### `llama-index-embeddings-oci-genai` [0.1.0] - add Oracle Cloud Infrastructure (OCI) Generative AI (#13631) ### `llama-index-embeddings-huggingface` [0.2.1] - Expose "safe_serialization" parameter from AutoModel (#11939) ### `llama-index-graph-stores-neo4j` [0.2.0] - Added `Neo4jPGStore` for property graph support (#13747) ### `llama-index-indices-managed-dashscope` [0.1.1] - Added dashscope managed index (#13378) ### `llama-index-llms-oci-genai` [0.1.0] - add Oracle Cloud Infrastructure (OCI) Generative AI (#13631) ### `llama-index-readers-feishu-wiki` [0.1.1] - fix undefined variable (#13768) ### `llama-index-packs-secgpt` [0.1.0] - SecGPT - LlamaIndex Integration #13127 ### `llama-index-vector-stores-hologres` [0.1.0] - Add Hologres vector db (#13619) ### `llama-index-vector-stores-milvus` [0.1.16] - Remove FlagEmbedding as Milvus's dependency (#13767) Unify the collection construction regardless of the value of enable_sparse (#13773) ### `llama-index-vector-stores-opensearch` [0.1.9] - refactor to put helper methods inside class definition (#13749) ## [2024-05-24] ### `llama-index-core` [0.10.39] - Add VectorMemory and SimpleComposableMemory (#13352) - Improve MarkdownReader to ignore headers in code blocks (#13694) - proper async element node parsers (#13698) - return only the message content in function calling worker (#13677) - nit: fix multimodal query engine to use metadata (#13712) - Add notebook with workaround for lengthy tool descriptions and QueryPlanTool (#13701) ### `llama-index-embeddings-ipex-llm` [0.1.2] - Improve device selection (#13644) ### `llama-index-indices-managed-postgresml` [0.1.3] - Add the PostgresML Managed Index (#13623) ### `llama-index-indices-managed-vectara` [0.1.4] - Added chat engine, streaming, factual consistency score, and more (#13639) ### `llama-index-llms-deepinfra` [0.0.1] - Add Integration for DeepInfra LLM Models (#13652) ### `llama-index-llm-ipex-llm` [0.1.3] - add GPU support for llama-index-llm-ipex-llm (#13691) ### `llama-index-llms-lmstudio` [0.1.0] - lmstudio integration (#13557) ### `llama-index-llms-ollama` [0.1.5] - Use aiter_lines function to iterate over lines in ollama integration (#13699) ### `llama-index-llms-vertex` [0.1.6] - Added safety_settings parameter for gemini (#13568) ### `llama-index-postprocessor-voyageai-rerank` [0.1.3] - VoyageAI reranking bug fix (#13622) ### `llama-index-retrievers-mongodb-atlas-bm25-retriever` [0.1.4] - Add missing return (#13720) ### `llama-index-readers-web` [0.1.17] - Add Scrapfly Web Loader (#13654) ### `llama-index-vector-stores-postgres` [0.1.9] - fix bug with delete and special chars (#13651) ### `llama-index-vector-stores-supabase` [0.1.5] - Try-catch in case the .\_client attribute is not present (#13681) ## [2024-05-21] ### `llama-index-core` [0.10.38] - Enabling streaming in BaseSQLTableQueryEngine (#13599) - Fix nonetype errors in relational node parsers (#13615) - feat(instrumentation): new spans for ALL llms (#13565) - Properly Limit the number of generated questions (#13596) - Pass 'exclude_llm_metadata_keys' and 'exclude_embed_metadata_keys' in element Node Parsers (#13567) - Add batch mode to QueryPipeline (#13203) - Improve SentenceEmbeddingOptimizer to respect Settings.embed_model (#13514) - ReAct output parser robustness changes (#13459) - fix for pydantic tool calling with a single argument (#13522) - Avoid unexpected error when stream chat doesn't yield (#13422) ### `llama-index-embeddings-nomic` [0.2.0] - Implement local Nomic Embed with the inference_mode parameter (#13607) ### `llama-index-embeddings-nvidia` [0.1.3] - Deprecate `mode()` in favor of `__init__(base_url=...)` (#13572) - add snowflake/arctic-embed-l support (#13555) ### `llama-index-embeddings-openai` [0.1.10] - update how retries get triggered for openai (#13608) ### `llama-index-embeddings-upstage` [0.1.0] - Integrations: upstage LLM and Embeddings (#13193) ### `llama-index-llms-gemini` [0.1.8] - feat: add gemini new models to multimodal LLM and regular (#13539) ### `llama-index-llms-groq` [0.1.4] - fix: enable tool use (#13566) ### `llama-index-llms-lmstudio` [0.1.0] - Add support for lmstudio integration (#13557) ### `llama-index-llms-nvidia` [0.1.3] - Deprecate `mode()` in favor of `__init__(base_url=...)` (#13572) ### `llama-index-llms-openai` [0.1.20] - update how retries get triggered for openai (#13608) ### `llama-index-llms-unify` [0.1.0] - Add Unify LLM Support (#12921) ### `llama-index-llms-upstage` [0.1.0] - Integrations: upstage LLM and Embeddings (#13193) ### `llama-index-llms-vertex` [0.1.6] - Adding Support for MedLM Models (#11911) ### `llama_index.postprocessor.dashscope_rerank` [0.1.0] - Add dashscope rerank for postprocessor (#13353) ### `llama-index-postprocessor-nvidia-rerank` [0.1.2] - Deprecate `mode()` in favor of `__init__(base_url=...)` (#13572) ### `llama-index-readers-mongodb` [0.1.5] - SimpleMongoReader should allow optional fields in metadata (#13575) ### `llama-index-readers-papers` [0.1.5] - fix: (ArxivReader) set exclude_hidden to False when reading data from hidden directory (#13578) ### `llama-index-readers-sec-filings` [0.1.5] - fix: sec_filings header when making requests to sec.gov #13548 ### `llama-index-readers-web` [0.1.16] - Added firecrawl search mode (#13560) - Updated Browserbase web reader (#13535) ### `llama-index-tools-cassandra` [0.1.0] - added Cassandra database tool spec for agents (#13423) ### `llama-index-vector-stores-azureaisearch` [0.1.7] - Allow querying AzureAISearch without non-null metadata field (#13531) ### `llama-index-vector-stores-elasticsearch` [0.2.0] - Integrate VectorStore from Elasticsearch client (#13291) ### `llama-index-vector-stores-milvus` [0.1.14] - Fix the filter expression construction of Milvus vector store (#13591) ### `llama-index-vector-stores-supabase` [0.1.4] - Disconnect when deleted (#13611) ### `llama-index-vector-stores-wordlift` [0.1.0] - Added the WordLift Vector Store (#13028) ## [2024-05-14] ### `llama-index-core` [0.10.37] - Add image_documents at call time for `MultiModalLLMCompletionProgram` (#13467) - fix RuntimeError by switching to asyncio from threading (#13486) - Add support for prompt kwarg (#13405) - VectorStore -> BasePydanticVectorStore (#13439) - fix: user_message does not exist bug (#13432) - import missing response type (#13382) - add `CallbackManager` to `MultiModalLLM` (#13400) ### `llama-index-llms-bedrock` [0.1.8] - Remove "Truncate" parameter from Bedrock Cohere invoke model request (#13442) ### `llama-index-readers-web` [0.1.14] - Trafilatura kwargs and progress bar for trafilatura web reader (#13454) ### `llama-index-vector-stores-postgres` [0.1.8] - Fix #9522 - SQLAlchemy warning when using hybrid search (#13476) ### `llama-index-vector-stores-lantern` [0.1.4] - Fix #9522 - SQLAlchemy warning when using hybrid search (#13476) ### `llama-index-callbacks-uptrain` [0.2.0] - update UpTrain Callback Handler to support new Upgratin eval schema (#13479) ### `llama-index-vector-stores-zep` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-vearch` [0.1.1] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-upstash` [0.1.4] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-typesense` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-timescalerevector` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-tencentvectordb` [0.1.4] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-tair` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-singlestoredb` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-rocksetdb` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-neptune` [0.1.1] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-neo4jvector` [0.1.5] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-myscale` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-metal` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-jaguar` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-epsilla` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-dynamodb` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-dashvector` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-chatgpt-plugin` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-baiduvectordb` [0.1.1] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-bagel` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-awsdocdb` [0.1.5] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-awadb` [0.1.3] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-vector-stores-alibabacloud-opensearch` [0.1.1] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-readers-wordlift` [0.1.4] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-readers-guru` [0.1.4] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-readers-pebblo` [0.1.1] - VectorStore -> BasePydanticVectorStore (#13439) ### `llama-index-postprocessor-voyageai-rerank` [0.1.2] - bump rerank versions (#13465) ### `llama-index-postprocessor-sbert-rerank` [0.1.4] - bump rerank versions (#13465) ### `llama-index-postprocessor-rankllm-rerank` [0.1.3] - bump rerank versions (#13465) ### `llama-index-postprocessor-rankgpt-rerank` [0.1.4] - bump rerank versions (#13465) ### `llama-index-postprocessor-openvino-rerank` [0.1.3] - bump rerank versions (#13465) ### `llama-index-postprocessor-nvidia-rerank` [0.1.1] - bump rerank versions (#13465) ### `llama-index-postprocessor-jinaai-rerank` [0.1.3] - bump rerank versions (#13465) ### `llama-index-postprocessor-flag-embedding-rerank` [0.1.3] - bump rerank versions (#13465) ### `llama-index-postprocessor-colbert-rerank` [0.1.2] - bump rerank versions (#13465) ### `llama-index-postprocessor-cohere-rerank` [0.1.6] - bump rerank versions (#13465) ### `llama-index-multi-modal-llms-openai` [0.1.6] - gpt-4o support (#13463) ### `llama-index-llms-openai` [0.1.19] - gpt-4o support (#13463) ### `llama-index-packs-rag-fusion-query-pipeline` [0.1.4] - fix the RAG fusion pipeline (#13413) ### `llama-index-agent-openai` [0.2.5] - fix: update OpenAIAssistantAgent to use attachments (#13341) ### `llama-index-embeddings-deepinfra` [0.1.0] - new embeddings integration (#13323) ### `llama-index-llms-mlx` [0.1.0] - new llm integration (#13231) ### `llama-index-vector-stores-milvus` [0.1.12] - fix: Corrected connection parameters in connections.connect() (#13448) ### `llama-index-vector-stores-azureaisearch` [0.1.6] - fix AzureAiSearchVectorStore metadata f-string (#13435) ### `llama-index-vector-stores-mongodb` [0.1.5] - adds Unit and Integration tests for MongoDBAtlasVectorSearch (#12854) ### `llama-index-llms-huggingface` [0.2.0] - update llama-index-llms-huggingface dependency (#13420) ### `llama-index-vector-store-relyt` [0.1.0] - new vector store integration ### `llama-index-storage-kvstore-redis` [0.1.5] - Implement async methods in RedisKVStore (#12943) ### `llama-index-packs-cohere-citation-chat` [0.1.5] - pin llama-index-llms-cohere dependency (#13417) ### `llama-index-llms-cohere` [0.2.0] - pin cohere dependency (#13417) ### `llama-index-tools-azure-code-interpreter` [0.1.1] - fix indexing issue and runtime error message (#13414) ### `llama-index-postprocessor-cohere-rerank` [0.1.5] - fix Cohere Rerank bug (#13410) ### `llama-index-indices-managed-llama-cloud` [0.1.7] - fix retriever integration (#13409) ### `llama-index-tools-azure-code-interpreter` [0.1.0] - new tool ### `llama-index-readers-google` [0.2.5] - fix missing authorized_user_info check on GoogleDriveReader (#13394) ### `llama-index-storage-kvstore-firestore` [0.2.1] - await Firestore's AsyncDocumentReference (#13386) ### `llama-index-llms-nvidia` [0.1.2] - add dynamic model listing support (#13398) ## [2024-05-09] ### `llama-index-core` [0.10.36] - add start_char_idx and end_char_idx with MarkdownElementParser (#13377) - use handlers from global default (#13368) ### `llama-index-readers-pebblo` [0.1.0] - Initial release (#13128) ### `llama-index-llms-cohere` [0.1.7] - Call Cohere RAG inference with documents argument (#13196) ### `llama-index-vector-scores-kdbai` [0.1.6] - update add method decode utf-8 (#13194) ### `llama-index-vector-stores-alibabacloud-opensearch` [0.1.0] - Initial release (#13286) ### `llama-index-tools-multion` [0.2.0] - update tool to use updated api/sdk (#13373) ### `llama-index-vector-sores-weaviate` [1.0.0] - Update to weaviate client v4 (#13229) ### `llama-index-readers-file` [0.1.22] - fix bug where PDFReader ignores extra_info (#13369) ### `llama-index-llms-azure-openai` [0.1.8] - Add sync httpx client support (#13370) ### `llama-index-llms-openai` [0.1.18] - Add sync httpx client support (#13370) - Add missing openai model token context (#13337) ### `llama-index-readers-github` [0.1.9] - Add fail_on_http_error (#13366) ### `llama-index-vector-stores-pinecone` [0.1.7] - Add attribution tag for pinecone (#13329) ### `llama-index-llms-nvidia` [0.1.1] - set default max_tokens to 1024 (#13371) ### `llama-index-readers-papers` [0.1.5] - Fix hiddent temp directory issue for arxiv reader (#13351) ### `llama-index-embeddings-nvidia` [0.1.1] - fix truncate passing aget_query_embedding and get_text_embedding (#13367) ### `llama-index-llms-anyscare` [0.1.4] - Add llama-3 models (#13336) ## [2024-05-07] ### `llama-index-agent-introspective` [0.1.0] - Add CRITIC and reflection agent integrations (#13108) ### `llama-index-core` [0.10.35] - fix `from_defaults()` erasing summary memory buffer history (#13325) - use existing async event loop instead of `asyncio.run()` in core (#13309) - fix async streaming from query engine in condense question chat engine (#13306) - Handle ValueError in extract_table_summaries in element node parsers (#13318) - Handle llm properly for QASummaryQueryEngineBuilder and RouterQueryEngine (#13281) - expand instrumentation payloads (#13302) - Fix Bug in sql join statement missing schema (#13277) ### `llama-index-embeddings-jinaai` [0.1.5] - add encoding_type parameters in JinaEmbedding class (#13172) - fix encoding type access in JinaEmbeddings (#13315) ### `llama-index-embeddings-nvidia` [0.1.0] - add nvidia nim embeddings support (#13177) ### `llama-index-llms-mistralai` [0.1.12] - Fix async issue when streaming with Mistral AI (#13292) ### `llama-index-llms-nvidia` [0.1.0] - add nvidia nim llm support (#13176) ### `llama-index-postprocessor-nvidia-rerank` [0.1.0] - add nvidia nim rerank support (#13178) ### `llama-index-readers-file` [0.1.21] - Update MarkdownReader to parse text before first header (#13327) ### `llama-index-readers-web` [0.1.13] - feat: Spider Web Loader (#13200) ### `llama-index-vector-stores-vespa` [0.1.0] - Add VectorStore integration for Vespa (#13213) ### `llama-index-vector-stores-vertexaivectorsearch` [0.1.0] - Add support for Vertex AI Vector Search as Vector Store (#13186) ## [2024-05-02] ### `llama-index-core` [0.10.34] - remove error ignoring during chat engine streaming (#13160) - add structured planning agent (#13149) - update base class for planner agent (#13228) - Fix: Error when parse file using SimpleFileNodeParser and file's extension doesn't in FILE_NODE_PARSERS (#13156) - add matching `source_node.node_id` verification to node parsers (#13109) - Retrieval Metrics: Updating HitRate and MRR for Evaluation@K documents retrieved. Also adding RR as a separate metric (#12997) - Add chat summary memory buffer (#13155) ### `llama-index-indices-managed-zilliz` [0.1.3] - ZillizCloudPipelineIndex accepts flexible params to create pipelines (#10134, #10112) ### `llama-index-llms-huggingface` [0.1.7] - Add tool usage support with text-generation-inference integration from Hugging Face (#12471) ### `llama-index-llms-maritalk` [0.2.0] - Add streaming for maritalk (#13207) ### `llama-index-llms-mistral-rs` [0.1.0] - Integrate mistral.rs LLM (#13105) ### `llama-index-llms-mymagic` [0.1.7] - mymagicai api update (#13148) ### `llama-index-llms-nvidia-triton` [0.1.5] - Streaming Support for Nvidia's Triton Integration (#13135) ### `llama-index-llms-ollama` [0.1.3] - added async support to ollama llms (#13150) ### `llama-index-readers-microsoft-sharepoint` [0.2.2] - Exclude access control metadata keys from LLMs and embeddings - SharePoint Reader (#13184) ### `llama-index-readers-web` [0.1.11] - feat: Browserbase Web Reader (#12877) ### `llama-index-readers-youtube-metadata` [0.1.0] - Added YouTube Metadata Reader (#12975) ### `llama-index-storage-kvstore-redis` [0.1.4] - fix redis kvstore key that was in bytes (#13201) ### `llama-index-vector-stores-azureaisearch` [0.1.5] - Respect filter condition for Azure AI Search (#13215) ### `llama-index-vector-stores-chroma` [0.1.7] - small bump for new chroma client version (#13158) ### `llama-index-vector-stores-firestore` [0.1.0] - Adding Firestore Vector Store (#12048) ### `llama-index-vector-stores-kdbai` [0.1.5] - small fix to returned IDs after `add()` (#12515) ### `llama-index-vector-stores-milvus` [0.1.11] - Add hybrid retrieval mode to MilvusVectorStore (#13122) ### `llama-index-vector-stores-postgres` [0.1.7] - parameterize queries in pgvector store (#13199) ## [2024-04-27] ### `llama-index-core` [0.10.33] - add agent_worker.as_agent() (#13061) ### `llama-index-embeddings-bedrock` [0.1.5] - Use Bedrock cohere character limit (#13126) ### `llama-index-tools-google` [0.1.5] - Change default value for attendees to empty list (#13134) ### `llama-index-graph-stores-falkordb` [0.1.4] - Skip index creation error when index already exists (#13085) ### `llama-index-tools-google` [0.1.4] - Fix datetime for google calendar create_event api (#13132) ### `llama-index-llms-anthropic` [0.1.11] - Merge multiple prompts into one (#13131) ### `llama-index-indices-managed-llama-cloud` [0.1.6] - Use MetadataFilters in LlamaCloud Retriever (#13117) ### `llama-index-graph-stores-kuzu` [0.1.3] - Fix kuzu integration .execute() calls (#13100) ### `llama-index-vector-stores-lantern` [0.1.3] - Maintenance update to keep up to date with lantern builds (#13116) ## [2024-04-25] ### `llama-index-core` [0.10.32] - Corrected wrong output type for `OutputKeys.from_keys()` (#13086) - add run_jobs to aws base embedding (#13096) - allow user to customize the keyword extractor prompt template (#13083) - (CondenseQuestionChatEngine) Do not condense the question if there's no conversation history (#13069) - QueryPlanTool: Execute tool calls in subsequent (dependent) nodes in the query plan (#13047) - Fix for fusion retriever sometime return Nonetype query(s) before similarity search (#13112) ### `llama-index-embeddings-ipex-llm` [0.1.1] - Support llama-index-embeddings-ipex-llm for Intel GPUs (#13097) ### `llama-index-packs-raft-dataset` [0.1.4] - Fix bug in raft dataset generator - multiple system prompts (#12751) ### `llama-index-readers-microsoft-sharepoint` [0.2.1] - Add access control related metadata to SharePoint reader (#13067) ### `llama-index-vector-stores-pinecone` [0.1.6] - Nested metadata filter support (#13113) ### `llama-index-vector-stores-qdrant` [0.2.8] - Nested metadata filter support (#13113) ## [2024-04-23] ### `llama-index-core` [0.10.31] - fix async streaming response from query engine (#12953) - enforce uuid in element node parsers (#12951) - add function calling LLM program (#12980) - make the PydanticSingleSelector work with async api (#12964) - fix query pipeline's arun_with_intermediates (#13002) ### `llama-index-agent-coa` [0.1.0] - Add COA Agent integration (#13043) ### `llama-index-agent-lats` [0.1.0] - Official LATs agent integration (#13031) ### `llama-index-agent-llm-compiler` [0.1.0] - Add LLMCompiler Agent Integration (#13044) ### `llama-index-llms-anthropic` [0.1.10] - Add the ability to pass custom headers to Anthropic LLM requests (#12819) ### `llama-index-llms-bedrock` [0.1.7] - Adding claude 3 opus to BedRock integration (#13033) ### `llama-index-llms-fireworks` [0.1.5] - Add new Llama 3 and Mixtral 8x22b model into Llama Index for Fireworks (#12970) ### `llama-index-llms-openai` [0.1.16] - Fix AsyncOpenAI "RuntimeError: Event loop is closed bug" when instances of AsyncOpenAI are rapidly created & destroyed (#12946) - Don't retry on all OpenAI APIStatusError exceptions - just InternalServerError (#12947) ### `llama-index-llms-watsonx` [0.1.7] - Updated IBM watsonx foundation models (#12973) ### `llama-index-packs-code-hierarchy` [0.1.6] - Return the parent node if the query node is not present (#12983) - fixed bug when function is defined twice (#12941) ### `llama-index-program-openai` [0.1.6] - dding support for streaming partial instances of Pydantic output class in OpenAIPydanticProgram (#13021) ### `llama-index-readers-openapi` [0.1.0] - add reader for openapi files (#12998) ### `llama-index-readers-slack` [0.1.4] - Avoid infinite loop when not handled exception is raised (#12963) ### `llama-index-readers-web` [0.1.10] - Improve whole site reader to remove duplicate links (#12977) ### `llama-index-retrievers-bedrock` [0.1.1] - Fix Bedrock KB retriever to use query bundle (#12910) ### `llama-index-vector-stores-awsdocdb` [0.1.0] - Integrating AWS DocumentDB as a vector storage method (#12217) ### `llama-index-vector-stores-databricks` [0.1.2] - Fix databricks vector search metadata (#12999) ### `llama-index-vector-stores-neo4j` [0.1.4] - Neo4j metadata filtering support (#12923) ### `llama-index-vector-stores-pinecone` [0.1.5] - Fix error querying PineconeVectorStore using sparse query mode (#12967) ### `llama-index-vector-stores-qdrant` [0.2.5] - Many fixes for async and checking if collection exists (#12916) ### `llama-index-vector-stores-weaviate` [0.1.5] - Adds the index deletion functionality to the WeviateVectoreStore (#12993) ## [2024-04-17] ### `llama-index-core` [0.10.30] - Add intermediate outputs to QueryPipeline (#12683) - Fix show progress causing results to be out of order (#12897) - add OR filter condition support to simple vector store (#12823) - improved custom agent init (#12824) - fix pipeline load without docstore (#12808) - Use async `_aprocess_actions` in `_arun_step_stream` (#12846) - provide the exception to the StreamChatErrorEvent (#12879) - fix bug in load and search tool spec (#12902) ### `llama-index-embeddings-azure-opena` [0.1.7] - Expose azure_ad_token_provider argument to support token expiration (#12818) ### `llama-index-embeddings-cohere` [0.1.8] - Add httpx_async_client option (#12896) ### `llama-index-embeddings-ipex-llm` [0.1.0] - add ipex-llm embedding integration (#12740) ### `llama-index-embeddings-octoai` [0.1.0] - add octoai embeddings (#12857) ### `llama-index-llms-azure-openai` [0.1.6] - Expose azure_ad_token_provider argument to support token expiration (#12818) ### `llama-index-llms-ipex-llm` [0.1.2] - add support for loading "low-bit format" model to IpexLLM integration (#12785) ### `llama-index-llms-mistralai` [0.1.11] - support `open-mixtral-8x22b` (#12894) ### `llama-index-packs-agents-lats` [0.1.0] - added LATS agent pack (#12735) ### `llama-index-readers-smart-pdf-loader` [0.1.4] - Use passed in metadata for documents (#12844) ### `llama-index-readers-web` [0.1.9] - added Firecrawl Web Loader (#12825) ### `llama-index-vector-stores-milvus` [0.1.10] - use batch insertions into Milvus vector store (#12837) ### `llama-index-vector-stores-vearch` [0.1.0] - add vearch to vector stores (#10972) ## [2024-04-13] ### `llama-index-core` [0.10.29] - **BREAKING** Moved `PandasQueryEngine` and `PandasInstruction` parser to `llama-index-experimental` (#12419) - new install: `pip install -U llama-index-experimental` - new import: `from llama_index.experimental.query_engine import PandasQueryEngine` - Fixed some core dependencies to make python3.12 work nicely (#12762) - update async utils `run_jobs()` to include tqdm description (#12812) - Refactor kvdocstore delete methods (#12681) ### `llama-index-llms-bedrock` [0.1.6] - Support for Mistral Large from Bedrock (#12804) ### `llama-index-llms-openvino` [0.1.0] - Added OpenVino LLMs (#12639) ### `llama-index-llms-predibase` [0.1.4] - Update LlamaIndex-Predibase Integration to latest API (#12736) - Enable choice of either Predibase-hosted or HuggingFace-hosted fine-tuned adapters in LlamaIndex-Predibase integration (#12789) ### `llama-index-output-parsers-guardrails` [0.1.3] - Modernize GuardrailsOutputParser (#12676) ### `llama-index-packs-agents-coa` [0.1.0] - Chain-of-Abstraction Agent Pack (#12757) ### `llama-index-packs-code-hierarchy` [0.1.3] - Fixed issue with chunking multi-byte characters (#12715) ### `llama-index-packs-raft-dataset` [0.1.4] - Fix bug in raft dataset generator - multiple system prompts (#12751) ### `llama-index-postprocessor-openvino-rerank` [0.1.2] - Add openvino rerank support (#12688) ### `llama-index-readers-file` [0.1.18] - convert to Path in docx reader if input path str (#12807) - make pip check work for optional pdf packages (#12758) ### `llama-index-readers-s3` [0.1.7] - wrong doc id when using default s3 endpoint in S3Reader (#12803) ### `llama-index-retrievers-bedrock` [0.1.0] - Add Amazon Bedrock knowledge base integration as retriever (#12737) ### `llama-index-retrievers-mongodb-atlas-bm25-retriever` [0.1.3] - Add mongodb atlas bm25 retriever (#12519) ### `llama-index-storage-chat-store-redis` [0.1.3] - fix message serialization in redis chat store (#12802) ### `llama-index-vector-stores-astra-db` [0.1.6] - Relax dependency version to accept astrapy `1.*` (#12792) ### `llama-index-vector-stores-couchbase` [0.1.0] - Add support for Couchbase as a Vector Store (#12680) ### `llama-index-vector-stores-elasticsearch` [0.1.7] - Fix elasticsearch hybrid rrf window_size (#12695) ### `llama-index-vector-stores-milvus` [0.1.8] - Added support to retrieve metadata fields from milvus (#12626) ### `llama-index-vector-stores-redis` [0.2.0] - Modernize redis vector store, use redisvl (#12386) ### `llama-index-vector-stores-qdrant` [0.2.0] - refactor: Switch default Qdrant sparse encoder (#12512) ## [2024-04-09] ### `llama-index-core` [0.10.28] - Support indented code block fences in markdown node parser (#12393) - Pass in output parser to guideline evaluator (#12646) - Added example of query pipeline + memory (#12654) - Add missing node postprocessor in CondensePlusContextChatEngine async mode (#12663) - Added `return_direct` option to tools /tool metadata (#12587) - Add retry for batch eval runner (#12647) - Thread-safe instrumentation (#12638) - Coroutine-safe instrumentation Spans #12589 - Add in-memory loading for non-default filesystems in PDFReader (#12659) - Remove redundant tokenizer call in sentence splitter (#12655) - Add SynthesizeComponent import to shortcut imports (#12655) - Improved truncation in SimpleSummarize (#12655) - adding err handling in eval_utils default_parser for correctness (#12624) - Add async_postprocess_nodes at RankGPT Postprocessor Nodes (#12620) - Fix MarkdownNodeParser ref_doc_id (#12615) ### `llama-index-embeddings-openvino` [0.1.5] - Added initial support for openvino embeddings (#12643) ### `llama-index-llms-anthropic` [0.1.9] - add anthropic tool calling (#12591) ### `llama-index-llms-ipex-llm` [0.1.1] - add ipex-llm integration (#12322) - add more data types support to ipex-llm llm integration (#12635) ### `llama-index-llms-openllm` [0.1.4] - Proper PrivateAttr usage in OpenLLM (#12655) ### `llama-index-multi-modal-llms-anthropic` [0.1.4] - Bumped anthropic dep version (#12655) ### `llama-index-multi-modal-llms-gemini` [0.1.5] - bump generativeai dep (#12645) ### `llama-index-packs-dense-x-retrieval` [0.1.4] - Add streaming support for DenseXRetrievalPack (#12607) ### `llama-index-readers-mongodb` [0.1.4] - Improve efficiency of MongoDB reader (#12664) ### `llama-index-readers-wikipedia` [0.1.4] - Added multilingual support for the Wikipedia reader (#12616) ### `llama-index-storage-index-store-elasticsearch` [0.1.3] - remove invalid chars from default collection name (#12672) ### `llama-index-vector-stores-milvus` [0.1.8] - Added support to retrieve metadata fields from milvus (#12626) - Bug fix - Similarity metric is always IP for MilvusVectorStore (#12611) ## [2024-04-04] ### `llama-index-agent-openai` [0.2.2] - Update imports for message thread typing (#12437) ### `llama-index-core` [0.10.27] - Fix for pydantic query engine outputs being blank (#12469) - Add span_id attribute to Events (instrumentation) (#12417) - Fix RedisDocstore node retrieval from docs property (#12324) - Add node-postprocessors to retriever_tool (#12415) - FLAREInstructQueryEngine : delegating retriever api if the query engine supports it (#12503) - Make chat message to dict safer (#12526) - fix check in batch eval runner for multi-kwargs (#12563) - Fixes agent_react_multimodal_step.py bug with partial args (#12566) ### `llama-index-embeddings-clip` [0.1.5] - Added support to load clip model from local file path (#12577) ### `llama-index-embeddings-cloudflar-workersai` [0.1.0] - text embedding integration: Cloudflare Workers AI (#12446) ### `llama-index-embeddings-voyageai` [0.1.4] - Fix pydantic issue in class definition (#12469) ### `llama-index-finetuning` [0.1.5] - Small typo fix in QA generation prompt (#12470) ### `llama-index-graph-stores-falkordb` [0.1.3] - Replace redis driver with FalkorDB driver (#12434) ### `llama-index-llms-anthropic` [0.1.8] - Add ability to pass custom HTTP headers to Anthropic client (#12558) ### `llama-index-llms-cohere` [0.1.6] - Add support for Cohere Command R+ model (#12581) ### `llama-index-llms-databricks` [0.1.0] - Integrations with DataBricks LLM API (#12432) ### `llama-index-llms-watsonx` [0.1.6] - Updated Watsonx foundation models (#12493) - Updated base model name on watsonx integration #12491 ### `lama-index-postprocessor-rankllm-rerank` [0.1.2] - Add RankGPT support inside RankLLM (#12475) ### `llama-index-readers-microsoft-sharepoint` [0.1.7] - Use recursive strategy by default for SharePoint (#12557) ### `llama-index-readers-web` [0.1.8] - Readability web page reader fix playwright async api bug (#12520) ### `llama-index-vector-stores-kdbai` [0.1.5] - small `to_list` fix (#12515) ### `llama-index-vector-stores-neptune` [0.1.0] - Add support for Neptune Analytics as a Vector Store (#12423) ### `llama-index-vector-stores-postgres` [0.1.5] - fix(postgres): numeric metadata filters (#12583) ## [2024-03-31] ### `llama-index-core` [0.10.26] - pass proper query bundle in QueryFusionRetriever (#12387) - Update llama_parse_json_element.py to fix error on lists (#12402) - Add node postprocessors to retriever tool (#12415) - Fix bug where user specified llm is not respected in fallback logic in element node parsers(#12403) - log proper LLM response key for async callback manager events (#12421) - Deduplicate the two built-in react system prompts; Also make it read from a Markdown file (#12307) - fix bug in BatchEvalRunner for multi-evaluator eval_kwargs_lists (#12418) - add the callback manager event for vector store index insert_nodes (#12443) - fixes an issue with serializing chat messages into chat stores when they contain pydantic API objects (#12394) - fixes an issue with slow memory.get() operation (caused by multiple calls to get_all()) (#12394) - fixes an issue where an agent+tool message pair is cut from the memory (#12394) - Added `FnNodeMapping` for object index (#12391) - Make object mapping optional / hidden for object index (#12391) - Make object index easier to create from existing vector db (#12391) - When LLM failed to follow the react response template, tell it so #12300 ### `llama-index-embeddings-cohere` [0.1.5] - Bump cohere version to 5.1.1 (#12279) ### `llama-index-embeddings-itrex` [0.1.0] - add Intel Extension for Transformers embedding model (#12410) ### `llama-index-graph-stores-neo4j` [0.1.4] - make neo4j query insensitive (#12337) ### `llama-index-llms-cohere` [0.1.5] - Bump cohere version to 5.1.1 (#12279) ### `llama-index-llms-ipex-llm` [0.1.0] - add ipex-llm integration (#12322) ### `llama-index-llms-litellm` [0.1.4] - Fix litellm ChatMessage role validation error (#12449) ### `llama-index-llms-openai` [0.1.14] - Use `FunctionCallingLLM` base class in OpenAI (#12227) ### `llama-index-packs-self-rag` [0.1.4] - Fix llama-index-core dep (#12374) ### `llama-index-postprocessor-cohere-rerank` [0.1.4] - Bump cohere version to 5.1.1 (#12279) ### `llama-index-postprocessor-rankllm-rerank` [0.1.1] - Added RankLLM rerank (#12296) - RankLLM fixes (#12399) ### `llama-index-readers-papers` [0.1.4] - Fixed bug with path names (#12366) ### `llama-index-vector-stores-analyticdb` [0.1.1] - Add AnalyticDB VectorStore (#12230) ### `llama-index-vector-stores-kdbai` [0.1.4] - Fixed typo in imports/readme (#12370) ### `llama-index-vector-stores-qdrant` [0.1.5] - add `in` filter operator for qdrant (#12376) ## [2024-03-27] ### `llama-index-core` [0.10.25] - Add score to NodeWithScore in KnowledgeGraphQueryEngine (#12326) - Batch eval runner fixes (#12302) ### `llama-index-embeddings-cohere` [0.1.5] - Added support for binary / quantized embeddings (#12321) ### `llama-index-llms-mistralai` [0.1.10] - add support for custom endpoints to MistralAI (#12328) ### `llama-index-storage-kvstore-redis` [0.1.3] - Fix RedisDocstore node retrieval from docs property (#12324) ## [2024-03-26] ### `llama-index-core` [0.10.24] - pretty prints in `LlamaDebugHandler` (#12216) - stricter interpreter constraints on pandas query engine (#12278) - PandasQueryEngine can now execute 'pd.\*' functions (#12240) - delete proper metadata in docstore delete function (#12276) - improved openai agent parsing function hook (#12062) - add raise_on_error flag for SimpleDirectoryReader (#12263) - remove un-caught openai import in core (#12262) - Fix download_llama_dataset and download_llama_pack (#12273) - Implement EvalQueryEngineTool (#11679) - Expand instrumenation Span coverage for AgentRunner (#12249) - Adding concept of function calling agent/llm (mistral supported for now) (#12222, ) ### `llama-index-embeddings-huggingface` [0.2.0] - Use `sentence-transformers` as a backend (#12277) ### `llama-index-postprocessor-voyageai-rerank` [0.1.0] - Added voyageai as a reranker (#12111) ### `llama-index-readers-gcs` [0.1.0] - Added google cloud storage reader (#12259) ### `llama-index-readers-google` [0.2.1] - Support for different drives (#12146) - Remove unnecessary PyDrive dependency from Google Drive Reader (#12257) ### `llama-index-readers-readme` [0.1.0] - added readme.com reader (#12246) ### `llama-index-packs-raft` [0.1.3] - added pack for RAFT (#12275) ## [2024-03-23] ### `llama-index-core` [0.10.23] - Added `(a)predict_and_call()` function to base LLM class + openai + mistralai (#12188) - fixed bug with `wait()` in async agent streaming (#12187) ### `llama-index-embeddings-alephalpha` [0.1.0] - Added alephalpha embeddings (#12149) ### `llama-index-llms-alephalpha` [0.1.0] - Added alephalpha LLM (#12149) ### `llama-index-llms-openai` [0.1.7] - fixed bug with `wait()` in async agent streaming (#12187) ### `llama-index-readers-docugami` [0.1.4] - fixed import errors in docugami reader (#12154) ### `llama-index-readers-file` [0.1.12] - fix PDFReader for remote fs (#12186) ## [2024-03-21] ### `llama-index-core` [0.10.22] - Updated docs backend from sphinx to mkdocs, added ALL api reference, some light re-org, better search (#11301) - Added async loading to `BaseReader` class (although its fake async for now) (#12156) - Fix path implementation for non-local FS in `SimpleDirectoryReader` (#12141) - add args/kwargs to spans, payloads for retrieval events, in instrumentation (#12147) - [react agent] Upon exception, say so, so that Agent can correct itself (#12137) ### `llama-index-embeddings-together` [0.1.3] - Added rate limit handling (#12127) ### `llama-index-graph-stores-neptune` [0.1.3] - Add Amazon Neptune Support as Graph Store (#12097) ### `llama-index-llms-vllm` [0.1.7] - fix VllmServer to work without CUDA-required vllm core (#12003) ### `llama-index-readers-s3` [0.1.4] - Use S3FS in S3Reader (#12061) ### `llama-index-storage-docstore-postgres` [0.1.3] - Added proper kvstore dep (#12157) ### `llama-index-storage-index-store-postgres` [0.1.3] - Added proper kvstore dep (#12157) ### `llama-index-vector-stores-elasticsearch` [0.1.6] - fix unclosed session in es add function #12135 ### `llama-index-vector-stores-kdbai` [0.1.3] - Add support for `KDBAIVectorStore` (#11967) ## [2024-03-20] ### `llama-index-core` [0.10.21] - Lazy init for async elements StreamingAgentChatResponse (#12116) - Fix streaming generators get bug by SynthesisEndEvent (#12092) - CLIP embedding more models (#12063) ### `llama-index-packs-raptor` [0.1.3] - Add `num_workers` to summary module (#) ### `llama-index-readers-telegram` [0.1.5] - Fix datetime fields (#12112) - Add ability to select time period of posts/messages (#12078) ### `llama-index-embeddings-openai` [0.1.7] - Add api version/base api as optional for open ai embedding (#12091) ### `llama-index-networks` [0.2.1] - Add node postprocessing to network retriever (#12027) - Add privacy-safe networks demo (#12027) ### `llama-index-callbacks-langfuse` [0.1.3] - Chore: bumps min version of langfuse dep (#12077) ### `llama-index-embeddings-google` [0.1.4] - Chore: bumps google-generativeai dep (#12085) ### `llama-index-embeddings-gemini` [0.1.5] - Chore: bumps google-generativeai dep (#12085) ### `llama-index-llms-gemini` [0.1.6] - Chore: bumps google-generativeai dep (#12085) ### `llama-index-llms-palm` [0.1.4] - Chore: bumps google-generativeai dep (#12085) ### `llama-index-multi-modal-llms-google` [0.1.4] - Chore: bumps google-generativeai dep (#12085) ### `llama-index-vector-stores-google` [0.1.5] - Chore: bumps google-generativeai dep (#12085) ### `llama-index-storage-kvstore-elasticsearch` [0.1.0] - New integration (#12068) ### `llama-index-readers-google` [0.1.7] - Fix - Google Drive Issue of not loading same name files (#12022) ### `llama-index-vector-stores-upstash` [0.1.3] - Adding Metadata Filtering support for UpstashVectorStore (#12054) ### `llama-index-packs-raptor` [0.1.2] - Fix: prevent RaptorPack infinite recursion (#12008) ### `llama-index-embeddings-huggingface-optimum` [0.1.4] - Fix(OptimumEmbedding): removing token_type_ids causing ONNX validation issues ### `llama-index-llms-anthropic` [0.1.7] - Fix: Anthropic LLM merge consecutive messages with same role (#12013) ### `llama-index-packs-diff-private-simple-dataset` [0.1.0] - DiffPrivacy ICL Pack - OpenAI Completion LLMs (#11881) ### `llama-index-cli` [0.1.11] - Remove llama_hub_url keyword from download_llama_dataset of command (#12038) ## [2024-03-14] ### `llama-index-core` [0.10.20] - New `instrumentation` module for observability (#11831) - Allow passing in LLM for `CitationQueryEngine` (#11914) - Updated keyval docstore to allow changing suffix in addition to namespace (#11873) - Add (some) async streaming support to query_engine #11949 ### `llama-index-embeddings-dashscope` [0.1.3] - Fixed embedding type for query texts (#11901) ### `llama-index-embeddings-premai` [0.1.3] - Support for premai embeddings (#11954) ### `llama-index-networks` [0.2.0] - Added support for network retrievers (#11800) ### `llama-index-llms-anthropic` [0.1.6] - Added support for haiku (#11916) ### `llama-index-llms-mistralai` [0.1.6] - Fixed import error for ChatMessage (#11902) ### `llama-index-llms-openai` [0.1.11] - added gpt-35-turbo-0125 for AZURE_TURBO_MODELS (#11956) - fixed error with nontype in logprobs (#11967) ### `llama-index-llms-premai` [0.1.4] - Support for premai llm (#11954) ### `llama-index-llms-solar` [0.1.3] - Support for solar as an LLM class (#11710) ### `llama-index-llms-vertex` [0.1.5] - Add support for medlm in vertex (#11911) ### `llama-index-readers-goolge` [0.1.6] - added README files and query string for google drive reader (#11724) ### `llama-index-readers-file` [0.1.11] - Updated ImageReader to add `plain_text` option to trigger pytesseract (#11913) ### `llama-index-readers-pathway` [0.1.3] - use pure requests to reduce deps, simplify code (#11924) ### `llama-index-retrievers-pathway` [0.1.3] - use pure requests to reduce deps, simplify code (#11924) ### `llama-index-storage-docstore-mongodb` [0.1.3] - Allow changing suffix for mongodb docstore (#11873) ### `llama-index-vector-stores-databricks` [0.1.1] - Support for databricks vector search as a vector store (#10754) ### `llama-index-vector-stores-opensearch` [0.1.8] - (re)implement proper delete (#11959) ### `llama-index-vector-stores-postgres` [0.1.4] - Fixes for IN filters and OR text search (#11872, #11927) ## [2024-03-12] ### `llama-index-cli` [0.1.9] - Removed chroma as a bundled dep to reduce `llama-index` deps ### `llama-index-core` [0.10.19] - Introduce retries for rate limits in `OpenAI` llm class (#11867) - Added table comments to SQL table schemas in `SQLDatabase` (#11774) - Added `LogProb` type to `ChatResponse` object (#11795) - Introduced `LabelledSimpleDataset` (#11805) - Fixed insert `IndexNode` objects with unserializable objects (#11836) - Fixed stream chat type error when writing response to history in `CondenseQuestionChatEngine` (#11856) - Improve post-processing for json query engine (#11862) ### `llama-index-embeddings-cohere` [0.1.4] - Fixed async kwarg error (#11822) ### `llama-index-embeddings-dashscope` [0.1.2] - Fixed pydantic import (#11765) ### `llama-index-graph-stores-neo4j` [0.1.3] - Properly close connection after verifying connectivity (#11821) ### `llama-index-llms-cohere` [0.1.3] - Add support for new `command-r` model (#11852) ### `llama-index-llms-huggingface` [0.1.4] - Fixed streaming decoding with special tokens (#11807) ### `llama-index-llms-mistralai` [0.1.5] - Added support for latest and open models (#11792) ### `llama-index-tools-finance` [0.1.1] - Fixed small bug when passing in the API get for stock news (#11772) ### `llama-index-vector-stores-chroma` [0.1.6] - Slimmed down chroma deps (#11775) ### `llama-index-vector-stores-lancedb` [0.1.3] - Fixes for deleting (#11825) ### `llama-index-vector-stores-postgres` [0.1.3] - Support for nested metadata filters (#11778) ## [2024-03-07] ### `llama-index-callbacks-deepeval` [0.1.3] - Update import path for callback handler (#11754) ### `llama-index-core` [0.10.18] - Ensure `LoadAndSearchToolSpec` loads document objects (#11733) - Fixed bug for no nodes in `QueryFusionRetriever` (#11759) - Allow using different runtime kwargs for different evaluators in `BatchEvalRunner` (#11727) - Fixed issues with fsspec + `SimpleDirectoryReader` (#11665) - Remove `asyncio.run()` requirement from guideline evaluator (#11719) ### `llama-index-embeddings-voyageai` [0.1.3] - Update voyage embeddings to use proper clients (#11721) ### `llama-index-indices-managed-vectara` [0.1.3] - Fixed issues with vectara query engine in non-summary mode (#11668) ### `llama-index-llms-mymagic` [0.1.5] - Add `return_output` option for json output with query and response (#11761) ### `llama-index-packs-code-hierarchy` [0.1.0] - Added support for a `CodeHiearchyAgentPack` that allows for agentic traversal of a codebase (#10671) ### `llama-index-packs-cohere-citation-chat` [0.1.3] - Added a new llama-pack for citations + chat with cohere (#11697) ### `llama-index-vector-stores-milvus` [0.1.6] - Prevent forced `flush()` on document add (#11734) ### `llama-index-vector-stores-opensearch` [0.1.7] - Small typo in metadata column name (#11751) ### `llama-index-vector-stores-tidbvector` [0.1.0] - Initial support for TiDB vector store (#11635) ### `llama-index-vector-stores-weaviate` [0.1.4] - Small fix for `int` fields in metadata filters (#11742) ## [2024-03-06] New format! Going to try out reporting changes per package. ### `llama-index-cli` [0.1.8] - Update mappings for `upgrade` command (#11699) ### `llama-index-core` [0.10.17] - add `relative_score` and `dist_based_score` to `QueryFusionRetriever` (#11667) - check for `none` in async agent queue (#11669) - allow refine template for `BaseSQLTableQueryEngine` (#11378) - update mappings for llama-packs (#11699) - fixed index error for extracting rel texts in KG index (#11695) - return proper response types from synthesizer when no nodes (#11701) - Inherit metadata to summaries in DocumentSummaryIndex (#11671) - Inherit callback manager in sql query engines (#11662) - Fixed bug with agent streaming not being written to chat history (#11675) - Fixed a small bug with `none` deltas when streaming a function call with an agent (#11713) ### `llama-index-multi-modal-llms-anthropic` [0.1.2] - Added support for new multi-modal models `haiku` and `sonnet` (#11656) ### `llama-index-packs-finchat` [0.1.0] - Added a new llama-pack for hierarchical agents + finance chat (#11387) ### `llama-index-readers-file` [0.1.8] - Added support for checking if NLTK files are already downloaded (#11676) ### `llama-index-readers-json` [0.1.4] - Use the metadata passed in when creating documents (#11626) ### `llama-index-vector-stores-astra-db` [0.1.4] - Update wording in warning message (#11702) ### `llama-index-vector-stores-opensearch` [0.1.7] - Avoid calling `nest_asyncio.apply()` in code to avoid confusing errors for users (#11707) ### `llama-index-vector-stores-qdrant` [0.1.4] - Catch RPC errors (#11657) ## [0.10.16] - 2024-03-05 ### New Features - Anthropic support for new models (#11623, #11612) - Easier creation of chat prompts (#11583) - Added a raptor retriever llama-pack (#11527) - Improve batch cohere embeddings through bedrock (#11572) - Added support for vertex AI embeddings (#11561) ### Bug Fixes / Nits - Ensure order in async embeddings generation (#11562) - Fixed empty metadata for csv reader (#11563) - Serializable fix for composable retrievers (#11617) - Fixed milvus metadata filter support (#11566) - FIxed pydantic import in clickhouse vector store (#11631) - Fixed system prompts for gemini/vertext-gemini (#11511) ## [0.10.15] - 2024-03-01 ### New Features - Added FeishuWikiReader (#11491) - Added videodb retriever integration (#11463) - Added async to opensearch vector store (#11513) - New LangFuse one-click callback handler (#11324) ### Bug Fixes / Nits - Fixed deadlock issue with async chat streaming (#11548) - Improved hidden file check in SimpleDirectoryReader (#11496) - Fixed null values in document metadata when using SimpleDirectoryReader (#11501) - Fix for sqlite utils in jsonalyze query engine (#11519) - Added base url and timeout to ollama multimodal LLM (#11526) - Updated duplicate handling in query fusion retriever (#11542) - Fixed bug in kg indexx struct updating (#11475) ## [0.10.14] - 2024-02-28 ### New Features - Released llama-index-networks (#11413) - Jina reranker (#11291) - Added DuckDuckGo agent search tool (#11386) - helper functions for chatml (#10272) - added brave search tool for agents (#11468) - Added Friendli LLM integration (#11384) - metadata only queries for chromadb (#11328) ### Bug Fixes / Nits - Fixed inheriting llm callback in synthesizers (#11404) - Catch delete error in milvus (#11315) - Fixed pinecone kwargs issue (#11422) - Supabase metadata filtering fix (#11428) - api base fix in gemini embeddings (#11393) - fix elasticsearch vector store await (#11438) - vllm server cuda fix (#11442) - fix for passing LLM to context chat engine (#11444) - set input types for cohere embeddings (#11288) - default value for azure ad token (#10377) - added back prompt mixin for react agent (#10610) - fixed system roles for gemini (#11481) - fixed mean agg pooling returning numpy float values (#11458) - improved json path parsing for JSONQueryEngine (#9097) ## [0.10.13] - 2024-02-26 ### New Features - Added a llama-pack for KodaRetriever, for on-the-fly alpha tuning (#11311) - Added support for `mistral-large` (#11398) - Last token pooling mode for huggingface embeddings models like SFR-Embedding-Mistral (#11373) - Added fsspec support to SimpleDirectoryReader (#11303) ### Bug Fixes / Nits - Fixed an issue with context window + prompt helper (#11379) - Moved OpenSearch vector store to BasePydanticVectorStore (#11400) - Fixed function calling in fireworks LLM (#11363) - Made cohere embedding types more automatic (#11288) - Improve function calling in react agent (#11280) - Fixed MockLLM imports (#11376) ## [0.10.12] - 2024-02-22 ### New Features - Added `llama-index-postprocessor-colbert-rerank` package (#11057) - `MyMagicAI` LLM (#11263) - `MariaTalk` LLM (#10925) - Add retries to github reader (#10980) - Added FireworksAI embedding and LLM modules (#10959) ### Bug Fixes / Nits - Fixed string formatting in weaviate (#11294) - Fixed off-by-one error in semantic splitter (#11295) - Fixed `download_llama_pack` for multiple files (#11272) - Removed `BUILD` files from packages (#11267) - Loosened python version reqs for all packages (#11267) - Fixed args issue with chromadb (#11104) ## [0.10.11] - 2024-02-21 ### Bug Fixes / Nits - Fixed multi-modal LLM for async acomplete (#11064) - Fixed issue with llamaindex-cli imports (#11068) ## [0.10.10] - 2024-02-20 I'm still a bit wonky with our publishing process -- apologies. This is just a version bump to ensure the changes that were supposed to happen in 0.10.9 actually did get published. (AF) ## [0.10.9] - 2024-02-20 - add llama-index-cli dependency ## [0.10.7] - 2024-02-19 ### New Features - Added Self-Discover llamapack (#10951) ### Bug Fixes / Nits - Fixed linting in CICD (#10945) - Fixed using remote graph stores (#10971) - Added missing LLM kwarg in NoText response synthesizer (#10971) - Fixed openai import in rankgpt (#10971) - Fixed resolving model name to string in openai embeddings (#10971) - Off by one error in sentence window node parser (#10971) ## [0.10.6] - 2024-02-17 First, apologies for missing the changelog the last few versions. Trying to figure out the best process with 400+ packages. At some point, each package will have a dedicated changelog. But for now, onto the "master" changelog. ### New Features - Added `NomicHFEmbedding` (#10762) - Added `MinioReader` (#10744) ### Bug Fixes / Nits - Various fixes for clickhouse vector store (#10799) - Fix index name in neo4j vector store (#10749) - Fixes to sagemaker embeddings (#10778) - Fixed performance issues when splitting nodes (#10766) - Fix non-float values in reranker + b25 (#10930) - OpenAI-agent should be a dep of openai program (#10930) - Add missing shortcut imports for query pipeline components (#10930) - Fix NLTK and tiktoken not being bundled properly with core (#10930) - Add back `llama_index.core.__version__` (#10930) ## [0.10.3] - 2024-02-13 ### Bug Fixes / Nits - Fixed passing in LLM to `as_chat_engine` (#10605) - Fixed system prompt formatting for anthropic (#10603) - Fixed elasticsearch vector store error on `__version__` (#10656) - Fixed import on openai pydantic program (#10657) - Added client back to opensearch vector store exports (#10660) - Fixed bug in SimpleDirectoryReader not using file loaders properly (#10655) - Added lazy LLM initialization to RankGPT (#10648) - Fixed bedrock embedding `from_credentials` passing ing the model name (#10640) - Added back recent changes to TelegramReader (#10625) ## [0.10.0, 0.10.1] - 2024-02-12 ### Breaking Changes - Several changes are introduced. See the [full blog post](https://blog.llamaindex.ai/llamaindex-v0-10-838e735948f8) for complete details. ## [0.9.48] - 2024-02-12 ### Bug Fixes / Nits - Add back deprecated API for BedrockEmbdding (#10581) ## [0.9.47] - 2024-02-11 Last patch before v0.10! ### New Features - add conditional links to query pipeline (#10520) - refactor conditional links + add to cookbook (#10544) - agent + query pipeline cleanups (#10563) ### Bug Fixes / Nits - Add sleep to fix lag in chat stream (#10339) - OllamaMultiModal kwargs (#10541) - Update Ingestion Pipeline to handle empty documents (#10543) - Fixing minor spelling error (#10516) - fix elasticsearch async check (#10549) - Docs/update slack demo colab (#10534) - Adding the possibility to use the IN operator for PGVectorStore (#10547) - fix agent reset (#10562) - Fix MD duplicated Node id from multiple docs (#10564) ## [0.9.46] - 2024-02-08 ### New Features - Update pooling strategy for embedding models (#10536) - Add Multimodal Video RAG example (#10530) - Add SECURITY.md (#10531) - Move agent module guide up one-level (#10519) ### Bug Fixes / Nits - Deeplake fixes (#10529) - Add Cohere section for llamaindex (#10523) - Fix md element (#10510) ## [0.9.45.post1] - 2024-02-07 ### New Features - Upgraded deeplake vector database to use BasePydanticVectorStore (#10504) ### Bug Fixes / Nits - Fix MD parser for inconsistency tables (#10488) - Fix ImportError for pypdf in MetadataExtractionSEC.ipynb (#10491) ## [0.9.45] - 2024-02-07 ### New Features - Refactor: add AgentRunner.from_llm method (#10452) - Support custom prompt formatting for non-chat LLMS (#10466) - Bump cryptography from 41.0.7 to 42.0.0 (#10467) - Add persist and load method for Colbert Index (#10477) - Allow custom agent to take in user inputs (#10450) ### Bug Fixes / Nits - remove exporter from arize-phoenix global callback handler (#10465) - Fixing Dashscope qwen llm bug (#10471) - Fix: calling AWS Bedrock models (#10443) - Update Azure AI Search (fka Azure Cognitive Search) vector store integration to latest client SDK 11.4.0 stable + updating jupyter notebook sample (#10416) - fix some imports (#10485) ## [0.9.44] - 2024-02-05 ### New Features - ollama vision cookbook (#10438) - Support Gemini "transport" configuration (#10457) - Add Upstash Vector (#10451) ## [0.9.43] - 2024-02-03 ### New Features - Add multi-modal ollama (#10434) ### Bug Fixes / Nits - update base class for astradb (#10435) ## [0.9.42.post1] - 2024-02-02 ### New Features - Add Async support for Base nodes parser (#10418) ## [0.9.42] - 2024-02-02 ### New Features - Add support for `gpt-3.5-turbo-0125` (#10412) - Added `create-llama` support to rag cli (#10405) ### Bug Fixes / Nits - Fixed minor bugs in lance-db vector store (#10404) - Fixed streaming bug in ollama (#10407) ## [0.9.41] - 2024-02-01 ### New Features - Nomic Embedding (#10388) - Dashvector support sparse vector (#10386) - Table QA with MarkDownParser and Benchmarking (#10382) - Simple web page reader (#10395) ### Bug Fixes / Nits - fix full node content in KeywordExtractor (#10398) ## [0.9.40] - 2024-01-30 ### New Features - Improve and fix bugs for MarkdownElementNodeParser (#10340) - Fixed and improve Perplexity support for new models (#10319) - Ensure system_prompt is passed to Perplexity LLM (#10326) - Extended BaseRetrievalEvaluator to include an optional PostProcessor (#10321) ## [0.9.39] - 2024-01-26 ### New Features - Support for new GPT Turbo Models (#10291) - Support Multiple docs for Sentence Transformer Fine tuning(#10297) ### Bug Fixes / Nits - Marvin imports fixed (#9864) ## [0.9.38] - 2024-01-25 ### New Features - Support for new OpenAI v3 embedding models (#10279) ### Bug Fixes / Nits - Extra checks on sparse embeddings for qdrant (#10275) ## [0.9.37] - 2024-01-24 ### New Features - Added a RAG CLI utility (#10193) - Added a textai vector store (#10240) - Added a Postgresql based docstore and index store (#10233) - specify tool spec in tool specs (#10263) ### Bug Fixes / Nits - Fixed serialization error in ollama chat (#10230) - Added missing fields to `SentenceTransformerRerank` (#10225) - Fixed title extraction (#10209, #10226) - nit: make chainable output parser more exposed in library/docs (#10262) - :bug: summary index not carrying over excluded metadata keys (#10259) ## [0.9.36] - 2024-01-23 ### New Features - Added support for `SageMakerEmbedding` (#10207) ### Bug Fixes / Nits - Fix duplicated `file_id` on openai assistant (#10223) - Fix circular dependencies for programs (#10222) - Run `TitleExtractor` on groups of nodes from the same parent document (#10209) - Improve vectara auto-retrieval (#10195) ## [0.9.35] - 2024-01-22 ### New Features - `beautifulsoup4` dependency to new optional extra `html` (#10156) - make `BaseNode.hash` an `@property` (#10163) - Neutrino (#10150) - feat: JSONalyze Query Engine (#10067) - [wip] add custom hybrid retriever notebook (#10164) - add from_collection method to ChromaVectorStore class (#10167) - CLI experiment v0: ask (#10168) - make react agent prompts more editable (#10154) - Add agent query pipeline (#10180) ### Bug Fixes / Nits - Update supabase vecs metadata filter function to support multiple fields (#10133) - Bugfix/code improvement for LanceDB integration (#10144) - `beautifulsoup4` optional dependency (#10156) - Fix qdrant aquery hybrid search (#10159) - make hash a @property (#10163) - fix: bug on poetry install of llama-index[postgres] (#10171) - [doc] update jaguar vector store documentation (#10179) - Remove use of not-launched finish_message (#10188) - Updates to Lantern vector stores docs (#10192) - fix typo in multi_document_agents.ipynb (#10196) ## [0.9.34] - 2024-01-19 ### New Features - Added SageMakerEndpointLLM (#10140) - Added support for Qdrant filters (#10136) ### Bug Fixes / Nits - Update bedrock utils for Claude 2:1 (#10139) - BugFix: deadlocks using multiprocessing (#10125) ## [0.9.33] - 2024-01-17 ### New Features - Added RankGPT as a postprocessor (#10054) - Ensure backwards compatibility with new Pinecone client version bifucation (#9995) - Recursive retriever all the things (#10019) ### Bug Fixes / Nits - BugFix: When using markdown element parser on a table containing comma (#9926) - extend auto-retrieval notebook (#10065) - Updated the Attribute name in llm_generators (#10070) - jaguar vector store add text_tag to add_kwargs in add() (#10057) ## [0.9.32] - 2024-01-16 ### New Features - added query-time row retrieval + fix nits with query pipeline over structured data (#10061) - ReActive Agents w/ Context + updated stale link (#10058) ## [0.9.31] - 2024-01-15 ### New Features - Added selectors and routers to query pipeline (#9979) - Added sparse-only search to qdrant vector store (#10041) - Added Tonic evaluators (#10000) - Adding async support to firestore docstore (#9983) - Implement mongodb docstore `put_all` method (#10014) ### Bug Fixes / Nits - Properly truncate sql results based on `max_string_length` (#10015) - Fixed `node.resolve_image()` for base64 strings (#10026) - Fixed cohere system prompt role (#10020) - Remove redundant token counting operation in SentenceSplitter (#10053) ## [0.9.30] - 2024-01-11 ### New Features - Implements a Node Parser using embeddings for Semantic Splitting (#9988) - Add Anyscale Embedding model support (#9470) ### Bug Fixes / Nits - nit: fix pandas get prompt (#10001) - Fix: Token counting bug (#9912) - Bump jinja2 from 3.1.2 to 3.1.3 (#9997) - Fix corner case for qdrant hybrid search (#9993) - Bugfix: sphinx generation errors (#9944) - Fix: `language` used before assignment in `CodeSplitter` (#9987) - fix inconsistent name "text_parser" in section "Use a Text Splitter… (#9980) - :bug: fixing batch size (#9982) - add auto-async execution to query pipelines (#9967) - :bug: fixing init (#9977) - Parallel Loading with SimpleDirectoryReader (#9965) - do not force delete an index in milvus (#9974) ## [0.9.29] - 2024-01-10 ### New Features - Added support for together.ai models (#9962) - Added support for batch redis/firestore kvstores, async firestore kvstore (#9827) - Parallelize `IngestionPipeline.run()` (#9920) - Added new query pipeline components: function, argpack, kwargpack (#9952) ### Bug Fixes / Nits - Updated optional langchain imports to avoid warnings (#9964) - Raise an error if empty nodes are embedded (#9953) ## [0.9.28] - 2024-01-09 ### New Features - Added support for Nvidia TenorRT LLM (#9842) - Allow `tool_choice` to be set during agent construction (#9924) - Added streaming support for `QueryPipeline` (#9919) ### Bug Fixes / Nits - Set consistent doc-ids for llama-index readers (#9923, #9916) - Remove unneeded model inputs for HuggingFaceEmbedding (#9922) - Propagate `tool_choice` flag to downstream APIs (#9901) - Add `chat_store_key` to chat memory `from_defaults()` (#9928) ## [0.9.27] - 2024-01-08 ### New Features - add query pipeline (#9908) - Feature: Azure Multi Modal (fixes: #9471) (#9843) - add postgres docker (#9906) - Vectara auto_retriever (#9865) - Redis Chat Store support (#9880) - move more classes to core (#9871) ### Bug Fixes / Nits / Smaller Features - Propagate `tool_choice` flag to downstream APIs (#9901) - filter out negative indexes from faiss query (#9907) - added NE filter for qdrant payloads (#9897) - Fix incorrect id assignment in MyScale query result (#9900) - Qdrant Text Match Filter (#9895) - Fusion top k for hybrid search (#9894) - Fix (#9867) sync_to_async to avoid blocking during asynchronous calls (#9869) - A single node passed into compute_scores returns as a float (#9866) - Remove extra linting steps (#9878) - add vectara links (#9886) ## [0.9.26] - 2024-01-05 ### New Features - Added a `BaseChatStore` and `SimpleChatStore` abstraction for dedicated chat memory storage (#9863) - Enable custom `tree_sitter` parser to be passed into `CodeSplitter` (#9845) - Created a `BaseAutoRetriever` base class, to allow other retrievers to extend to auto modes (#9846) - Added support for Nvidia Triton LLM (#9488) - Added `DeepEval` one-click observability (#9801) ### Bug Fixes / Nits - Updated the guidance integration to work with the latest version (#9830) - Made text storage optional for doctores/ingestion pipeline (#9847) - Added missing `sphinx-automodapi` dependency for docs (#9852) - Return actual node ids in weaviate query results (#9854) - Added prompt formatting to LangChainLLM (#9844) ## [0.9.25] - 2024-01-03 ### New Features - Added concurrancy limits for dataset generation (#9779) - New `deepeval` one-click observability handler (#9801) - Added jaguar vector store (#9754) - Add beta multimodal ReAct agent (#9807) ### Bug Fixes / Nits - Changed default batch size for OpenAI embeddings to 100 (#9805) - Use batch size properly for qdrant upserts (#9814) - `_verify_source_safety` uses AST, not regexes, for proper safety checks (#9789) - use provided LLM in element node parsers (#9776) - updated legacy vectordb loading function to be more robust (#9773) - Use provided http client in AzureOpenAI (#9772) ## [0.9.24] - 2023-12-30 ### New Features - Add reranker for BEIR evaluation (#9743) - Add Pathway integration. (#9719) - custom agents implementation + notebook (#9746) ### Bug Fixes / Nits - fix beam search for vllm: add missing parameter (#9741) - Fix alpha for hrbrid search (#9742) - fix token counter (#9744) - BM25 tokenizer lowercase (#9745) ## [0.9.23] - 2023-12-28 ### Bug Fixes / Nits - docs: fixes qdrant_hybrid.ipynb typos (#9729) - make llm completion program more general (#9731) - Refactor MM Vector store and Index for empty collection (#9717) - Adding IF statement to check for Schema using "Select" (#9712) - allow skipping module loading in `download_module` and `download_llama_pack` (#9734) ## [0.9.22] - 2023-12-26 ### New Features - Added `.iter_data()` method to `SimpleDirectoryReader` (#9658) - Added async support to `Ollama` LLM (#9689) - Expanding pinecone filter support for `in` and `not in` (#9683) ### Bug Fixes / Nits - Improve BM25Retriever performance (#9675) - Improved qdrant hybrid search error handling (#9707) - Fixed `None` handling in `ChromaVectorStore` (#9697) - Fixed postgres schema creation if not existing (#9712) ## [0.9.21] - 2023-12-23 ### New Features - Added zilliz cloud as a managed index (#9605) ### Bug Fixes / Nits - Bedrock client and LLM fixes (#9671, #9646) ## [0.9.20] - 2023-12-21 ### New Features - Added `insert_batch_size` to limit number of embeddings held in memory when creating an index, defaults to 2048 (#9630) - Improve auto-retrieval (#9647) - Configurable Node ID Generating Function (#9574) - Introduced action input parser (#9575) - qdrant sparse vector support (#9644) - Introduced upserts and delete in ingestion pipeline (#9643) - Add Zilliz Cloud Pipeline as a Managed Index (#9605) - Add support for Google Gemini models via VertexAI (#9624) - support allowing additional metadata filters on autoretriever (#9662) ### Bug Fixes / Nits - Fix pip install commands in LM Format Enforcer notebooks (#9648) - Fixing some more links and documentations (#9633) - some bedrock nits and fixes (#9646) ## [0.9.19] - 2023-12-20 ### New Features - new llama datasets `LabelledEvaluatorDataset` & `LabelledPairwiseEvaluatorDataset` (#9531) ## [0.9.18] - 2023-12-20 ### New Features - multi-doc auto-retrieval guide (#9631) ### Bug Fixes / Nits - fix(vllm): make Vllm's 'complete' method behave the same as other LLM class (#9634) - FIx Doc links and other documentation issue (#9632) ## [0.9.17] - 2023-12-19 ### New Features - [example] adding user feedback (#9601) - FEATURE: Cohere ReRank Relevancy Metric for Retrieval Eval (#9495) ### Bug Fixes / Nits - Fix Gemini Chat Mode (#9599) - Fixed `types-protobuf` from being a primary dependency (#9595) - Adding an optional auth token to the TextEmbeddingInference class (#9606) - fix: out of index get latest tool call (#9608) - fix(azure_openai.py): add missing return to subclass override (#9598) - fix mix up b/w 'formatted' and 'format' params for ollama api call (#9594) ## [0.9.16] - 2023-12-18 ### New Features - agent refactor: step-wise execution (#9584) - Add OpenRouter, with Mixtral demo (#9464) - Add hybrid search to neo4j vector store (#9530) - Add support for auth service accounts for Google Semantic Retriever (#9545) ### Bug Fixes / Nits - Fixed missing `default=None` for `LLM.system_prompt` (#9504) - Fix #9580 : Incorporate metadata properly (#9582) - Integrations: Gradient[Embeddings,LLM] - sdk-upgrade (#9528) - Add mixtral 8x7b model to anyscale available models (#9573) - Gemini Model Checks (#9563) - Update OpenAI fine-tuning with latest changes (#9564) - fix/Reintroduce `WHERE` filter to the Sparse Query for PgVectorStore (#9529) - Update Ollama API to ollama v0.1.16 (#9558) - ollama: strip invalid `formatted` option (#9555) - add a device in optimum push #9541 (#9554) - Title vs content difference for Gemini Embedding (#9547) - fix pydantic fields to float (#9542) ## [0.9.15] - 2023-12-13 ### New Features - Added full support for Google Gemini text+vision models (#9452) - Added new Google Semantic Retriever (#9440) - added `from_existing()` method + async support to OpenAI assistants (#9367) ### Bug Fixes / Nits - Fixed huggingface LLM system prompt and messages to prompt (#9463) - Fixed ollama additional kwargs usage (#9455) ## [0.9.14] - 2023-12-11 ### New Features - Add MistralAI LLM (#9444) - Add MistralAI Embeddings (#9441) - Add `Ollama` Embedding class (#9341) - Add `FlagEmbeddingReranker` for reranking (#9285) - feat: PgVectorStore support advanced metadata filtering (#9377) - Added `sql_only` parameter to SQL query engines to avoid executing SQL (#9422) ### Bug Fixes / Nits - Feat/PgVector Support custom hnsw.ef_search and ivfflat.probes (#9420) - fix F1 score definition, update copyright year (#9424) - Change more than one image input for Replicate Multi-modal models from error to warning (#9360) - Removed GPT-Licensed `aiostream` dependency (#9403) - Fix result of BedrockEmbedding with Cohere model (#9396) - Only capture valid tool names in react agent (#9412) - Fixed `top_k` being multiplied by 10 in azure cosmos (#9438) - Fixed hybrid search for OpenSearch (#9430) ### Breaking Changes - Updated the base `LLM` interface to match `LLMPredictor` (#9388) - Deprecated `LLMPredictor` (#9388) ## [0.9.13] - 2023-12-06 ### New Features - Added batch prediction support for `LabelledRagDataset` (#9332) ### Bug Fixes / Nits - Fixed save and load for faiss vector store (#9330) ## [0.9.12] - 2023-12-05 ### New Features - Added an option `reuse_client` to openai/azure to help with async timeouts. Set to `False` to see improvements (#9301) - Added support for `vLLM` llm (#9257) - Add support for python 3.12 (#9304) - Support for `claude-2.1` model name (#9275) ### Bug Fixes / Nits - Fix embedding format for bedrock cohere embeddings (#9265) - Use `delete_kwargs` for filtering in weaviate vector store (#9300) - Fixed automatic qdrant client construction (#9267) ## [0.9.11] - 2023-12-03 ### New Features - Make `reference_contexts` optional in `LabelledRagDataset` (#9266) - Re-organize `download` module (#9253) - Added document management to ingestion pipeline (#9135) - Add docs for `LabelledRagDataset` (#9228) - Add submission template notebook and other doc updates for `LabelledRagDataset` (#9273) ### Bug Fixes / Nits - Convert numpy to list for `InstructorEmbedding` (#9255) ## [0.9.10] - 2023-11-30 ### New Features - Advanced Metadata filter for vector stores (#9216) - Amazon Bedrock Embeddings New models (#9222) - Added PromptLayer callback integration (#9190) - Reuse file ids for `OpenAIAssistant` (#9125) ### Breaking Changes / Deprecations - Deprecate ExactMatchFilter (#9216) ## [0.9.9] - 2023-11-29 ### New Features - Add new abstractions for `LlamaDataset`'s (#9165) - Add metadata filtering and MMR mode support for `AstraDBVectorStore` (#9193) - Allowing newest `scikit-learn` versions (#9213) ### Breaking Changes / Deprecations - Added `LocalAI` demo and began deprecation cycle (#9151) - Deprecate `QueryResponseDataset` and `DatasetGenerator` of `evaluation` module (#9165) ### Bug Fixes / Nits - Fix bug in `download_utils.py` with pointing to wrong repo (#9215) - Use `azure_deployment` kwarg in `AzureOpenAILLM` (#9174) - Fix similarity score return for `AstraDBVectorStore` Integration (#9193) ## [0.9.8] - 2023-11-26 ### New Features - Add `persist` and `persist_from_dir` methods to `ObjectIndex` that are able to support it (#9064) - Added async metadata extraction + pipeline support (#9121) - Added back support for start/end char idx in nodes (#9143) ### Bug Fixes / Nits - Fix for some kwargs not being set properly in global service context (#9137) - Small fix for `memory.get()` when system/prefix messages are large (#9149) - Minor fixes for global service context (#9137) ## [0.9.7] - 2023-11-24 ### New Features - Add support for `PGVectoRsStore` (#9087) - Enforcing `requests>=2.31` for security, while unpinning `urllib3` (#9108) ### Bug Fixes / Nits - Increased default memory token limit for context chat engine (#9123) - Added system prompt to `CondensePlusContextChatEngine` that gets prepended to the `context_prompt` (#9123) - Fixed bug in `CondensePlusContextChatEngine` not using chat history properly (#9129) ## [0.9.6] - 2023-11-22 ### New Features - Added `default_headers` argument to openai LLMs (#9090) - Added support for `download_llama_pack()` and LlamaPack integrations - Added support for `llamaindex-cli` command line tool ### Bug Fixed / Nits - store normalize as bool for huggingface embedding (#9089) ## [0.9.5] - 2023-11-21 ### Bug Fixes / Nits - Fixed bug with AzureOpenAI logic for inferring if stream chunk is a tool call (#9018) ### New Features - `FastEmbed` embeddings provider (#9043) - More precise testing of `OpenAILike` (#9026) - Added callback manager to each retriever (#8871) - Ability to bypass `max_tokens` inference with `OpenAILike` (#9032) ### Bug Fixes / Nits - Fixed bug in formatting chat prompt templates when estimating chunk sizes (#9025) - Sandboxed Pandas execution, remediate CVE-2023-39662 (#8890) - Restored `mypy` for Python 3.8 (#9031) - Loosened `dataclasses-json` version range, and removes unnecessary `jinja2` extra from `pandas` (#9042) ## [0.9.4] - 2023-11-19 ### New Features - Added `CondensePlusContextChatEngine` (#8949) ### Smaller Features / Bug Fixes / Nits - Fixed bug with `OpenAIAgent` inserting errors into chat history (#9000) - Fixed various bugs with LiteLLM and the new OpenAI client (#9003) - Added context window attribute to perplexity llm (#9012) - Add `node_parser` attribute back to service context (#9013) - Refactor MM retriever classes (#8998) - Fix TextNode instantiation on SupabaseVectorIndexDemo (#8994) ## [0.9.3] - 2023-11-17 ### New Features - Add perplexity LLM integration (#8734) ### Bug Fixes / Nits - Fix token counting for new openai client (#8981) - Fix small pydantic bug in postgres vector db (#8962) - Fixed `chunk_overlap` and `doc_id` bugs in `HierarchicalNodeParser` (#8983) ## [0.9.2] - 2023-11-16 ### New Features - Added new notebook guide for Multi-Modal Rag Evaluation (#8945) - Added `MultiModalRelevancyEvaluator`, and `MultiModalFaithfulnessEvaluator` (#8945) ## [0.9.1] - 2023-11-15 ### New Features - Added Cohere Reranker fine-tuning (#8859) - Support for custom httpx client in `AzureOpenAI` LLM (#8920) ### Bug Fixes / Nits - Fixed issue with `set_global_service_context` not propagating settings (#8940) - Fixed issue with building index with Google Palm embeddings (#8936) - Fixed small issue with parsing ImageDocuments/Nodes that have no text (#8938) - Fixed issue with large data inserts in Astra DB (#8937) - Optimize `QueryEngineTool` for agents (#8933) ## [0.9.0] - 2023-11-15 ### New Features / Breaking Changes / Deprecations - New `IngestionPipeline` concept for ingesting and transforming data - Data ingestion and transforms are now automatically cached - Updated interface for node parsing/text splitting/metadata extraction modules - Changes to the default tokenizer, as well as customizing the tokenizer - Packaging/Installation changes with PyPi (reduced bloat, new install options) - More predictable and consistent import paths - Plus, in beta: MultiModal RAG Modules for handling text and images! - Find more details at: `https://medium.com/@llama_index/719f03282945` ## [0.8.69.post1] - 2023-11-13 ### Bug Fixes / Nits - Increase max weaivate delete size to max of 10,000 (#8887) - Final pickling remnant fix (#8902) ## [0.8.69] - 2023-11-13 ### Bug Fixes / Nits - Fixed bug in loading pickled objects (#8880) - Fix `custom_path` vs `custom_dir` in `download_loader` (#8865) ## [0.8.68] - 2023-11-11 ### New Features - openai assistant agent + advanced retrieval cookbook (#8863) - add retrieval API benchmark (#8850) - Add JinaEmbedding class (#8704) ### Bug Fixes / Nits - Improved default timeouts/retries for OpenAI (#8819) - Add back key validation for OpenAI (#8819) - Disable automatic LLM/Embedding model downloads, give informative error (#8819) - fix openai assistant tool creation + retrieval notebook (#8862) - Quick fix Replicate MultiModal example (#8861) - fix: paths treated as hidden (#8860) - fix Replicate multi-modal LLM + notebook (#8854) - Feature/citation metadata (#8722) - Fix ImageNode type from NodeWithScore for SimpleMultiModalQueryEngine (#8844) ## [0.8.67] - 2023-11-10 ### New Features - Advanced Multi Modal Retrieval Example and docs (#8822, #8823) ### Bug Fixes / Nits - Fix retriever node postprocessors for `CitationQueryEngine` (#8818) - Fix `cannot pickle 'builtins.CoreBPE' object` in most scenarios (#8835) ## [0.8.66] - 2023-11-09 ### New Features - Support parallel function calling with new OpenAI client in `OpenAIPydanticProgram` (#8793) ### Bug Fixes / Nits - Fix bug in pydantic programs with new OpenAI client (#8793) - Fixed bug with un-listable fsspec objects (#8795) ## [0.8.65] - 2023-11-08 ### New Features - `OpenAIAgent` parallel function calling (#8738) ### New Features - Properly supporting Hugging Face recommended model (#8784) ### Bug Fixes / Nits - Fixed missing import for `embeddings.__all__` (#8779) ### Breaking Changes / Deprecations - Use `tool_choice` over `function_call` and `tool` over `functions` in `OpenAI(LLM)` (#8738) - Deprecate `to_openai_function` in favor of `to_openai_tool` (#8738) ## [0.8.64] - 2023-11-06 ### New Features - `OpenAIAgent` parallel function calling (#8738) - Add AI assistant agent (#8735) - OpenAI GPT4v Abstraction (#8719) - Add support for `Lantern` VectorStore (#8714) ### Bug Fixes / Nits - Fix returning zero nodes in elastic search vector store (#8746) - Add try/except for `SimpleDirectoryReader` loop to avoid crashing on a single document (#8744) - Fix for `deployment_name` in async embeddings (#8748) ## [0.8.63] - 2023-11-05 ### New Features - added native sync and async client support for the lasted `openai` client package (#8712) - added support for `AzureOpenAIEmbedding` (#8712) ### Bug Fixes / Nits - Fixed errors about "no host supplied" with `download_loader` (#8723) ### Breaking Changes - `OpenAIEmbedding` no longer supports azure, moved into the `AzureOpenAIEmbedding` class (#8712) ## [0.8.62.post1] - 2023-11-05 ### Breaking Changes - add new devday models (#8713) - moved `max_docs` parameter from constructor to `lazy_load_data()` for `SimpleMongoReader` (#8686) ## [0.8.61] - 2023-11-05 ### New Features - [experimental] Hyperparameter tuner (#8687) ### Bug Fixes / Nits - Fix typo error in CohereAIModelName class: cohere light models was missing v3 (#8684) - Update deeplake.py (#8683) ## [0.8.60] - 2023-11-04 ### New Features - prompt optimization guide (#8659) - VoyageEmbedding (#8634) - Multilingual support for `YoutubeTranscriptReader` (#8673) - emotion prompt guide (#8674) ### Bug Fixes / Nits - Adds mistral 7b instruct v0.1 to available anyscale models (#8652) - Make pgvector's setup (extension, schema, and table creation) optional (#8656) - Allow init of stores_text variable for Pinecone vector store (#8633) - fix: azure ad support (#8667) - Fix nltk bug in multi-threaded environments (#8668) - Fix google colab link in cohereai notebook (#8677) - passing max_tokens to the `Cohere` llm (#8672) ## [0.8.59] - 2023-11-02 - Deepmemory support (#8625) - Add CohereAI embeddings (#8650) - Add Azure AD (Microsoft Entra ID) support (#8667) ## [0.8.58] - 2023-11-02 ### New Features - Add `lm-format-enforcer` integration for structured output (#8601) - Google Vertex Support (#8626) ## [0.8.57] - 2023-10-31 ### New Features - Add `VoyageAIEmbedding` integration (#8634) - Add fine-tuning evaluator notebooks (#8596) - Add `SingleStoreDB` integration (#7991) - Add support for ChromaDB PersistentClient (#8582) - Add DataStax Astra DB support (#8609) ### Bug Fixes / Nits - Update dataType in Weaviate (#8608) - In Knowledge Graph Index with hybrid retriever_mode, - return the nodes found by keyword search when 'No Relationship found' - Fix exceed context length error in chat engines (#8530) - Retrieve actual content of all the triplets from KG (#8579) - Return the nodes found by Keywords when no relationship is found by embeddings in hybrid retriever_mode in `KnowledgeGraphIndex` (#8575) - Optimize content of retriever tool and minor bug fix (#8588) ## [0.8.56] - 2023-10-30 ### New Features - Add Amazon `BedrockEmbedding` (#8550) - Moves `HuggingFaceEmbedding` to center on `Pooling` enum for pooling (#8467) - Add IBM WatsonX LLM support (#8587) ### Bug Fixes / Nits - [Bug] Patch Clarifai classes (#8529) - fix retries for bedrock llm (#8528) - Fix : VectorStore’s QueryResult always returns saved Node as TextNode (#8521) - Added default file_metadata to get basic metadata that many postprocessors use, for SimpleDirectoryReader (#8486) - Handle metadata with None values in chromadb (#8584) ## [0.8.55] - 2023-10-29 ### New Features - allow prompts to take in functions with `function_mappings` (#8548) - add advanced prompt + "prompt engineering for RAG" notebook (#8555) - Leverage Replicate API for serving LLaVa modal (#8539) ### Bug Fixes / Nits - Update pull request template with google colab support inclusion (#8525) ## [0.8.54] - 2023-10-28 ### New Features - notebook showing how to fine-tune llama2 on structured outputs (#8540) - added GradientAIFineTuningHandler - added pydantic_program_mode to ServiceContext - Initialize MultiModal Retrieval using LlamaIndex (#8507) ### Bug Fixes / Nits - Add missing import to `ChatEngine` usage pattern `.md` doc (#8518) - :bug: fixed async add (#8531) - fix: add the needed CondenseQuestionChatEngine import in the usage_pa… (#8518) - Add import LongLLMLinguaPostprocessor for LongLLMLingua.ipynb (#8519) ## [0.8.53] - 2023-10-27 ### New Features - Docs refactor (#8500) An overhaul of the docs organization. Major changes - Added a big new "understanding" section - Added a big new "optimizing" section - Overhauled Getting Started content - Categorized and moved module guides to a single section ## [0.8.52] - 2023-10-26 ### New Features - Add longllmlingua (#8485) - Add google colab support for notebooks (#7560) ### Bug Fixes / Nits - Adapt Cassandra VectorStore constructor DB connection through cassio.init (#8255) - Allow configuration of service context and storage context in managed index (#8487) ## [0.8.51.post1] - 2023-10-25 ### New Features - Add Llava MultiModal QA examples for Tesla 10k RAG (#8271) - fix bug streaming on react chat agent not working as expected (#8459) ### Bug Fixes / Nits - patch: add selected result to response metadata for router query engines, fix bug (#8483) - add Jina AI embeddings notebook + huggingface embedding fix (#8478) - add `is_chat_model` to replicate (#8469) - Brought back `toml-sort` to `pre-commit` (#8267) - Added `LocationConstraint` for local `test_s3_kvstore` (#8263) ## [0.8.50] - 2023-10-24 ### New Features - Expose prompts in different modules (query engines, synthesizers, and more) (#8275) ## [0.8.49] - 2023-10-23 ### New Features - New LLM integrations - Support for Hugging Face Inference API's `conversational`, `text_generation`, and `feature_extraction` endpoints via `huggingface_hub[inference]` (#8098) - Add Amazon Bedrock LLMs (#8223) - Add AI21 Labs LLMs (#8233) - Add OpenAILike LLM class for OpenAI-compatible api servers (#7973) - New / updated vector store integrations - Add DashVector (#7772) - Add Tencent VectorDB (#8173) - Add option for custom Postgres schema on PGVectorStore instead of only allowing public schema (#8080) - Add Gradient fine tuning engine (#8208) - docs(FAQ): frequently asked questions (#8249) ### Bug Fixes / Nits - Fix inconsistencies with `ReActAgent.stream_chat` (#8147) - Deprecate some functions for GuardrailsOutputParser (#8016) - Simplify dependencies (#8236) - Bug fixes for LiteLLM (#7885) - Update for Predibase LLM (#8211) ## [0.8.48] - 2023-10-20 ### New Features - Add `DELETE` for MyScale vector store (#8159) - Add SQL Retriever (#8197) - add semantic kernel document format (#8226) - Improve MyScale Hybrid Search and Add `DELETE` for MyScale vector store (#8159) ### Bug Fixes / Nits - Fixed additional kwargs in ReActAgent.from_tools() (#8206) - Fixed missing spaces in prompt templates (#8190) - Remove auto-download of llama2-13B on exception (#8225) ## [0.8.47] - 2023-10-19 ### New Features - add response synthesis to text-to-SQL (#8196) - Added support for `LLMRailsEmbedding` (#8169) - Inferring MPS device with PyTorch (#8195) - Consolidated query/text prepending (#8189) ## [0.8.46] - 2023-10-18 ### New Features - Add fine-tuning router support + embedding selector (#8174) - add more document converters (#8156) ### Bug Fixes / Nits - Add normalization to huggingface embeddings (#8145) - Improve MyScale Hybrid Search (#8159) - Fixed duplicate `FORMAT_STR` being inside prompt (#8171) - Added: support for output_kwargs={'max_colwidth': xx} for PandasQueryEngine (#8110) - Minor fix in the description for an argument in cohere llm (#8163) - Fix Firestore client info (#8166) ## [0.8.45] - 2023-10-13 ### New Features - Added support for fine-tuning cross encoders (#7705) - Added `QueryFusionRetriever` for merging multiple retrievers + query augmentation (#8100) - Added `nb-clean` to `pre-commit` to minimize PR diffs (#8108) - Support for `TextEmbeddingInference` embeddings (#8122) ### Bug Fixes / Nits - Improved the `BM25Retriever` interface to accept `BaseNode` objects (#8096) - Fixed bug with `BM25Retriever` tokenizer not working as expected (#8096) - Brought mypy to pass in Python 3.8 (#8107) - `ReActAgent` adding missing `super().__init__` call (#8125) ## [0.8.44] - 2023-10-12 ### New Features - add pgvector sql query engine (#8087) - Added HoneyHive one-click observability (#7944) - Add support for both SQLAlchemy V1 and V2 (#8060) ## [0.8.43.post1] - 2023-10-11 ### New Features - Moves `codespell` to `pre-commit` (#8040) - Added `prettier` for autoformatting extensions besides `.py` (#8072) ### Bug Fixes / Nits - Fixed forgotten f-str in `HuggingFaceLLM` (#8075) - Relaxed numpy/panadas reqs ## [0.8.43] - 2023-10-10 ### New Features - Added support for `GradientEmbedding` embed models (#8050) ### Bug Fixes / Nits - added `messages_to_prompt` kwarg to `HuggingFaceLLM` (#8054) - improved selection and sql parsing for open-source models (#8054) - fixed bug when agents hallucinate too many kwargs for a tool (#8054) - improved prompts and debugging for selection+question generation (#8056) ## [0.8.42] - 2023-10-10 ### New Features - `LocalAI` more intuitive module-level var names (#8028) - Enable `codespell` for markdown docs (#7972) - add unstructured table element node parser (#8036) - Add: Async upserting for Qdrant vector store (#7968) - Add cohere llm (#8023) ### Bug Fixes / Nits - Parse multi-line outputs in react agent answers (#8029) - Add properly named kwargs to keyword `as_retriever` calls (#8011) - Updating Reference to RAGAS LlamaIndex Integration (#8035) - Vectara bugfix (#8032) - Fix: ChromaVectorStore can attempt to add in excess of chromadb batch… (#8019) - Fix get_content method in Mbox reader (#8012) - Apply kwarg filters in WeaviateVectorStore (#8017) - Avoid ZeroDivisionError (#8027) - `LocalAI` intuitive module-level var names (#8028) - zep/fix: imports & typing (#8030) - refactor: use `str.join` (#8020) - use proper metadata str for node parsing (#7987) ## [0.8.41] - 2023-10-07 ### New Features - You.com retriever (#8024) - Pull fields from mongodb into metadata with `metadata_names` argument (#8001) - Simplified `LocalAI.__init__` preserving the same behaviors (#7982) ### Bug Fixes / Nits - Use longest metadata string for metadata aware text splitting (#7987) - Handle lists of strings in mongodb reader (#8002) - Removes `OpenAI.class_type` as it was dead code (#7983) - Fixing `HuggingFaceLLM.device_map` type hint (#7989) ## [0.8.40] - 2023-10-05 ### New Features - Added support for `Clarifai` LLM (#7967) - Add support for function fine-tuning (#7971) ### Breaking Changes - Update document summary index (#7815) - change default retrieval mode to embedding - embed summaries into vector store by default at indexing time (instead of calculating embedding on the fly) - support configuring top k in llm retriever ## [0.8.39] - 2023-10-03 ### New Features - Added support for pydantic object outputs with query engines (#7893) - `ClarifaiEmbedding` class added for embedding support (#7940) - Markdown node parser, flat file reader and simple file node parser (#7863) - Added support for mongdb atlas `$vectorSearch` (#7866) ### Bug Fixes / Nits - Adds support for using message metadata in discord reader (#7906) - Fix `LocalAI` chat capability without `max_tokens` (#7942) - Added `codespell` for automated checking (#7941) - `ruff` modernization and autofixes (#7889) - Implement own SQLDatabase class (#7929) - Update LlamaCPP context_params property (#7945) - fix duplicate embedding (#7949) - Adds `codespell` tool for enforcing good spelling (#7941) - Supporting `mypy` local usage with `venv` (#7952) - Vectara - minor update (#7954) - Avoiding `pydantic` reinstalls in CI (#7956) - move tree_sitter_languages into data_requirements.txt (#7955) - Add `cache_okay` param to `PGVectorStore` to help suppress TSVector warnings (#7950) ## [0.8.38] - 2023-10-02 ### New Features - Updated `KeywordNodePostprocessor` to use spacy to support more languages (#7894) - `LocalAI` supporting global or per-query `/chat/completions` vs `/completions` (#7921) - Added notebook on using REBEL + Wikipedia filtering for knowledge graphs (#7919) - Added support for `ElasticsearchEmbedding` (#7914) ## [0.8.37] - 2023-09-30 ### New Features - Supporting `LocalAI` LLMs (#7913) - Validations protecting against misconfigured chunk sizes (#7917) ### Bug Fixes / Nits - Simplify NL SQL response to SQL parsing, with expanded NL SQL prompt (#7868) - Improve vector store retrieval speed for vectordb integrations (#7876) - Added replacing {{ and }}, and fixed JSON parsing recursion (#7888) - Nice-ified JSON decoding error (#7891) - Nice-ified SQL error from LLM not providing SQL (#7900) - Nice-ified `ImportError` for `HuggingFaceLLM` (#7904) - eval fixes: fix dataset response generation, add score to evaluators (#7915) ## [0.8.36] - 2023-09-27 ### New Features - add "build RAG from scratch notebook" - OSS/local (#7864) ### Bug Fixes / Nits - Fix elasticsearch hybrid scoring (#7852) - Replace `get_color_mapping` and `print_text` Langchain dependency with internal implementation (#7845) - Fix async streaming with azure (#7856) - Avoid `NotImplementedError()` in sub question generator (#7855) - Patch predibase initialization (#7859) - Bumped min langchain version and changed prompt imports from langchain (#7862) ## [0.8.35] - 2023-09-27 ### Bug Fixes / Nits - Fix dropping textnodes in recursive retriever (#7840) - share callback_manager between agent and its llm when callback_manager is None (#7844) - fix pandas query engine (#7847) ## [0.8.34] - 2023-09-26 ### New Features - Added `Konko` LLM support (#7775) - Add before/after context sentence (#7821) - EverlyAI integration with LlamaIndex through OpenAI library (#7820) - add Arize Phoenix tracer to global handlers (#7835) ### Bug Fixes / Nits - Normalize scores returned from ElasticSearch vector store (#7792) - Fixed `refresh_ref_docs()` bug with order of operations (#7664) - Delay postgresql connection for `PGVectorStore` until actually needed (#7793) - Fix KeyError in delete method of `SimpleVectorStore` related to metadata filters (#7829) - Fix KeyError in delete method of `SimpleVectorStore` related to metadata filters (#7831) - Addressing PyYAML import error (#7784) - ElasticsearchStore: Update User-Agent + Add example docker compose (#7832) - `StorageContext.persist` supporting `Path` (#7783) - Update ollama.py (#7839) - fix bug for self.\_session_pool (#7834) ## [0.8.33] - 2023-09-25 ### New Features - add pairwise evaluator + benchmark auto-merging retriever (#7810) ### Bug Fixes / Nits - Minor cleanup in embedding class (#7813) - Misc updates to `OpenAIEmbedding` (#7811) ## [0.8.32] - 2023-09-24 ### New Features - Added native support for `HuggingFaceEmbedding`, `InstructorEmbedding`, and `OptimumEmbedding` (#7795) - Added metadata filtering and hybrid search to MyScale vector store (#7780) - Allowing custom text field name for Milvus (#7790) - Add support for `vector_store_query_mode` to `VectorIndexAutoRetriever` (#7797) ### Bug Fixes / Nits - Update `LanceDBVectorStore` to handle score and distance (#7754) - Pass LLM to `memory_cls` in `CondenseQuestionChatEngine` (#7785) ## [0.8.31] - 2023-09-22 ### New Features - add pydantic metadata extractor (#7778) - Allow users to set the embedding dimensions in azure cognitive vector store (#7734) - Add semantic similarity evaluator (#7770) ### Bug Fixes / Nits - 📝docs: Update Chatbot Tutorial and Notebook (#7767) - Fixed response synthesizers with empty nodes (#7773) - Fix `NotImplementedError` in auto vector retriever (#7764) - Multiple kwargs values in "KnowledgeGraphQueryEngine" bug-fix (#7763) - Allow setting azure cognitive search dimensionality (#7734) - Pass service context to index for dataset generator (#7748) - Fix output parsers for selector templates (#7774) - Update Chatbot_SEC.ipynb (#7711) - linter/typechecker-friendly improvements to cassandra test (#7771) - Expose debug option of `PgVectorStore` (#7776) - llms/openai: fix Azure OpenAI by considering `prompt_filter_results` field (#7755) ## [0.8.30] - 2023-09-21 ### New Features - Add support for `gpt-3.5-turbo-instruct` (#7729) - Add support for `TimescaleVectorStore` (#7727) - Added `LongContextReorder` for lost-in-the-middle issues (#7719) - Add retrieval evals (#7738) ### Bug Fixes / Nits - Added node post-processors to async context chat engine (#7731) - Added unique index name for postgres tsv column (#7741) ## [0.8.29.post1] - 2023-09-18 ### Bug Fixes / Nits - Fix langchain import error for embeddings (#7714) ## [0.8.29] - 2023-09-18 ### New Features - Added metadata filtering to the base simple vector store (#7564) - add low-level router guide (#7708) - Add CustomQueryEngine class (#7703) ### Bug Fixes / Nits - Fix context window metadata in lite-llm (#7696) ## [0.8.28] - 2023-09-16 ### New Features - Add CorrectnessEvaluator (#7661) - Added support for `Ollama` LLMs (#7635) - Added `HWPReader` (#7672) - Simplified portkey LLM interface (#7669) - Added async operation support to `ElasticsearchStore` vector store (#7613) - Added support for `LiteLLM` (#7600) - Added batch evaluation runner (#7692) ### Bug Fixes / Nits - Avoid `NotImplementedError` for async langchain embeddings (#7668) - Imrpoved reliability of LLM selectors (#7678) - Fixed `query_wrapper_prompt` and `system_prompt` for output parsers and completion models (#7678) - Fixed node attribute inheritance in citation query engine (#7675) ### Breaking Changes - Refactor and update `BaseEvaluator` interface to be more consistent (#7661) - Use `evaluate` function for generic input - Use `evaluate_response` function with `Response` objects from llama index query engine - Update existing evaluators with more explicit naming - `ResponseEvaluator` -> `FaithfulnessEvaluator` - `QueryResponseEvaluator` -> `RelevancyEvaluator` - old names are kept as class aliases for backwards compatibility ## [0.8.27] - 2023-09-14 ### New Features - add low-level tutorial section (#7673) ### Bug Fixes / Nits - default delta should be a dict (#7665) - better query wrapper logic on LLMPredictor (#7667) ## [0.8.26] - 2023-09-12 ### New Features - add non-linear embedding adapter (#7658) - Add "finetune + RAG" evaluation to knowledge fine-tuning notebook (#7643) ### Bug Fixes / Nits - Fixed chunk-overlap for sentence splitter (#7590) ## [0.8.25] - 2023-09-12 ### New Features - Added `AGENT_STEP` callback event type (#7652) ### Bug Fixes / Nits - Allowed `simple` mode to work with `as_chat_engine()` (#7637) - Fixed index error in azure streaming (#7646) - Removed `pdb` from llama-cpp (#7651) ## [0.8.24] - 2023-09-11 ## New Features - guide: fine-tuning to memorize knowledge (#7626) - added ability to customize prompt template for eval modules (#7626) ### Bug Fixes - Properly detect `llama-cpp-python` version for loading the default GGML or GGUF `llama2-chat-13b` model (#7616) - Pass in `summary_template` properly with `RetrieverQueryEngine.from_args()` (#7621) - Fix span types in wandb callback (#7631) ## [0.8.23] - 2023-09-09 ### Bug Fixes - Make sure context and system prompt is included in prompt for first chat for llama2 (#7597) - Avoid negative chunk size error in refine process (#7607) - Fix relationships for small documents in hierarchical node parser (#7611) - Update Anyscale Endpoints integration with full streaming and async support (#7602) - Better support of passing credentials as LLM constructor args in `OpenAI`, `AzureOpenAI`, and `Anyscale` (#7602) ### Breaking Changes - Update milvus vector store to support filters and dynamic schemas (#7286) - See the [updated notebook](https://docs.llamaindex.ai/en/stable/examples/vector_stores/MilvusIndexDemo.html) for usage - Added NLTK to core dependencies to support the default sentence splitter (#7606) ## [0.8.22] - 2023-09-07 ### New Features - Added support for ElasticSearch Vector Store (#7543) ### Bug Fixes / Nits - Fixed small `_index` bug in `ElasticSearchReader` (#7570) - Fixed bug with prompt helper settings in global service contexts (#7576) - Remove newlines from openai embeddings again (#7588) - Fixed small bug with setting `query_wrapper_prompt` in the service context (#7585) ### Breaking/Deprecated API Changes - Clean up vector store interface to use `BaseNode` instead of `NodeWithEmbedding` - For majority of users, this is a no-op change - For users directly operating with the `VectorStore` abstraction and manually constructing `NodeWithEmbedding` objects, this is a minor breaking change. Use `TextNode` with `embedding` set directly, instead of `NodeWithEmbedding`. ## [0.8.21] - 2023-09-06 ### New Features - add embedding adapter fine-tuning engine + guide (#7565) - Added support for Azure Cognitive Search vector store (#7469) - Support delete in supabase (#6951) - Added support for Espilla vector store (#7539) - Added support for AnyScale LLM (#7497) ### Bug Fixes / Nits - Default to user-configurable top-k in `VectorIndexAutoRetriever` (#7556) - Catch validation errors for structured responses (#7523) - Fix streaming refine template (#7561) ## [0.8.20] - 2023-09-04 ### New Features - Added Portkey LLM integration (#7508) - Support postgres/pgvector hybrid search (#7501) - upgrade recursive retriever node reference notebook (#7537) ## [0.8.19] - 2023-09-03 ### New Features - replace list index with summary index (#7478) - rename list index to summary index part 2 (#7531) ## [0.8.18] - 2023-09-03 ### New Features - add agent finetuning guide (#7526) ## [0.8.17] - 2023-09-02 ### New Features - Make (some) loaders serializable (#7498) - add node references to recursive retrieval (#7522) ### Bug Fixes / Nits - Raise informative error when metadata is too large during splitting (#7513) - Allow langchain splitter in simple node parser (#7517) ## [0.8.16] - 2023-09-01 ### Bug Fixes / Nits - fix link to Marvin notebook in docs (#7504) - Ensure metadata is not `None` in `SimpleWebPageReader` (#7499) - Fixed KGIndex visualization (#7493) - Improved empty response in KG Index (#7493) ## [0.8.15] - 2023-08-31 ### New Features - Added support for `MarvinEntityExtractor` metadata extractor (#7438) - Added a url_metadata callback to SimpleWebPageReader (#7445) - Expanded callback logging events (#7472) ### Bug Fixes / Nits - Only convert newlines to spaces for text 001 embedding models in OpenAI (#7484) - Fix `KnowledgeGraphRagRetriever` for non-nebula indexes (#7488) - Support defined embedding dimension in `PGVectorStore` (#7491) - Greatly improved similarity calculation speed for the base vector store (#7494) ## [0.8.14] - 2023-08-30 ### New Features - feat: non-kg heterogeneous graph support in Graph RAG (#7459) - rag guide (#7480) ### Bug Fixes / Nits - Improve openai fine-tuned model parsing (#7474) - doing some code de-duplication (#7468) - support both str and templates for query_wrapper_prompt in HF LLMs (#7473) ## [0.8.13] - 2023-08-29 ### New Features - Add embedding finetuning (#7452) - Added support for RunGPT LLM (#7401) - Integration guide and notebook with DeepEval (#7425) - Added `VectorIndex` and `VectaraRetriever` as a managed index (#7440) - Added support for `to_tool_list` to detect and use async functions (#7282) ## [0.8.12] - 2023-08-28 ### New Features - add openai finetuning class (#7442) - Service Context to/from dict (#7395) - add finetuning guide (#7429) ### Smaller Features / Nits / Bug Fixes - Add example how to run FalkorDB docker (#7441) - Update root.md to use get_response_synthesizer expected type. (#7437) - Bugfix MonsterAPI Pydantic version v2/v1 support. Doc Update (#7432) ## [0.8.11.post3] - 2023-08-27 ### New Features - AutoMergingRetriever (#7420) ## [0.8.10.post1] - 2023-08-25 ### New Features - Added support for `MonsterLLM` using MonsterAPI (#7343) - Support comments fields in NebulaGraphStore and int type VID (#7402) - Added configurable endpoint for DynamoDB (#6777) - Add structured answer filtering for Refine response synthesizer (#7317) ### Bug Fixes / Nits - Use `utf-8` for json file reader (#7390) - Fix entity extractor initialization (#7407) ## [0.8.9] - 2023-08-24 ### New Features - Added support for FalkorDB/RedisGraph graph store (#7346) - Added directed sub-graph RAG (#7378) - Added support for `BM25Retriever` (#7342) ### Bug Fixes / Nits - Added `max_tokens` to `Xinference` LLM (#7372) - Support cache dir creation in multithreaded apps (#7365) - Ensure temperature is a float for openai (#7382) - Remove duplicate subjects in knowledge graph retriever (#7378) - Added support for both pydantic v1 and v2 to allow other apps to move forward (#7394) ### Breaking/Deprecated API Changes - Refactor prompt template (#7319) - Use `BasePromptTemplate` for generic typing - Use `PromptTemplate`, `ChatPromptTemplate`, `SelectorPromptTemplate` as core implementations - Use `LangchainPromptTemplate` for compatibility with Langchain prompt templates - Fully replace specific prompt classes (e.g. `SummaryPrompt`) with generic `BasePromptTemplate` for typing in codebase. - Keep `Prompt` as an alias for `PromptTemplate` for backwards compatibility. - BREAKING CHANGE: remove support for `Prompt.from_langchain_prompt`, please use `template=LangchainPromptTemplate(lc_template)` instead. ## [0.8.8] - 2023-08-23 ### New Features - `OpenAIFineTuningHandler` for collecting LLM inputs/outputs for OpenAI fine tuning (#7367) ### Bug Fixes / Nits - Add support for `claude-instant-1.2` (#7369) ## [0.8.7] - 2023-08-22 ### New Features - Support fine-tuned OpenAI models (#7364) - Added support for Cassandra vector store (#6784) - Support pydantic fields in tool functions (#7348) ### Bug Fixes / Nits - Fix infinite looping with forced function call in `OpenAIAgent` (#7363) ## [0.8.6] - 2023-08-22 ### New Features - auto vs. recursive retriever notebook (#7353) - Reader and Vector Store for BagelDB with example notebooks (#7311) ### Bug Fixes / Nits - Use service context for intermediate index in retry source query engine (#7341) - temp fix for prompt helper + chat models (#7350) - Properly skip unit-tests when packages not installed (#7351) ## [0.8.5.post2] - 2023-08-20 ### New Features - Added FireStore docstore/index store support (#7305) - add recursive agent notebook (#7330) ### Bug Fixes / Nits - Fix Azure pydantic error (#7329) - fix callback trace ids (make them a context var) (#7331) ## [0.8.5.post1] - 2023-08-18 ### New Features - Awadb Vector Store (#7291) ### Bug Fixes / Nits - Fix bug in OpenAI llm temperature type ## [0.8.5] - 2023-08-18 ### New Features - Expose a system prompt/query wrapper prompt in the service context for open-source LLMs (#6647) - Changed default MyScale index format to `MSTG` (#7288) - Added tracing to chat engines/agents (#7304) - move LLM and embeddings to pydantic (#7289) ### Bug Fixes / Nits - Fix sentence splitter bug (#7303) - Fix sentence splitter infinite loop (#7295) ## [0.8.4] - 2023-08-17 ### Bug Fixes / Nits - Improve SQL Query parsing (#7283) - Fix loading embed_model from global service context (#7284) - Limit langchain version until we migrate to pydantic v2 (#7297) ## [0.8.3] - 2023-08-16 ### New Features - Added Knowledge Graph RAG Retriever (#7204) ### Bug Fixes / Nits - accept `api_key` kwarg in OpenAI LLM class constructor (#7263) - Fix to create separate queue instances for separate instances of `StreamingAgentChatResponse` (#7264) ## [0.8.2.post1] - 2023-08-14 ### New Features - Added support for Rockset as a vector store (#7111) ### Bug Fixes - Fixed bug in service context definition that could disable LLM (#7261) ## [0.8.2] - 2023-08-14 ### New Features - Enable the LLM or embedding model to be disabled by setting to `None` in the service context (#7255) - Resolve nearly any huggingface embedding model using the `embed_model="local:"` syntax (#7255) - Async tool-calling support (#7239) ### Bug Fixes / Nits - Updated supabase kwargs for add and query (#7103) - Small tweak to default prompts to allow for more general purpose queries (#7254) - Make callback manager optional for `CustomLLM` + docs update (#7257) ## [0.8.1] - 2023-08-13 ### New Features - feat: add node_postprocessors to ContextChatEngine (#7232) - add ensemble query engine tutorial (#7247) ### Smaller Features - Allow EMPTY keys for Fastchat/local OpenAI API endpoints (#7224) ## [0.8.0] - 2023-08-11 ### New Features - Added "LLAMA_INDEX_CACHE_DIR" to control cached files (#7233) - Default to pydantic selectors when possible (#7154, #7223) - Remove the need for langchain wrappers on `embed_model` in the service context (#7157) - Metadata extractors take an `LLM` object now, in addition to `LLMPredictor` (#7202) - Added local mode + fallback to llama.cpp + llama2 (#7200) - Added local fallback for embeddings to `BAAI/bge-small-en` (#7200) - Added `SentenceWindowNodeParser` + `MetadataReplacementPostProcessor` (#7211) ### Breaking Changes - Change default LLM to gpt-3.5-turbo from text-davinci-003 (#7223) - Change prompts for compact/refine/tree_summarize to work better with gpt-3.5-turbo (#7150, #7179, #7223) - Increase default LLM temperature to 0.1 (#7180) ## [0.7.24.post1] - 2023-08-11 ### Other Changes - Reverted #7223 changes to defaults (#7235) ## [0.7.24] - 2023-08-10 ### New Features - Default to pydantic selectors when possible (#7154, #7223) - Remove the need for langchain wrappers on `embed_model` in the service context (#7157) - Metadata extractors take an `LLM` object now, in addition to `LLMPredictor` (#7202) - Added local mode + fallback to llama.cpp + llama2 (#7200) - Added local fallback for embeddings to `BAAI/bge-small-en` (#7200) - Added `SentenceWindowNodeParser` + `MetadataReplacementPostProcessor` (#7211) ### Breaking Changes - Change default LLM to gpt-3.5-turbo from text-davinci-003 (#7223) - Change prompts for compact/refine/tree_summarize to work better with gpt-3.5-turbo (#7150, #7179, #7223) - Increase default LLM temperature to 0.1 (#7180) ### Other Changes - docs: Improvements to Mendable Search (#7220) - Refactor openai agent (#7077) ### Bug Fixes / Nits - Use `1 - cosine_distance` for pgvector/postgres vector db (#7217) - fix metadata formatting and extraction (#7216) - fix(readers): Fix non-ASCII JSON Reader bug (#7086) - Chore: change PgVectorStore variable name from `sim` to `distance` for clarity (#7226) ## [0.7.23] - 2023-08-10 ### Bug Fixes / Nits - Fixed metadata formatting with custom tempalates and inheritance (#7216) ## [0.7.23] - 2023-08-10 ### New Features - Add "one click observability" page to docs (#7183) - Added Xorbits inference for local deployments (#7151) - Added Zep vector store integration (#7203) - feat/zep vectorstore (#7203) ### Bug Fixes / Nits - Update the default `EntityExtractor` model (#7209) - Make `ChatMemoryBuffer` pickleable (#7205) - Refactored `BaseOpenAIAgent` (#7077) ## [0.7.22] - 2023-08-08 ### New Features - add ensemble retriever notebook (#7190) - DOCS: added local llama2 notebook (#7146) ### Bug Fixes / Nits - Fix for `AttributeError: 'OpenAIAgent' object has no attribute 'callback_manager'` by calling super constructor within `BaseOpenAIAgent` - Remove backticks from nebula queries (#7192) ## [0.7.21] - 2023-08-07 ### New Features - Added an `EntityExtractor` for metadata extraction (#7163) ## [0.7.20] - 2023-08-06 ### New Features - add router module docs (#7171) - add retriever router (#7166) ### New Features - Added a `RouterRetriever` for routing queries to specific retrievers (#7166) ### Bug Fixes / Nits - Fix for issue where having multiple concurrent streamed responses from `OpenAIAgent` would result in interleaving of tokens across each response stream. (#7164) - fix llms callbacks issue (args[0] error) (#7165) ## [0.7.19] - 2023-08-04 ### New Features - Added metadata filtering to weaviate (#7130) - Added token counting (and all callbacks) to agents and streaming (#7122) ## [0.7.18] - 2023-08-03 ### New Features - Added `to/from_string` and `to/from_dict` methods to memory objects (#7128) - Include columns comments from db tables in table info for SQL queries (#7124) - Add Neo4j support (#7122) ### Bug Fixes / Nits - Added `Azure AD` validation support to the `AzureOpenAI` class (#7127) - add `flush=True` when printing agent/chat engine response stream (#7129) - Added `Azure AD` support to the `AzureOpenAI` class (#7127) - Update LLM question generator prompt to mention JSON markdown (#7105) - Fixed `astream_chat` in chat engines (#7139) ## [0.7.17] - 2023-08-02 ### New Features - Update `ReActAgent` to support memory modules (minor breaking change since the constructor takes `memory` instead of `chat_history`, but the main `from_tools` method remains backward compatible.) (#7116) - Update `ReActAgent` to support streaming (#7119) - Added Neo4j graph store and query engine integrations (#7122) - add object streaming (#7117) ## [0.7.16] - 2023-07-30 ### New Features - Chat source nodes (#7078) ## [0.7.15] - 2023-07-29 ### Bug Fixes / Nits - anthropic api key customization (#7082) - Fix broken link to API reference in Contributor Docs (#7080) - Update vector store docs (#7076) - Update comment (#7073) ## [0.7.14] - 2023-07-28 ### New Features - Added HotpotQADistractor benchmark evaluator (#7034) - Add metadata filter and delete support for LanceDB (#7048) - Use MetadataFilters in opensearch (#7005) - Added support for `KuzuGraphStore` (#6970) - Added `kg_triplet_extract_fn` to customize how KGs are built (#7068) ### Bug Fixes / Nits - Fix string formatting in context chat engine (#7050) - Fixed tracing for async events (#7052) - Less strict triplet extraction for KGs (#7059) - Add configurable limit to KG data retrieved (#7059) - Nebula connection improvements (#7059) - Bug fix in building source nodes for agent response (#7067) ## [0.7.13] - 2023-07-26 ### New Features - Support function calling api for AzureOpenAI (#7041) ### Bug Fixes / Nits - tune prompt to get rid of KeyError in SubQ engine (#7039) - Fix validation of Azure OpenAI keys (#7042) ## [0.7.12] - 2023-07-25 ### New Features - Added `kwargs` to `ComposableGraph` for the underlying query engines (#6990) - Validate openai key on init (#6940) - Added async embeddings and async RetrieverQueryEngine (#6587) - Added async `aquery` and `async_add` to PGVectorStore (#7031) - Added `.source_nodes` attribute to chat engine and agent responses (#7029) - Added `OpenInferenceCallback` for storing generation data in OpenInference format (#6998) ### Bug Fixes / Nits - Fix achat memory initialization for data agents (#7000) - Add `print_response_stream()` to agengt/chat engine response class (#7018) ### Bug Fixes / Nits - Fix achat memory initialization for data agents (#7000) - Add `print_response_stream()` to agengt/chat engine response class (#7018) ## [v0.7.11.post1] - 2023-07-20 ### New Features - Default to pydantic question generation when possible for sub-question query engine (#6979) ### Bug Fixes / Nits - Fix returned order of messages in large chat memory (#6979) ## [v0.7.11] - 2023-07-19 ### New Features - Added a `SentenceTransformerRerank` node post-processor for fast local re-ranking (#6934) - Add numpy support for evaluating queries in pandas query engine (#6935) - Add metadata filtering support for Postgres Vector Storage integration (#6968) - Proper llama2 support for agents and query engines (#6969) ### Bug Fixes / Nits - Added `model_name` to LLMMetadata (#6911) - Fallback to retriever service context in query engines (#6911) - Fixed `as_chat_engine()` ValueError with extra kwargs (#6971 ## [v0.7.10.post1] - 2023-07-18 ### New Features - Add support for Replicate LLM (vicuna & llama 2!) ### Bug Fixes / Nits - fix streaming for condense chat engine (#6958) ## [v0.7.10] - 2023-07-17 ### New Features - Add support for chroma v0.4.0 (#6937) - Log embedding vectors to callback manager (#6962) ### Bug Fixes / Nits - add more robust embedding timeouts (#6779) - improved connection session management on postgres vector store (#6843) ## [v0.7.9] - 2023-07-15 ### New Features - specify `embed_model="local"` to use default local embbeddings in the service context (#6806) - Add async `acall` endpoint to `BasePydanticProgram` (defaults to sync version). Implement for `OpenAIPydanticProgram` ### Bug Fixes / Nits - fix null metadata for searching existing vector dbs (#6912) - add module guide docs for `SimpleDirectoryReader` (#6916) - make sure `CondenseQuestionChatEngine` streaming chat endpoints work even if not explicitly setting `streaming=True` in the underlying query engine. ## [v0.7.8] - 2023-07-13 ### New Features - Added embedding speed benchmark (#6876) - Added BEIR retrieval benchmark (#6825) ### Bug Fixes / Nits - remove toctrees from deprecated_terms (#6895) - Relax typing dependencies (#6879) - docs: modification to evaluation notebook (#6840) - raise error if the model does not support functions (#6896) - fix(bench embeddings): bug not taking into account string length (#6899)x ## [v0.7.7] - 2023-07-13 ### New Features - Improved milvus consistency support and output fields support (#6452) - Added support for knowledge graph querying w/ cypyer+nebula (#6642) - Added `Document.example()` to create documents for fast prototyping (#6739) - Replace react chat engine to use native reactive agent (#6870) ### Bug Fixes / Nits - chore: added a help message to makefile (#6861) ### Bug Fixes / Nits - Fixed support for using SQLTableSchema context_str attribute (#6891) ## [v0.7.6] - 2023-07-12 ### New Features - Added sources to agent/chat engine responses (#6854) - Added basic chat buffer memory to agents / chat engines (#6857) - Adding load and search tool (#6871) - Add simple agent benchmark (#6869) - add agent docs (#6866) - add react agent (#6865) ### Breaking/Deprecated API Changes - Replace react chat engine with native react agent (#6870) - Set default chat mode to "best": use openai agent when possible, otherwise use react agent (#6870) ### Bug Fixes / Nits - Fixed support for legacy vector store metadata (#6867) - fix chroma notebook in docs (#6872) - update LC embeddings docs (#6868) ## [v0.7.5] - 2023-07-11 ### New Features - Add `Anthropic` LLM implementation (#6855) ### Bug Fixes / Nits - Fix indexing error in `SentenceEmbeddingOptimizer` (#6850) - fix doc for custom embedding model (#6851) - fix(silent error): Add validation to `SimpleDirectoryReader` (#6819) - Fix link in docs (#6833) - Fixes Azure gpt-35-turbo model not recognized (#6828) - Update Chatbot_SEC.ipynb (#6808) - Rename leftover original name to LlamaIndex (#6792) - patch nested traces of the same type (#6791) ## [v0.7.4] - 2023-07-08 ### New Features - `MetadataExtractor` - Documnent Metadata Augmentation via LLM-based feature extractors (#6764) ### Bug Fixes / Nits - fixed passing in query bundle to node postprocessors (#6780) - fixed error in callback manager with nested traces (#6791) ## [v0.7.3] - 2023-07-07 ### New Features - Sub question query engine returns source nodes of sub questions in the callback manager (#6745) - trulens integration (#6741) - Add sources to subquestion engine (#6745) ### Bug Fixes / Nits - Added/Fixed streaming support to simple and condense chat engines (#6717) - fixed `response_mode="no_text"` response synthesizer (#6755) - fixed error setting `num_output` and `context_window` in service context (#6766) - Fix missing as_query_engine() in tutorial (#6747) - Fixed variable sql_query_engine in the notebook (#6778) - fix required function fields (#6761) - Remove usage of stop token in Prompt, SQL gen (#6782) ## [v0.7.2] - 2023-07-06 ### New Features - Support Azure OpenAI (#6718) - Support prefix messages (e.g. system prompt) in chat engine and OpenAI agent (#6723) - Added `CBEventType.SUB_QUESTIONS` event type for tracking sub question queries/responses (#6716) ### Bug Fixes / Nits - Fix HF LLM output error (#6737) - Add system message support for langchain message templates (#6743) - Fixed applying node-postprocessors (#6749) - Add missing `CustomLLM` import under `llama_index.llms` (#6752) - fix(typo): `get_transformer_tokenizer_fn` (#6729) - feat(formatting): `black[jupyter]` (#6732) - fix(test): `test_optimizer_chinese` (#6730) ## [v0.7.1] - 2023-07-05 ### New Features - Streaming support for OpenAI agents (#6694) - add recursive retriever + notebook example (#6682) ## [v0.7.0] - 2023-07-04 ### New Features - Index creation progress bars (#6583) ### Bug Fixes/ Nits - Improved chat refine template (#6645) ### Breaking/Deprecated API Changes - Change `BaseOpenAIAgent` to use `llama_index.llms.OpenAI`. Adjust `chat_history` to use `List[ChatMessage]]` as type. - Remove (previously deprecated) `llama_index.langchain_helpers.chain_wrapper` module. - Remove (previously deprecated) `llama_index.token_counter.token_counter` module. See [migration guide](/how_to/callbacks/token_counting_migration.html) for more details on new callback based token counting. - Remove `ChatGPTLLMPredictor` and `HuggingFaceLLMPredictor`. See [migration guide](/how_to/customization/llms_migration_guide.html) for more details on replacements. - Remove support for setting `cache` via `LLMPredictor` constructor. - Update `BaseChatEngine` interface: - adjust `chat_history` to use `List[ChatMessage]]` as type - expose `chat_history` state as a property - support overriding `chat_history` in `chat` and `achat` endpoints - Remove deprecated arguments for `PromptHelper`: `max_input_size`, `embedding_limit`, `max_chunk_overlap` - Update all notebooks to use native openai integration (#6696) ## [v0.6.38] - 2023-07-02 ### New Features - add optional tqdm progress during index creation (#6583) - Added async support for "compact" and "refine" response modes (#6590) - [feature]add transformer tokenize functionalities for optimizer (chinese) (#6659) - Add simple benchmark for vector store (#6670) - Introduce `llama_index.llms` module, with new `LLM` interface, and `OpenAI`, `HuggingFaceLLM`, `LangChainLLM` implementations. (#6615) - Evaporate pydantic program (#6666) ### Bug Fixes / Nits - Improve metadata/node storage and retrieval for RedisVectorStore (#6678) - Fixed node vs. document filtering in vector stores (#6677) - add context retrieval agent notebook link to docs (#6660) - Allow null values for the 'image' property in the ImageNode class and se… (#6661) - Fix broken links in docs (#6669) - update milvus to store node content (#6667) ## [v0.6.37] - 2023-06-30 ### New Features - add context augmented openai agent (#6655) ## [v0.6.36] - 2023-06-29 ### New Features - Redis support for index stores and docstores (#6575) - DuckDB + SQL query engine notebook (#6628) - add notebook showcasing deplot data loader (#6638) ### Bug Fixes / Nits - More robust JSON parsing from LLM for `SelectionOutputParser` (#6610) - bring our loaders back in line with llama-hub (#6630) - Remove usage of SQLStructStoreIndex in notebooks (#6585) - MD reader: remove html tags and leave linebreaks alone (#6618) - bump min langchain version to latest version (#6632) - Fix metadata column name in postgres vector store (#6622) - Postgres metadata fixes (#6626, #6634) - fixed links to dataloaders in contribution.md (#6636) - fix: typo in docs in creating custom_llm huggingface example (#6639) - Updated SelectionOutputParser to handle JSON objects and arrays (#6610) - Fixed docstring argument typo (#6652) ## [v0.6.35] - 2023-06-28 - refactor structured output + pydantic programs (#6604) ### Bug Fixes / Nits - Fix serialization for OpenSearch vector stores (#6612) - patch docs relationships (#6606) - Bug fix for ignoring directories while parsing git repo (#4196) - updated Chroma notebook (#6572) - Backport old node name (#6614) - Add the ability to change chroma implementation (#6601) ## [v0.6.34] - 2023-06-26 ### Patch Update (v0.6.34.post1) - Patch imports for Document obj for backwards compatibility (#6597) ### New Features - New `TextNode`/`Document` object classes based on pydantic (#6586) - `TextNode`/`Document` objects support metadata customization (metadata templates, exclude metadata from LLM or embeddings) (#6586) - Nodes no longer require flat metadata dictionaries, unless the vector store you use requires it (#6586) ### Bug Fixes / Nits - use `NLTK_DATA` env var to control NLTK download location (#6579) - [discord] save author as metadata in group_conversations.py (#6592) - bs4 -> beautifulsoup4 in requirements (#6582) - negate euclidean distance (#6564) - add df output parser notebook link to docs (#6581) ### Breaking/Deprecated API Changes - `Node` has been renamed to `TextNode` and is imported from `llama_index.schema` (#6586) - `TextNode` and `Document` must be instantiated with kwargs: `Document(text=text)` (#6586) - `TextNode` (fka `Node`) has a `id_` or `node_id` property, rather than `doc_id` (#6586) - `TextNode` and `Document` have a metadata property, which replaces the extra_info property (#6586) - `TextNode` no longer has a `node_info` property (start/end indexes are accessed directly with `start/end_char_idx` attributes) (#6586) ## [v0.6.33] - 2023-06-25 ### New Features - Add typesense vector store (#6561) - add df output parser (#6576) ### Bug Fixes / Nits - Track langchain dependency via bridge module. (#6573) ## [v0.6.32] - 2023-06-23 ### New Features - add object index (#6548) - add SQL Schema Node Mapping + SQLTableRetrieverQueryEngine + obj index fixes (#6569) - sql refactor (NLSQLTableQueryEngine) (#6529) ### Bug Fixes / Nits - Update vector_stores.md (#6562) - Minor `BaseResponseBuilder` interface cleanup (#6557) - Refactor TreeSummarize (#6550) ## [v0.6.31] - 2023-06-22 ### Bug Fixes / Nits - properly convert weaviate distance to score (#6545) - refactor tree summarize and fix bug to not truncate context (#6550) - fix custom KG retrieval notebook nits (#6551) ## [v0.6.30] - 2023-06-21 ### New Features - multi-selector support in router query engine (#6518) - pydantic selector support in router query engine using OpenAI function calling API (#6518) - streaming response support in `CondenseQuestionChatEngine` and `SimpleChatEngine` (#6524) - metadata filtering support in `QdrantVectorStore` (#6476) - add `PGVectorStore` to support postgres with pgvector (#6190) ### Bug Fixes / Nits - better error handling in the mbox reader (#6248) - Fix blank similarity score when using weaviate (#6512) - fix for sorted nodes in `PrevNextNodePostprocessor` (#6048) ### Breaking/Deprecated API Changes - Refactor PandasQueryEngine to take in df directly, deprecate PandasIndex (#6527) ## [v0.6.29] - 2023-06-20 ### New Features - query planning tool with OpenAI Function API (#6520) - docs: example of kg+vector index (#6497) - Set context window sizes for Cohere and AI21(J2 model) (#6485) ### Bug Fixes / Nits - add default input size for Cohere and AI21 (#6485) - docs: replace comma with colon in dict object (#6439) - extra space in prompt and error message update (#6443) - [Issue 6417] Fix prompt_templates docs page (#6499) - Rip out monkey patch and update model to context window mapping (#6490) ## [v0.6.28] - 2023-06-19 ### New Features - New OpenAI Agent + Query Engine Cookbook (#6496) - allow recursive data extraction (pydantic program) (#6503) ### Bug Fixes / Nits - update mongo interface (#6501) - fixes that we forgot to include for openai pydantic program (#6503) (#6504) - Fix github pics in Airbyte notebook (#6493) ## [v0.6.27] - 2023-06-16 ### New Features - Add node doc_id filtering to weaviate (#6467) - New `TokenCountingCallback` to customize and track embedding, prompt, and completion token usage (#6440) - OpenAI Retrieval Function Agent (#6491) ### Breaking/Deprecated API Changes - Deprecated current token tracking (llm predictor and embed model will no longer track tokens in the future, please use the `TokenCountingCallback` (#6440) - Add maximal marginal relevance to the Simple Vector Store, which can be enabled as a query mode (#6446) ### Bug Fixes / Nits - `as_chat_engine` properly inherits the current service context (#6470) - Use namespace when deleting from pinecone (#6475) - Fix paths when using fsspec on windows (#3778) - Fix for using custom file readers in `SimpleDirectoryReader` (#6477) - Edit MMR Notebook (#6486) - FLARE fixes (#6484) ## [v0.6.26] - 2023-06-14 ### New Features - Add OpenAIAgent and tutorial notebook for "build your own agent" (#6461) - Add OpenAIPydanticProgram (#6462) ### Bug Fixes / Nits - Fix citation engine import (#6456) ## [v0.6.25] - 2023-06-13 ### New Features - Added FLARE query engine (#6419). ## [v0.6.24] - 2023-06-12 ### New Features - Added better support for vector store with existing data (e.g. allow configurable text key) for Pinecone and Weaviate. (#6393) - Support batched upsert for Pineone (#6393) - Added initial [guidance](https://github.com/microsoft/guidance/) integration. Added `GuidancePydanticProgram` for generic structured output generation and `GuidanceQuestionGenerator` for generating sub-questions in `SubQuestionQueryEngine` (#6246). ## [v0.6.23] - 2023-06-11 ### Bug Fixes / Nits - Remove hardcoded chunk size for citation query engine (#6408) - Mongo demo improvements (#6406) - Fix notebook (#6418) - Cleanup RetryQuery notebook (#6381) ## [v0.6.22] - 2023-06-10 ### New Features - Added `SQLJoinQueryEngine` (generalization of `SQLAutoVectorQueryEngine`) (#6265) - Added support for graph stores under the hood, and initial support for Nebula KG. More docs coming soon! (#2581) - Added guideline evaluator to allow llm to provide feedback based on user guidelines (#4664) - Added support for MongoDB Vector stores to enable Atlas knnbeta search (#6379) - Added new CitationQueryEngine for inline citations of sources in response text (#6239) ### Bug Fixes - Fixed bug with `delete_ref_doc` not removing all metadata from the docstore (#6192) - FIxed bug with loading existing QDrantVectorStore (#6230) ### Miscellaneous - Added changelog officially to github repo (#6191) ## [v0.6.21] - 2023-06-06 ### New Features - SimpleDirectoryReader has new `filename_as_id` flag to automatically set the doc_id (useful for `refresh_ref_docs()`) - DocArray vector store integration - Tair vector store integration - Weights and Biases callback handler for tracing and versioning indexes - Can initialize indexes directly from a vector store: `index = VectorStoreIndex.from_vector_store(vector_store=vector_store)` ### Bug Fixes - Fixed multimodal notebook - Updated/fixed the SQL tutorial in the docs ### Miscellaneous - Minor docs updates - Added github pull-requset templates - Added github issue-forms ## [v0.6.20] - 2023-06-04 ### New Features - Added new JSONQueryEngine that uses JSON schema to deliver more accurate JSON query answers - Metadata support for redis vector-store - Added Supabase vector store integration ### Bug Fixes - Fixed typo in text-to-sql prompt ### Breaking/Deprecated API Changes - Removed GPT prefix from indexes (old imports/names are still supported though) ### Miscellaneous - Major docs updates, brought important modules to the top level ## [v0.6.19] - 2023-06-02 ### New Features - Added agent tool abstraction for llama-hub data loaders ### Miscellaneous - Minor doc updates ## [v0.6.18] - 2023-06-02 ### Miscellaneous - Added `Discover LlamaIndex` video series to the tutorials docs section - Minor docs updates