feat: bump dependencies (#1987)
This commit is contained in:
		
							parent
							
								
									c7212ac7cc
								
							
						
					
					
						commit
						b687dc8524
					
				
										
											
												File diff suppressed because it is too large
												Load Diff
											
										
									
								
							|  | @ -4,10 +4,10 @@ import typing | ||||||
| from injector import inject, singleton | from injector import inject, singleton | ||||||
| from llama_index.core.indices.vector_store import VectorIndexRetriever, VectorStoreIndex | from llama_index.core.indices.vector_store import VectorIndexRetriever, VectorStoreIndex | ||||||
| from llama_index.core.vector_stores.types import ( | from llama_index.core.vector_stores.types import ( | ||||||
|  |     BasePydanticVectorStore, | ||||||
|     FilterCondition, |     FilterCondition, | ||||||
|     MetadataFilter, |     MetadataFilter, | ||||||
|     MetadataFilters, |     MetadataFilters, | ||||||
|     VectorStore, |  | ||||||
| ) | ) | ||||||
| 
 | 
 | ||||||
| from private_gpt.open_ai.extensions.context_filter import ContextFilter | from private_gpt.open_ai.extensions.context_filter import ContextFilter | ||||||
|  | @ -32,7 +32,7 @@ def _doc_id_metadata_filter( | ||||||
| @singleton | @singleton | ||||||
| class VectorStoreComponent: | class VectorStoreComponent: | ||||||
|     settings: Settings |     settings: Settings | ||||||
|     vector_store: VectorStore |     vector_store: BasePydanticVectorStore | ||||||
| 
 | 
 | ||||||
|     @inject |     @inject | ||||||
|     def __init__(self, settings: Settings) -> None: |     def __init__(self, settings: Settings) -> None: | ||||||
|  | @ -54,7 +54,7 @@ class VectorStoreComponent: | ||||||
|                     ) |                     ) | ||||||
| 
 | 
 | ||||||
|                 self.vector_store = typing.cast( |                 self.vector_store = typing.cast( | ||||||
|                     VectorStore, |                     BasePydanticVectorStore, | ||||||
|                     PGVectorStore.from_params( |                     PGVectorStore.from_params( | ||||||
|                         **settings.postgres.model_dump(exclude_none=True), |                         **settings.postgres.model_dump(exclude_none=True), | ||||||
|                         table_name="embeddings", |                         table_name="embeddings", | ||||||
|  | @ -87,7 +87,7 @@ class VectorStoreComponent: | ||||||
|                 )  # TODO |                 )  # TODO | ||||||
| 
 | 
 | ||||||
|                 self.vector_store = typing.cast( |                 self.vector_store = typing.cast( | ||||||
|                     VectorStore, |                     BasePydanticVectorStore, | ||||||
|                     BatchedChromaVectorStore( |                     BatchedChromaVectorStore( | ||||||
|                         chroma_client=chroma_client, chroma_collection=chroma_collection |                         chroma_client=chroma_client, chroma_collection=chroma_collection | ||||||
|                     ), |                     ), | ||||||
|  | @ -115,7 +115,7 @@ class VectorStoreComponent: | ||||||
|                         **settings.qdrant.model_dump(exclude_none=True) |                         **settings.qdrant.model_dump(exclude_none=True) | ||||||
|                     ) |                     ) | ||||||
|                 self.vector_store = typing.cast( |                 self.vector_store = typing.cast( | ||||||
|                     VectorStore, |                     BasePydanticVectorStore, | ||||||
|                     QdrantVectorStore( |                     QdrantVectorStore( | ||||||
|                         client=client, |                         client=client, | ||||||
|                         collection_name="make_this_parameterizable_per_api_call", |                         collection_name="make_this_parameterizable_per_api_call", | ||||||
|  |  | ||||||
|  | @ -7,45 +7,48 @@ authors = ["Zylon <hi@zylon.ai>"] | ||||||
| [tool.poetry.dependencies] | [tool.poetry.dependencies] | ||||||
| python = ">=3.11,<3.12" | python = ">=3.11,<3.12" | ||||||
| # PrivateGPT | # PrivateGPT | ||||||
| fastapi = { extras = ["all"], version = "^0.110.0" } | fastapi = { extras = ["all"], version = "^0.111.0" } | ||||||
| python-multipart = "^0.0.9" | python-multipart = "^0.0.9" | ||||||
| injector = "^0.21.0" | injector = "^0.21.0" | ||||||
| pyyaml = "^6.0.1" | pyyaml = "^6.0.1" | ||||||
| watchdog = "^4.0.0" | watchdog = "^4.0.1" | ||||||
| transformers = "^4.38.2" | transformers = "^4.42.3" | ||||||
| docx2txt = "^0.8" | docx2txt = "^0.8" | ||||||
| cryptography = "^3.1" | cryptography = "^3.1" | ||||||
| # LlamaIndex core libs | # LlamaIndex core libs | ||||||
| llama-index-core = "^0.10.14" | llama-index-core = "^0.10.52" | ||||||
| llama-index-readers-file = "^0.1.6" | llama-index-readers-file = "^0.1.27" | ||||||
| # Optional LlamaIndex integration libs | # Optional LlamaIndex integration libs | ||||||
| llama-index-llms-llama-cpp = {version = "^0.1.3", optional = true} | llama-index-llms-llama-cpp = {version = "^0.1.4", optional = true} | ||||||
| llama-index-llms-openai = {version = "^0.1.6", optional = true} | llama-index-llms-openai = {version = "^0.1.25", optional = true} | ||||||
| llama-index-llms-openai-like = {version ="^0.1.3", optional = true} | llama-index-llms-openai-like = {version ="^0.1.3", optional = true} | ||||||
| llama-index-llms-ollama = {version ="^0.1.2", optional = true} | llama-index-llms-ollama = {version ="^0.1.5", optional = true} | ||||||
| llama-index-llms-azure-openai = {version ="^0.1.5", optional = true} | llama-index-llms-azure-openai = {version ="^0.1.8", optional = true} | ||||||
| llama-index-embeddings-ollama = {version ="^0.1.2", optional = true} | llama-index-embeddings-ollama = {version ="^0.1.2", optional = true} | ||||||
| llama-index-embeddings-huggingface = {version ="^0.1.4", optional = true} | llama-index-embeddings-huggingface = {version ="^0.2.2", optional = true} | ||||||
| llama-index-embeddings-openai = {version ="^0.1.6", optional = true} | llama-index-embeddings-openai = {version ="^0.1.10", optional = true} | ||||||
| llama-index-embeddings-azure-openai = {version ="^0.1.6", optional = true} | llama-index-embeddings-azure-openai = {version ="^0.1.10", optional = true} | ||||||
| llama-index-vector-stores-qdrant = {version ="^0.1.3", optional = true} | llama-index-vector-stores-qdrant = {version ="^0.2.10", optional = true} | ||||||
| llama-index-vector-stores-chroma = {version ="^0.1.4", optional = true} | llama-index-vector-stores-chroma = {version ="^0.1.10", optional = true} | ||||||
| llama-index-vector-stores-postgres = {version ="^0.1.2", optional = true} | llama-index-vector-stores-postgres = {version ="^0.1.11", optional = true} | ||||||
| llama-index-storage-docstore-postgres = {version ="^0.1.2", optional = true} | llama-index-storage-docstore-postgres = {version ="^0.1.3", optional = true} | ||||||
| llama-index-storage-index-store-postgres = {version ="^0.1.2", optional = true} | llama-index-storage-index-store-postgres = {version ="^0.1.4", optional = true} | ||||||
| # Postgres | # Postgres | ||||||
| psycopg2-binary = {version ="^2.9.9", optional = true} | psycopg2-binary = {version ="^2.9.9", optional = true} | ||||||
| asyncpg = {version="^0.29.0", optional = true} | asyncpg = {version="^0.29.0", optional = true} | ||||||
| 
 | 
 | ||||||
| # Optional Sagemaker dependency | # Optional Sagemaker dependency | ||||||
| boto3 = {version ="^1.34.51", optional = true} | boto3 = {version ="^1.34.139", optional = true} | ||||||
|  | 
 | ||||||
|  | # Optional Qdrant client | ||||||
|  | qdrant-client = {version ="^1.9.0", optional = true} | ||||||
| 
 | 
 | ||||||
| # Optional Reranker dependencies | # Optional Reranker dependencies | ||||||
| torch = {version ="^2.1.2", optional = true} | torch = {version ="^2.3.1", optional = true} | ||||||
| sentence-transformers = {version ="^2.6.1", optional = true} | sentence-transformers = {version ="^3.0.1", optional = true} | ||||||
| 
 | 
 | ||||||
| # Optional UI | # Optional UI | ||||||
| gradio = {version ="^4.19.2", optional = true} | gradio = {version ="^4.37.2", optional = true} | ||||||
| 
 | 
 | ||||||
| [tool.poetry.extras] | [tool.poetry.extras] | ||||||
| ui = ["gradio"] | ui = ["gradio"] | ||||||
|  |  | ||||||
		Loading…
	
		Reference in New Issue