Fix: Disable Chroma Telemetry
Opts-out of anonymized telemetry being tracked in Chroma. See: https://docs.trychroma.com/telemetry
This commit is contained in:
		
							parent
							
								
									60225698b6
								
							
						
					
					
						commit
						8c6a81a07f
					
				|  | @ -0,0 +1,11 @@ | ||||||
|  | from chromadb.config import Settings | ||||||
|  | 
 | ||||||
|  | # Define the folder for storing database | ||||||
|  | PERSIST_DIRECTORY = 'db' | ||||||
|  | 
 | ||||||
|  | # Define the Chroma settings | ||||||
|  | CHROMA_SETTINGS = Settings( | ||||||
|  |         chroma_db_impl='duckdb+parquet', | ||||||
|  |         persist_directory=PERSIST_DIRECTORY, | ||||||
|  |         anonymized_telemetry=False | ||||||
|  | ) | ||||||
|  | @ -3,6 +3,8 @@ from langchain.text_splitter import RecursiveCharacterTextSplitter | ||||||
| from langchain.vectorstores import Chroma | from langchain.vectorstores import Chroma | ||||||
| from langchain.embeddings import LlamaCppEmbeddings | from langchain.embeddings import LlamaCppEmbeddings | ||||||
| from sys import argv | from sys import argv | ||||||
|  | from chroma_preference import PERSIST_DIRECTORY | ||||||
|  | from chroma_preference import CHROMA_SETTINGS | ||||||
| 
 | 
 | ||||||
| def main(): | def main(): | ||||||
|     # Load document and split in chunks |     # Load document and split in chunks | ||||||
|  | @ -13,8 +15,7 @@ def main(): | ||||||
|     # Create embeddings |     # Create embeddings | ||||||
|     llama = LlamaCppEmbeddings(model_path="./models/ggml-model-q4_0.bin") |     llama = LlamaCppEmbeddings(model_path="./models/ggml-model-q4_0.bin") | ||||||
|     # Create and store locally vectorstore |     # Create and store locally vectorstore | ||||||
|     persist_directory = 'db' |     db = Chroma.from_documents(texts, llama, persist_directory=PERSIST_DIRECTORY, client_settings=CHROMA_SETTINGS) | ||||||
|     db = Chroma.from_documents(texts, llama, persist_directory=persist_directory) |  | ||||||
|     db.persist() |     db.persist() | ||||||
|     db = None |     db = None | ||||||
| 
 | 
 | ||||||
|  |  | ||||||
|  | @ -3,12 +3,13 @@ from langchain.embeddings import LlamaCppEmbeddings | ||||||
| from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler | from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler | ||||||
| from langchain.vectorstores import Chroma | from langchain.vectorstores import Chroma | ||||||
| from langchain.llms import GPT4All | from langchain.llms import GPT4All | ||||||
|  | from chroma_preference import PERSIST_DIRECTORY | ||||||
|  | from chroma_preference import CHROMA_SETTINGS | ||||||
| 
 | 
 | ||||||
| def main():         | def main():         | ||||||
|     # Load stored vectorstore |     # Load stored vectorstore | ||||||
|     llama = LlamaCppEmbeddings(model_path="./models/ggml-model-q4_0.bin") |     llama = LlamaCppEmbeddings(model_path="./models/ggml-model-q4_0.bin") | ||||||
|     persist_directory = 'db' |     db = Chroma(persist_directory=PERSIST_DIRECTORY, embedding_function=llama, client_settings=CHROMA_SETTINGS) | ||||||
|     db = Chroma(persist_directory=persist_directory, embedding_function=llama) |  | ||||||
|     retriever = db.as_retriever() |     retriever = db.as_retriever() | ||||||
|     # Prepare the LLM |     # Prepare the LLM | ||||||
|     callbacks = [StreamingStdOutCallbackHandler()] |     callbacks = [StreamingStdOutCallbackHandler()] | ||||||
|  |  | ||||||
		Loading…
	
		Reference in New Issue