Fix: Disable Chroma Telemetry
Opts-out of anonymized telemetry being tracked in Chroma. See: https://docs.trychroma.com/telemetry
This commit is contained in:
parent
60225698b6
commit
8c6a81a07f
|
@ -0,0 +1,11 @@
|
||||||
|
from chromadb.config import Settings
|
||||||
|
|
||||||
|
# Define the folder for storing database
|
||||||
|
PERSIST_DIRECTORY = 'db'
|
||||||
|
|
||||||
|
# Define the Chroma settings
|
||||||
|
CHROMA_SETTINGS = Settings(
|
||||||
|
chroma_db_impl='duckdb+parquet',
|
||||||
|
persist_directory=PERSIST_DIRECTORY,
|
||||||
|
anonymized_telemetry=False
|
||||||
|
)
|
|
@ -3,6 +3,8 @@ from langchain.text_splitter import RecursiveCharacterTextSplitter
|
||||||
from langchain.vectorstores import Chroma
|
from langchain.vectorstores import Chroma
|
||||||
from langchain.embeddings import LlamaCppEmbeddings
|
from langchain.embeddings import LlamaCppEmbeddings
|
||||||
from sys import argv
|
from sys import argv
|
||||||
|
from chroma_preference import PERSIST_DIRECTORY
|
||||||
|
from chroma_preference import CHROMA_SETTINGS
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
# Load document and split in chunks
|
# Load document and split in chunks
|
||||||
|
@ -13,8 +15,7 @@ def main():
|
||||||
# Create embeddings
|
# Create embeddings
|
||||||
llama = LlamaCppEmbeddings(model_path="./models/ggml-model-q4_0.bin")
|
llama = LlamaCppEmbeddings(model_path="./models/ggml-model-q4_0.bin")
|
||||||
# Create and store locally vectorstore
|
# Create and store locally vectorstore
|
||||||
persist_directory = 'db'
|
db = Chroma.from_documents(texts, llama, persist_directory=PERSIST_DIRECTORY, client_settings=CHROMA_SETTINGS)
|
||||||
db = Chroma.from_documents(texts, llama, persist_directory=persist_directory)
|
|
||||||
db.persist()
|
db.persist()
|
||||||
db = None
|
db = None
|
||||||
|
|
||||||
|
|
|
@ -3,12 +3,13 @@ from langchain.embeddings import LlamaCppEmbeddings
|
||||||
from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
|
from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
|
||||||
from langchain.vectorstores import Chroma
|
from langchain.vectorstores import Chroma
|
||||||
from langchain.llms import GPT4All
|
from langchain.llms import GPT4All
|
||||||
|
from chroma_preference import PERSIST_DIRECTORY
|
||||||
|
from chroma_preference import CHROMA_SETTINGS
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
# Load stored vectorstore
|
# Load stored vectorstore
|
||||||
llama = LlamaCppEmbeddings(model_path="./models/ggml-model-q4_0.bin")
|
llama = LlamaCppEmbeddings(model_path="./models/ggml-model-q4_0.bin")
|
||||||
persist_directory = 'db'
|
db = Chroma(persist_directory=PERSIST_DIRECTORY, embedding_function=llama, client_settings=CHROMA_SETTINGS)
|
||||||
db = Chroma(persist_directory=persist_directory, embedding_function=llama)
|
|
||||||
retriever = db.as_retriever()
|
retriever = db.as_retriever()
|
||||||
# Prepare the LLM
|
# Prepare the LLM
|
||||||
callbacks = [StreamingStdOutCallbackHandler()]
|
callbacks = [StreamingStdOutCallbackHandler()]
|
||||||
|
|
Loading…
Reference in New Issue