Skip to content

Commit

Permalink
Fix: Disable Chroma Telemetry
Browse files Browse the repository at this point in the history
Opts-out of anonymized telemetry being tracked in Chroma.

See: https://docs.trychroma.com/telemetry
  • Loading branch information
R-Y-M-R committed May 11, 2023
1 parent 6022569 commit 8c6a81a
Show file tree
Hide file tree
Showing 3 changed files with 17 additions and 4 deletions.
11 changes: 11 additions & 0 deletions chroma_preference.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
from chromadb.config import Settings

# Define the folder for storing database
PERSIST_DIRECTORY = 'db'

# Define the Chroma settings
CHROMA_SETTINGS = Settings(
chroma_db_impl='duckdb+parquet',
persist_directory=PERSIST_DIRECTORY,
anonymized_telemetry=False
)
5 changes: 3 additions & 2 deletions ingest.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,8 @@
from langchain.vectorstores import Chroma
from langchain.embeddings import LlamaCppEmbeddings
from sys import argv
from chroma_preference import PERSIST_DIRECTORY
from chroma_preference import CHROMA_SETTINGS

def main():
# Load document and split in chunks
Expand All @@ -13,8 +15,7 @@ def main():
# Create embeddings
llama = LlamaCppEmbeddings(model_path="./models/ggml-model-q4_0.bin")
# Create and store locally vectorstore
persist_directory = 'db'
db = Chroma.from_documents(texts, llama, persist_directory=persist_directory)
db = Chroma.from_documents(texts, llama, persist_directory=PERSIST_DIRECTORY, client_settings=CHROMA_SETTINGS)
db.persist()
db = None

Expand Down
5 changes: 3 additions & 2 deletions privateGPT.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,12 +3,13 @@
from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
from langchain.vectorstores import Chroma
from langchain.llms import GPT4All
from chroma_preference import PERSIST_DIRECTORY
from chroma_preference import CHROMA_SETTINGS

def main():
# Load stored vectorstore
llama = LlamaCppEmbeddings(model_path="./models/ggml-model-q4_0.bin")
persist_directory = 'db'
db = Chroma(persist_directory=persist_directory, embedding_function=llama)
db = Chroma(persist_directory=PERSIST_DIRECTORY, embedding_function=llama, client_settings=CHROMA_SETTINGS)
retriever = db.as_retriever()
# Prepare the LLM
callbacks = [StreamingStdOutCallbackHandler()]
Expand Down

0 comments on commit 8c6a81a

Please sign in to comment.