From a02597ed59a04a7a839fc49f0ce5f3c043ef073d Mon Sep 17 00:00:00 2001 From: Dev Khant Date: Mon, 18 Nov 2024 16:25:23 +0530 Subject: [PATCH] Update embedder docs to show openai key is used for LLM (#2033) --- docs/components/embedders/models/azure_openai.mdx | 2 ++ docs/components/embedders/models/gemini.mdx | 1 + docs/components/embedders/models/huggingface.mdx | 2 +- docs/components/embedders/models/ollama.mdx | 2 +- docs/components/embedders/models/together.mdx | 1 + docs/components/embedders/models/vertexai.mdx | 1 + 6 files changed, 7 insertions(+), 2 deletions(-) diff --git a/docs/components/embedders/models/azure_openai.mdx b/docs/components/embedders/models/azure_openai.mdx index fda2d32e8a..9848562b8c 100644 --- a/docs/components/embedders/models/azure_openai.mdx +++ b/docs/components/embedders/models/azure_openai.mdx @@ -15,6 +15,8 @@ os.environ["EMBEDDING_AZURE_DEPLOYMENT"] = "your-deployment-name" os.environ["EMBEDDING_AZURE_ENDPOINT"] = "your-api-base-url" os.environ["EMBEDDING_AZURE_API_VERSION"] = "version-to-use" +os.environ["OPENAI_API_KEY"] = "your_api_key" # For LLM + config = { "embedder": { diff --git a/docs/components/embedders/models/gemini.mdx b/docs/components/embedders/models/gemini.mdx index 125e3160b9..40e5cad330 100644 --- a/docs/components/embedders/models/gemini.mdx +++ b/docs/components/embedders/models/gemini.mdx @@ -11,6 +11,7 @@ import os from mem0 import Memory os.environ["GOOGLE_API_KEY"] = "key" +os.environ["OPENAI_API_KEY"] = "your_api_key" # For LLM config = { "embedder": { diff --git a/docs/components/embedders/models/huggingface.mdx b/docs/components/embedders/models/huggingface.mdx index 98e9b519ba..4b8d8c7e51 100644 --- a/docs/components/embedders/models/huggingface.mdx +++ b/docs/components/embedders/models/huggingface.mdx @@ -10,7 +10,7 @@ You can use embedding models from Huggingface to run Mem0 locally. import os from mem0 import Memory -os.environ["OPENAI_API_KEY"] = "your_api_key" +os.environ["OPENAI_API_KEY"] = "your_api_key" # For LLM config = { "embedder": { diff --git a/docs/components/embedders/models/ollama.mdx b/docs/components/embedders/models/ollama.mdx index e74722a6ed..263d82276f 100644 --- a/docs/components/embedders/models/ollama.mdx +++ b/docs/components/embedders/models/ollama.mdx @@ -6,7 +6,7 @@ You can use embedding models from Ollama to run Mem0 locally. import os from mem0 import Memory -os.environ["OPENAI_API_KEY"] = "your_api_key" +os.environ["OPENAI_API_KEY"] = "your_api_key" # For LLM config = { "embedder": { diff --git a/docs/components/embedders/models/together.mdx b/docs/components/embedders/models/together.mdx index 6788003941..160e474641 100644 --- a/docs/components/embedders/models/together.mdx +++ b/docs/components/embedders/models/together.mdx @@ -13,6 +13,7 @@ import os from mem0 import Memory os.environ["TOGETHER_API_KEY"] = "your_api_key" +os.environ["OPENAI_API_KEY"] = "your_api_key" # For LLM config = { "embedder": { diff --git a/docs/components/embedders/models/vertexai.mdx b/docs/components/embedders/models/vertexai.mdx index 1fe8b95eef..b30c12911c 100644 --- a/docs/components/embedders/models/vertexai.mdx +++ b/docs/components/embedders/models/vertexai.mdx @@ -10,6 +10,7 @@ from mem0 import Memory # Set the path to your Google Cloud credentials JSON file os.environ["GOOGLE_APPLICATION_CREDENTIALS"] = "/path/to/your/credentials.json" +os.environ["OPENAI_API_KEY"] = "your_api_key" # For LLM config = { "embedder": {