diff --git a/flowsettings.py b/flowsettings.py index 281fd63f..e4b95b95 100644 --- a/flowsettings.py +++ b/flowsettings.py @@ -184,14 +184,14 @@ }, "default": False, } -KH_LLMS["gemini"] = { - "spec": { - "__type__": "kotaemon.llms.chats.LCGeminiChat", - "model_name": "gemini-1.5-pro", - "api_key": "your-key", - }, - "default": False, -} +# KH_LLMS["gemini"] = { +# "spec": { +# "__type__": "kotaemon.llms.chats.LCGeminiChat", +# "model_name": "gemini-1.5-pro", +# "api_key": "your-key", +# }, +# "default": False, +# } KH_LLMS["groq"] = { "spec": { "__type__": "kotaemon.llms.ChatOpenAI", diff --git a/libs/kotaemon/pyproject.toml b/libs/kotaemon/pyproject.toml index 27170c07..ee28cfd7 100644 --- a/libs/kotaemon/pyproject.toml +++ b/libs/kotaemon/pyproject.toml @@ -31,7 +31,6 @@ dependencies = [ "langchain-community>=0.0.34,<0.1.0", "langchain-openai>=0.1.4,<0.2.0", "langchain-anthropic", - "langchain-google-genai", "llama-hub>=0.0.79,<0.1.0", "llama-index>=0.10.40,<0.11.0", "fastapi<=0.112.1",