From 5761529c7f6dac842a4ecd33246ebfdcbcf4a428 Mon Sep 17 00:00:00 2001 From: latte Date: Sun, 11 Jan 2026 20:23:03 +0100 Subject: [PATCH] Add Google Gemini as AI provider - Add gemini.py provider using google-genai SDK - Update config.py with gemini provider and GEMINI_API_KEY - Update ai_service.py factory to support gemini - Add google-genai to requirements.txt - Update .env.example, README.md, and CLAUDE.md documentation --- .env.example | 3 +- CLAUDE.md | 5 +- README.md | 6 +- requirements.txt | 1 + src/daemon_boyfriend/config.py | 4 +- src/daemon_boyfriend/services/ai_service.py | 4 +- .../services/providers/__init__.py | 2 + .../services/providers/gemini.py | 71 +++++++++++++++++++ 8 files changed, 89 insertions(+), 7 deletions(-) create mode 100644 src/daemon_boyfriend/services/providers/gemini.py diff --git a/.env.example b/.env.example index 9e99d47..de8b2bb 100644 --- a/.env.example +++ b/.env.example @@ -7,7 +7,7 @@ DISCORD_TOKEN=your_discord_bot_token_here # =========================================== # AI Provider Configuration # =========================================== -# Available providers: "openai", "openrouter", "anthropic" +# Available providers: "openai", "openrouter", "anthropic", "gemini" AI_PROVIDER=openai # Model to use (e.g., gpt-4o, gpt-4o-mini, claude-3-5-sonnet, etc.) @@ -17,6 +17,7 @@ AI_MODEL=gpt-4o OPENAI_API_KEY=sk-xxx OPENROUTER_API_KEY=sk-or-xxx ANTHROPIC_API_KEY=sk-ant-xxx +GEMINI_API_KEY=xxx # Maximum tokens in AI response (100-4096) AI_MAX_TOKENS=1024 diff --git a/CLAUDE.md b/CLAUDE.md index 0bed859..be9931f 100644 --- a/CLAUDE.md +++ b/CLAUDE.md @@ -25,9 +25,10 @@ This is a Discord bot that responds to @mentions with AI-generated responses (mu ### Provider Pattern The AI system uses a provider abstraction pattern: - `services/providers/base.py` defines `AIProvider` abstract class with `generate()` method -- `services/providers/openai.py`, `openrouter.py`, `anthropic.py` implement the interface +- `services/providers/openai.py`, `openrouter.py`, `anthropic.py`, `gemini.py` implement the interface - `services/ai_service.py` is the factory that creates the correct provider based on `AI_PROVIDER` env var - OpenRouter uses OpenAI's client with a different base URL +- Gemini uses the `google-genai` SDK ### Cog System Discord functionality is in `cogs/`: @@ -45,4 +46,4 @@ All config flows through `config.py` using pydantic-settings. The `settings` sin ## Environment Variables -Required: `DISCORD_TOKEN`, plus one of `OPENAI_API_KEY`, `OPENROUTER_API_KEY`, or `ANTHROPIC_API_KEY` depending on `AI_PROVIDER` setting. +Required: `DISCORD_TOKEN`, plus one of `OPENAI_API_KEY`, `OPENROUTER_API_KEY`, `ANTHROPIC_API_KEY`, or `GEMINI_API_KEY` depending on `AI_PROVIDER` setting. diff --git a/README.md b/README.md index b5181b1..1b05460 100644 --- a/README.md +++ b/README.md @@ -4,7 +4,7 @@ A customizable Discord bot that responds to @mentions with AI-generated response ## Features -- **Multi-Provider AI**: Supports OpenAI, OpenRouter, and Anthropic (Claude) +- **Multi-Provider AI**: Supports OpenAI, OpenRouter, Anthropic (Claude), and Google Gemini - **Fully Customizable**: Configure bot name, personality, and behavior - **Conversation Memory**: Remembers context per user - **Easy Deployment**: Docker support included @@ -50,10 +50,11 @@ All configuration is done via environment variables in `.env`. | Variable | Description | |----------|-------------| | `DISCORD_TOKEN` | Your Discord bot token | -| `AI_PROVIDER` | `openai`, `openrouter`, or `anthropic` | +| `AI_PROVIDER` | `openai`, `openrouter`, `anthropic`, or `gemini` | | `OPENAI_API_KEY` | OpenAI API key (if using OpenAI) | | `OPENROUTER_API_KEY` | OpenRouter API key (if using OpenRouter) | | `ANTHROPIC_API_KEY` | Anthropic API key (if using Anthropic) | +| `GEMINI_API_KEY` | Google Gemini API key (if using Gemini) | ### Bot Identity @@ -129,6 +130,7 @@ Mention the bot in any channel: | OpenAI | gpt-4o, gpt-4-turbo, gpt-3.5-turbo | Official OpenAI API | | OpenRouter | 100+ models | Access to Llama, Mistral, Claude, etc. | | Anthropic | claude-3-5-sonnet, claude-3-opus | Direct Claude API | +| Gemini | gemini-2.0-flash, gemini-1.5-pro | Google AI API | ## Project Structure diff --git a/requirements.txt b/requirements.txt index abbfc3a..22516f5 100644 --- a/requirements.txt +++ b/requirements.txt @@ -3,6 +3,7 @@ discord.py>=2.3.0 # AI Providers anthropic>=0.18.0 +google-genai>=1.0.0 openai>=1.12.0 # HTTP Client diff --git a/src/daemon_boyfriend/config.py b/src/daemon_boyfriend/config.py index b1f064c..0c19893 100644 --- a/src/daemon_boyfriend/config.py +++ b/src/daemon_boyfriend/config.py @@ -19,7 +19,7 @@ class Settings(BaseSettings): discord_token: str = Field(..., description="Discord bot token") # AI Provider Configuration - ai_provider: Literal["openai", "openrouter", "anthropic"] = Field( + ai_provider: Literal["openai", "openrouter", "anthropic", "gemini"] = Field( "openai", description="Which AI provider to use" ) ai_model: str = Field("gpt-4o", description="AI model to use") @@ -30,6 +30,7 @@ class Settings(BaseSettings): openai_api_key: str | None = Field(None, description="OpenAI API key") openrouter_api_key: str | None = Field(None, description="OpenRouter API key") anthropic_api_key: str | None = Field(None, description="Anthropic API key") + gemini_api_key: str | None = Field(None, description="Google Gemini API key") # Logging log_level: str = Field("INFO", description="Logging level") @@ -66,6 +67,7 @@ class Settings(BaseSettings): "openai": self.openai_api_key, "openrouter": self.openrouter_api_key, "anthropic": self.anthropic_api_key, + "gemini": self.gemini_api_key, } key = key_map.get(self.ai_provider) if not key: diff --git a/src/daemon_boyfriend/services/ai_service.py b/src/daemon_boyfriend/services/ai_service.py index fce3193..775ca28 100644 --- a/src/daemon_boyfriend/services/ai_service.py +++ b/src/daemon_boyfriend/services/ai_service.py @@ -9,6 +9,7 @@ from .providers import ( AIProvider, AIResponse, AnthropicProvider, + GeminiProvider, Message, OpenAIProvider, OpenRouterProvider, @@ -16,7 +17,7 @@ from .providers import ( logger = logging.getLogger(__name__) -ProviderType = Literal["openai", "openrouter", "anthropic"] +ProviderType = Literal["openai", "openrouter", "anthropic", "gemini"] class AIService: @@ -45,6 +46,7 @@ class AIService: "openai": OpenAIProvider, "openrouter": OpenRouterProvider, "anthropic": AnthropicProvider, + "gemini": GeminiProvider, } provider_class = providers.get(provider_type) diff --git a/src/daemon_boyfriend/services/providers/__init__.py b/src/daemon_boyfriend/services/providers/__init__.py index 71e99b5..b5e4b68 100644 --- a/src/daemon_boyfriend/services/providers/__init__.py +++ b/src/daemon_boyfriend/services/providers/__init__.py @@ -2,6 +2,7 @@ from .anthropic import AnthropicProvider from .base import AIProvider, AIResponse, Message +from .gemini import GeminiProvider from .openai import OpenAIProvider from .openrouter import OpenRouterProvider @@ -12,4 +13,5 @@ __all__ = [ "OpenAIProvider", "OpenRouterProvider", "AnthropicProvider", + "GeminiProvider", ] diff --git a/src/daemon_boyfriend/services/providers/gemini.py b/src/daemon_boyfriend/services/providers/gemini.py new file mode 100644 index 0000000..8ef1350 --- /dev/null +++ b/src/daemon_boyfriend/services/providers/gemini.py @@ -0,0 +1,71 @@ +"""Google Gemini provider implementation.""" + +import logging + +from google import genai +from google.genai import types + +from .base import AIProvider, AIResponse, Message + +logger = logging.getLogger(__name__) + + +class GeminiProvider(AIProvider): + """Google Gemini API provider.""" + + def __init__(self, api_key: str, model: str = "gemini-2.0-flash") -> None: + self.client = genai.Client(api_key=api_key) + self.model = model + + @property + def provider_name(self) -> str: + return "gemini" + + async def generate( + self, + messages: list[Message], + system_prompt: str | None = None, + max_tokens: int = 1024, + temperature: float = 0.7, + ) -> AIResponse: + """Generate a response using Gemini.""" + # Build contents list (Gemini format) + contents = [] + for m in messages: + # Gemini uses "user" and "model" roles + role = "model" if m.role == "assistant" else m.role + contents.append(types.Content(role=role, parts=[types.Part(text=m.content)])) + + logger.debug(f"Sending {len(contents)} messages to Gemini") + + # Build config + config = types.GenerateContentConfig( + max_output_tokens=max_tokens, + temperature=temperature, + ) + if system_prompt: + config.system_instruction = system_prompt + + response = await self.client.aio.models.generate_content( + model=self.model, + contents=contents, + config=config, + ) + + # Extract text from response + content = response.text or "" + + # Build usage dict + usage = {} + if response.usage_metadata: + usage = { + "prompt_tokens": response.usage_metadata.prompt_token_count or 0, + "completion_tokens": response.usage_metadata.candidates_token_count or 0, + "total_tokens": response.usage_metadata.total_token_count or 0, + } + + return AIResponse( + content=content, + model=self.model, + usage=usage, + )