Gemini 1.5 Pro Overview
Gemini Pro oferă cel mai mare context window (1M tokens) și integrare excelentă cu ecosistemul Google Cloud.
1M tokens
Context Window
~600-800ms
TTFT (streaming)
Good
Function Calling
Native
Multimodal
Key Advantages
1M Token Context
Largest context window available - entire conversation histories
Google Ecosystem
Native integration with Google Cloud services
Multimodal
Can process audio directly (future voice features)
Competitive Pricing
Good value for high-volume applications
Voice Configuration
import { GoogleGenerativeAI } from "@google/generative-ai";
const genAI = new GoogleGenerativeAI(process.env.GOOGLE_API_KEY);
const model = genAI.getGenerativeModel({
model: "gemini-1.5-pro",
generationConfig: {
maxOutputTokens: 150, // Keep responses short for voice
temperature: 0.7, // Natural but consistent
}
});
// Streaming for low latency
const result = await model.generateContentStream({
contents: [
{ role: "user", parts: [{ text: userMessage }] }
],
systemInstruction: systemPrompt,
});
for await (const chunk of result.stream) {
const text = chunk.text();
// Send to TTS immediately
sendToTTS(text);
}Google Cloud Integrations
Google Cloud Speech-to-Text
Google Cloud Text-to-Speech
Google Calendar API
Google Workspace
Dialogflow CX
Contact Center AI
When to Use Gemini
Ideal For
- • Very long conversations (hours)
- • Google Cloud infrastructure
- • Multimodal future plans
- • Budget-conscious at scale
Consider Alternatives For
- • Lowest possible latency
- • Complex function calling
- • Non-Google infrastructure
- • Specific persona consistency