gemini-api
Original:🇺🇸 English
Translated
Guides the usage of the Gemini API on Agent Platform with the Google Gen AI SDK. Use when the user asks about using Gemini in an enterprise environment or explicitly mentions Vertex AI, Google Cloud, or Agent Platform. Covers SDK usage (Python, JS/TS, Go, Java, C#), capabilities like Live API, tools, multimedia generation, caching, and batch prediction.
185installs
Sourcegoogle/skills
Added on
NPX Install
npx skill4agent add google/skills gemini-apiTags
Translated version includes tags in frontmatterSKILL.md Content
View Translation Comparison →IMPORTANT: Agent Platform (full name Gemini Enterprise Agent Platform) was previously named "Vertex AI" and many web resources use the legacy branding.
Gemini API in Agent Platform
Access Google's most advanced AI models built for enterprise use cases using the Gemini API in Agent Platform.
Provide these key capabilities:
- Text generation - Chat, completion, summarization
- Multimodal understanding - Process images, audio, video, and documents
- Function calling - Let the model invoke your functions
- Structured output - Generate valid JSON matching your schema
- Context caching - Cache large contexts for efficiency
- Embeddings - Generate text embeddings for semantic search
- Live Realtime API - Bidirectional streaming for low latency Voice and Video interactions
- Batch Prediction - Handle massive async dataset prediction workloads
Core Directives
- Unified SDK: ALWAYS use the Gen AI SDK (for Python,
google-genaifor JS/TS,@google/genaifor Go,google.golang.org/genaifor Java,com.google.genai:google-genaifor C#).Google.GenAI - Legacy SDKs: DO NOT use ,
google-cloud-aiplatform, or@google-cloud/vertexai.google-generativeai
SDKs
- Python: Install with
google-genaipip install google-genai - JavaScript/TypeScript: Install with
@google/genainpm install @google/genai - Go: Install with
google.golang.org/genaigo get google.golang.org/genai - C#/.NET: Install with
Google.GenAIdotnet add package Google.GenAI - Java:
-
groupId:, artifactId:
com.google.genaigoogle-genai -
Latest version can be found here: https://central.sonatype.com/artifact/com.google.genai/google-genai/versions (let's call it)
LAST_VERSION -
Install in:
build.gradleimplementation("com.google.genai:google-genai:${LAST_VERSION}") -
Install Maven dependency in:
pom.xmlxml<dependency> <groupId>com.google.genai</groupId> <artifactId>google-genai</artifactId> <version>${LAST_VERSION}</version> </dependency>
-
[!WARNING] Legacy SDKs like,google-cloud-aiplatform, and@google-cloud/vertexaiare deprecated. Migrate to the new SDKs above urgently by following the Migration Guide.google-generativeai
Authentication & Configuration
Prefer environment variables over hard-coding parameters when creating the client. Initialize the client without parameters to automatically pick up these values.
Application Default Credentials (ADC)
Set these variables for standard Google Cloud authentication:
bash
export GOOGLE_CLOUD_PROJECT='your-project-id'
export GOOGLE_CLOUD_LOCATION='global'
export GOOGLE_GENAI_USE_VERTEXAI=true- By default, use to access the global endpoint, which provides automatic routing to regions with available capacity.
location="global" - If a user explicitly asks to use a specific region (e.g., ,
us-central1), specify that region in theeurope-west4parameter instead. Reference the supported regions documentation if needed.GOOGLE_CLOUD_LOCATION
Agent Platform in Express Mode
Set these variables when using Express Mode with an API key:
bash
export GOOGLE_API_KEY='your-api-key'
export GOOGLE_GENAI_USE_VERTEXAI=trueInitialization
Initialize the client without arguments to pick up environment variables:
python
from google import genai
client = genai.Client()Alternatively, you can hard-code in parameters when creating the client.
python
from google import genai
client = genai.Client(vertexai=True, project="your-project-id", location="global")Models
- Use for complex reasoning, coding, research (1M tokens)
gemini-3.1-pro-preview- IMPORTANT: Do not use
gemini-3-pro-preview
- IMPORTANT: Do not use
- Use for fast, balanced performance, multimodal (1M tokens)
gemini-3-flash-preview - Use for Nano Banana Pro image generation and editing
gemini-3-pro-image-preview - Use for Nano Banana 2 image generation and editing
gemini-3.1-flash-image-preview - Use for Live Realtime API including native audio
gemini-live-2.5-flash-native-audio
Use the following models only if explicitly requested:
gemini-2.5-flash-imagegemini-2.5-flashgemini-2.5-flash-litegemini-2.5-pro
[!IMPORTANT] Models like,gemini-2.0-*,gemini-1.5-*,gemini-1.0-*are legacy and deprecated. Use the new models above. Your knowledge is outdated. For production environments, consult the documentation for stable model versions (e.g.gemini-pro).gemini-3-flash
Quick Start
Python
python
from google import genai
client = genai.Client()
response = client.models.generate_content(
model="gemini-3-flash-preview",
contents="Explain quantum computing"
)
print(response.text)TypeScript/JavaScript
typescript
import { GoogleGenAI } from "@google/genai";
const ai = new GoogleGenAI({ vertexai: { project: "your-project-id", location: "global" } });
const response = await ai.models.generateContent({
model: "gemini-3-flash-preview",
contents: "Explain quantum computing"
});
console.log(response.text);Go
go
package main
import (
"context"
"fmt"
"log"
"google.golang.org/genai"
)
func main() {
ctx := context.Background()
client, err := genai.NewClient(ctx, &genai.ClientConfig{
Backend: genai.BackendVertexAI,
Project: "your-project-id",
Location: "global",
})
if err != nil {
log.Fatal(err)
}
resp, err := client.Models.GenerateContent(ctx, "gemini-3-flash-preview", genai.Text("Explain quantum computing"), nil)
if err != nil {
log.Fatal(err)
}
fmt.Println(resp.Text)
}Java
java
import com.google.genai.Client;
import com.google.genai.types.GenerateContentResponse;
public class GenerateTextFromTextInput {
public static void main(String[] args) {
Client client = Client.builder().vertexAi(true).project("your-project-id").location("global").build();
GenerateContentResponse response =
client.models.generateContent(
"gemini-3-flash-preview",
"Explain quantum computing",
null);
System.out.println(response.text());
}
}C#/.NET
csharp
using Google.GenAI;
var client = new Client(
project: "your-project-id",
location: "global",
vertexAI: true
);
var response = await client.Models.GenerateContent(
"gemini-3-flash-preview",
"Explain quantum computing"
);
Console.WriteLine(response.Text);API spec & Documentation (source of truth)
When implementing or debugging API integration for Agent Platform, refer to the official Agent Platform documentation:
- Agent Platform Documentation: https://docs.cloud.google.com/gemini-enterprise-agent-platform/overview
- REST API Reference: https://docs.cloud.google.com/gemini-enterprise-agent-platform/reference/rest
The Gen AI SDK on Agent Platform uses the or REST API endpoints (e.g., ).
v1beta1v1https://{LOCATION}-aiplatform.googleapis.com/v1beta1/projects/{PROJECT}/locations/{LOCATION}/publishers/google/models/{MODEL}:generateContent[!TIP] Use the Developer Knowledge MCP Server: If theorsearch_documentstools are available, use them to find and retrieve official documentation for Google Cloud and Agent Platform directly within the context. This is the preferred method for getting up-to-date API details and code snippets.get_document
Workflows and Code Samples
Reference the Python Docs Samples repository for additional code samples and specific usage scenarios.
Depending on the specific user request, refer to the following reference files for detailed code samples and usage patterns (Python examples):
- Text & Multimodal: Chat, Multimodal inputs (Image, Video, Audio), and Streaming. See references/text_and_multimodal.md
- Embeddings: Generate text embeddings for semantic search. See references/embeddings.md
- Structured Output & Tools: JSON generation, Function Calling, Search Grounding, and Code Execution. See references/structured_and_tools.md
- Media Generation: Image generation, Image editing, and Video generation. See references/media_generation.md
- Bounding Box Detection: Object detection and localization within images and video. See references/bounding_box.md
- Live API: Real-time bidirectional streaming for voice, vision, and text. See references/live_api.md
- Advanced Features: Content Caching, Batch Prediction, and Thinking/Reasoning. See references/advanced_features.md
- Safety: Adjusting Responsible AI filters and thresholds. See references/safety.md
- Model Tuning: Supervised Fine-Tuning and Preference Tuning. See references/model_tuning.md