NotebookLM Deep Research & Slide Generator
Primary Workflow
When user requests a presentation or research on a topic:
Step 1: Deep Research via NotebookLM MCP
- Invoke the NotebookLM MCP to perform deep research on the topic
- Gather comprehensive sources and insights
- Extract key findings, data points, and citations
Step 2: Generate Highly Infographic Slide Presentation
Based on the user's topic outline, create visually-rich infographic slides:
Design Specifications:
- Background: White (#FFFFFF)
- Font: Arial (all text)
- Style: HIGHLY INFOGRAPHIC - visual-first design
- Content: Based exclusively on deep research sources
- Citations: Include source references on each slide
Infographic Elements (REQUIRED on every content slide):
- 📊 Charts & Graphs - Bar charts, pie charts, line graphs for data
- 📈 Statistics Callouts - Large numbers with context (e.g., "87% of users...")
- 🎯 Icons - Use relevant icons for each concept
- 🔲 Flowcharts - Process diagrams, decision trees
- 📋 Comparison Tables - Side-by-side comparisons
- 🗺️ Diagrams - Concept maps, timelines, hierarchies
- 💡 Highlight Boxes - Key takeaways in colored boxes
- ➡️ Visual Hierarchy - Clear information flow
Infographic Design Rules:
- Minimal text - Max 6 bullet points per slide, each under 10 words
- Data visualization - Convert all statistics to visual charts
- Icon-driven - Every section header has an icon
- Color coding - Use consistent colors for categories
- White space - 40% of slide should be empty for clarity
- One idea per slide - Don't overcrowd
Slide Structure:
- Title Slide - Topic name, subtitle, date, hero image/icon
- Agenda/Outline - Visual roadmap with icons
- Content Slides - Infographic-style, one concept per slide
- Key Statistics - Large numbers with visual context
- Key Findings - Summary infographic
- Sources/References - Clean list with icons
Step 3: Output Format
Generate slides as:
- Markdown format (for review)
- Or HTML/reveal.js format (for presentation)
NotebookLM API Reference
Complete programmatic access to Google NotebookLM—including capabilities not exposed in the web UI. Create notebooks, add sources (URLs, YouTube, PDFs, audio, video, images), chat with content, generate all artifact types, and download results in multiple formats.
Installation
From PyPI (Recommended):
bash
pip install notebooklm-py
From GitHub (use latest release tag, NOT main branch):
bash
# Get the latest release tag (using curl)
LATEST_TAG=$(curl -s https://api.github.com/repos/teng-lin/notebooklm-py/releases/latest | grep '"tag_name"' | cut -d'"' -f4)
pip install "git+https://github.com/teng-lin/notebooklm-py@${LATEST_TAG}"
⚠️
DO NOT install from main branch (
pip install git+https://github.com/teng-lin/notebooklm-py
). The main branch may contain unreleased/unstable changes. Always use PyPI or a specific release tag, unless you are testing unreleased features.
After installation, install the Claude Code skill:
Prerequisites
IMPORTANT: Before using any command, you MUST authenticate:
bash
notebooklm login # Opens browser for Google OAuth
notebooklm list # Verify authentication works
If commands fail with authentication errors, re-run
.
CI/CD, Multiple Accounts, and Parallel Agents
For automated environments, multiple accounts, or parallel agent workflows:
| Variable | Purpose |
|---|
| Custom config directory (default: ) |
| Inline auth JSON - no file writes needed |
CI/CD setup: Set
from a secret containing your
contents.
Multiple accounts: Use different
directories per account.
Parallel agents: The CLI stores notebook context in a shared file (
~/.notebooklm/context.json
). Multiple concurrent agents using
can overwrite each other's context.
Solutions for parallel workflows:
- Always use explicit notebook ID (recommended): Pass (for / commands) or (for others) instead of relying on
- Per-agent isolation: Set unique per agent:
export NOTEBOOKLM_HOME=/tmp/agent-$ID
- Use full UUIDs: Avoid partial IDs in automation (they can become ambiguous)
Agent Setup Verification
Before starting workflows, verify the CLI is ready:
- → Should show "Authenticated as: email@..."
- → Should return valid JSON (even if empty notebooks list)
- If either fails → Run
When This Skill Activates
Explicit: User says "/notebooklm", "use notebooklm", or mentions the tool by name
Intent detection: Recognize requests like:
- "Create a podcast about [topic]"
- "Summarize these URLs/documents"
- "Generate a quiz from my research"
- "Turn this into an audio overview"
- "Create flashcards for studying"
- "Generate a video explainer"
- "Make an infographic"
- "Create a mind map of the concepts"
- "Download the quiz as markdown"
- "Add these sources to NotebookLM"
Autonomy Rules
Run automatically (no confirmation):
- - check context
- - diagnose auth issues
- - list notebooks
- - list sources
- - list artifacts
- - list supported languages
- - get current language
- - set language (global setting)
- - wait for artifact completion (in subagent context)
- - wait for source processing (in subagent context)
notebooklm research status
- check research status
- - wait for research (in subagent context)
- - set context (⚠️ SINGLE-AGENT ONLY - use flag in parallel workflows)
- - create notebook
- - chat queries
- - add sources
Ask before running:
- - destructive
- - long-running, may fail
- - writes to filesystem
- - long-running (when in main conversation)
- - long-running (when in main conversation)
- - long-running (when in main conversation)
Quick Reference
| Task | Command |
|---|
| Authenticate | |
| Diagnose auth issues | |
| Diagnose auth (full) | notebooklm auth check --test
|
| List notebooks | |
| Create notebook | notebooklm create "Title"
|
| Set context | notebooklm use <notebook_id>
|
| Show context | |
| Add URL source | notebooklm source add "https://..."
|
| Add file | notebooklm source add ./file.pdf
|
| Add YouTube | notebooklm source add "https://youtube.com/..."
|
| List sources | |
| Wait for source processing | notebooklm source wait <source_id>
|
| Web research (fast) | notebooklm source add-research "query"
|
| Web research (deep) | notebooklm source add-research "query" --mode deep --no-wait
|
| Check research status | notebooklm research status
|
| Wait for research | notebooklm research wait --import-all
|
| Chat | notebooklm ask "question"
|
| Chat (new conversation) | notebooklm ask "question" --new
|
| Chat (specific sources) | notebooklm ask "question" -s src_id1 -s src_id2
|
| Chat (with references) | notebooklm ask "question" --json
|
| Get source fulltext | notebooklm source fulltext <source_id>
|
| Get source guide | notebooklm source guide <source_id>
|
| Generate podcast | notebooklm generate audio "instructions"
|
| Generate podcast (JSON) | notebooklm generate audio --json
|
| Generate podcast (specific sources) | notebooklm generate audio -s src_id1 -s src_id2
|
| Generate video | notebooklm generate video "instructions"
|
| Generate quiz | |
| Check artifact status | |
| Wait for completion | notebooklm artifact wait <artifact_id>
|
| Download audio | notebooklm download audio ./output.mp3
|
| Download video | notebooklm download video ./output.mp4
|
| Download report | notebooklm download report ./report.md
|
| Download mind map | notebooklm download mind-map ./map.json
|
| Download data table | notebooklm download data-table ./data.csv
|
| Download quiz | notebooklm download quiz quiz.json
|
| Download quiz (markdown) | notebooklm download quiz --format markdown quiz.md
|
| Download flashcards | notebooklm download flashcards cards.json
|
| Download flashcards (markdown) | notebooklm download flashcards --format markdown cards.md
|
| Delete notebook | notebooklm notebook delete <id>
|
| List languages | |
| Get language | |
| Set language | notebooklm language set zh_Hans
|
Parallel safety: Use explicit notebook IDs in parallel workflows. Commands supporting
shorthand:
,
,
,
. Download commands also support
. Other commands use
. For chat, use
to start fresh conversations (avoids conversation ID conflicts).
Partial IDs: Use first 6+ characters of UUIDs. Must be unique prefix (fails if ambiguous). Works for:
,
,
commands. For automation, prefer full UUIDs to avoid ambiguity.
Command Output Formats
Commands with
return structured data for parsing:
Create notebook:
$ notebooklm create "Research" --json
{"id": "abc123de-...", "title": "Research"}
Add source:
$ notebooklm source add "https://example.com" --json
{"source_id": "def456...", "title": "Example", "status": "processing"}
Generate artifact:
$ notebooklm generate audio "Focus on key points" --json
{"task_id": "xyz789...", "status": "pending"}
Chat with references:
$ notebooklm ask "What is X?" --json
{"answer": "X is... [1] [2]", "conversation_id": "...", "turn_number": 1, "is_follow_up": false, "references": [{"source_id": "abc123...", "citation_number": 1, "cited_text": "Relevant passage from source..."}, {"source_id": "def456...", "citation_number": 2, "cited_text": "Another passage..."}]}
Source fulltext (get indexed content):
$ notebooklm source fulltext <source_id> --json
{"source_id": "...", "title": "...", "char_count": 12345, "content": "Full indexed text..."}
Understanding citations: The
in references is often a snippet or section header, not the full quoted passage. The
/
positions reference NotebookLM's internal chunked index, not the raw fulltext. Use
SourceFulltext.find_citation_context()
to locate citations:
python
fulltext = await client.sources.get_fulltext(notebook_id, ref.source_id)
matches = fulltext.find_citation_context(ref.cited_text) # Returns list[(context, position)]
if matches:
context, pos = matches[0] # First match; check len(matches) > 1 for duplicates
Extract IDs: Parse the
,
, or
field from JSON output.
Generation Types
All generate commands support:
- to use specific source(s) instead of all sources
- to set output language (defaults to configured language or 'en')
- for machine-readable output (returns and )
- to automatically retry on rate limits with exponential backoff
| Type | Command | Options | Download |
|---|
| Podcast | | --format [deep-dive|brief|critique|debate]
, --length [short|default|long]
| .mp3 |
| Video | | --format [explainer|brief]
, --style [auto|classic|whiteboard|kawaii|anime|watercolor|retro-print|heritage|paper-craft]
| .mp4 |
| Slide Deck | | --format [detailed|presenter]
, | .pdf |
| Infographic | | --orientation [landscape|portrait|square]
, --detail [concise|standard|detailed]
| .png |
| Report | | --format [briefing-doc|study-guide|blog-post|custom]
| .md |
| Mind Map | | (sync, instant) | .json |
| Data Table | | description required | .csv |
| Quiz | | --difficulty [easy|medium|hard]
, --quantity [fewer|standard|more]
| .json/.md/.html |
| Flashcards | | --difficulty [easy|medium|hard]
, --quantity [fewer|standard|more]
| .json/.md/.html |
Features Beyond the Web UI
These capabilities are available via CLI but not in NotebookLM's web interface:
| Feature | Command | Description |
|---|
| Batch downloads | | Download all artifacts of a type at once |
| Quiz/Flashcard export | download quiz --format json
| Export as JSON, Markdown, or HTML (web UI only shows interactive view) |
| Mind map extraction | | Export hierarchical JSON for visualization tools |
| Data table export | | Download structured tables as CSV |
| Source fulltext | | Retrieve the indexed text content of any source |
| Programmatic sharing | commands | Manage sharing permissions without the UI |
Common Workflows
Research to Podcast (Interactive)
Time: 5-10 minutes total
notebooklm create "Research: [topic]"
— if fails: check auth with
- for each URL/document — if one fails: log warning, continue with others
- Wait for sources:
notebooklm source list --json
until all status=READY — required before generation
notebooklm generate audio "Focus on [specific angle]"
(confirm when asked) — if rate limited: wait 5 min, retry once
- Note the artifact ID returned
- Check later for status
notebooklm download audio ./podcast.mp3
when complete (confirm when asked)
Research to Podcast (Automated with Subagent)
Time: 5-10 minutes, but continues in background
When user wants full automation (generate and download when ready):
- Create notebook and add sources as usual
- Wait for sources to be ready (use or check )
- Run
notebooklm generate audio "..." --json
→ parse from output
- Spawn a background agent using Task tool:
Task(
prompt="Wait for artifact {artifact_id} in notebook {notebook_id} to complete, then download.
Use: notebooklm artifact wait {artifact_id} -n {notebook_id} --timeout 600
Then: notebooklm download audio ./podcast.mp3 -a {artifact_id} -n {notebook_id}",
subagent_type="general-purpose"
)
- Main conversation continues while agent waits
Error handling in subagent:
- If returns exit code 2 (timeout): Report timeout, suggest checking
- If download fails: Check if artifact status is COMPLETED first
Benefits: Non-blocking, user can do other work, automatic download on completion
Document Analysis
Time: 1-2 minutes
notebooklm create "Analysis: [project]"
notebooklm source add ./doc.pdf
(or URLs)
notebooklm ask "Summarize the key points"
notebooklm ask "What are the main arguments?"
- Continue chatting as needed
Bulk Import
Time: Varies by source count
notebooklm create "Collection: [name]"
- Add multiple sources:
bash
notebooklm source add "https://url1.com"
notebooklm source add "https://url2.com"
notebooklm source add ./local-file.pdf
- to verify
Source limits: Max 50 sources per notebook
Supported types: PDFs, YouTube URLs, web URLs, Google Docs, text files, Markdown, Word docs, audio files, video files, images
Bulk Import with Source Waiting (Subagent Pattern)
Time: Varies by source count
When adding multiple sources and needing to wait for processing before chat/generation:
- Add sources with to capture IDs:
bash
notebooklm source add "https://url1.com" --json # → {"source_id": "abc..."}
notebooklm source add "https://url2.com" --json # → {"source_id": "def..."}
- Spawn a background agent to wait for all sources:
Task(
prompt="Wait for sources {source_ids} in notebook {notebook_id} to be ready.
For each: notebooklm source wait {id} -n {notebook_id} --timeout 120
Report when all ready or if any fail.",
subagent_type="general-purpose"
)
- Main conversation continues while agent waits
- Once sources are ready, proceed with chat or generation
Why wait for sources? Sources must be indexed before chat or generation. Takes 10-60 seconds per source.
Deep Web Research (Subagent Pattern)
Time: 2-5 minutes, runs in background
Deep research finds and analyzes web sources on a topic:
- Create notebook:
notebooklm create "Research: [topic]"
- Start deep research (non-blocking):
bash
notebooklm source add-research "topic query" --mode deep --no-wait
- Spawn a background agent to wait and import:
Task(
prompt="Wait for research in notebook {notebook_id} to complete and import sources.
Use: notebooklm research wait -n {notebook_id} --import-all --timeout 300
Report how many sources were imported.",
subagent_type="general-purpose"
)
- Main conversation continues while agent waits
- When agent completes, sources are imported automatically
Alternative (blocking): For simple cases, omit
:
bash
notebooklm source add-research "topic" --mode deep --import-all
# Blocks for up to 5 minutes
When to use each mode:
- : Specific topic, quick overview needed (5-10 sources, seconds)
- : Broad topic, comprehensive analysis needed (20+ sources, 2-5 min)
Research sources:
- : Search the web (default)
- : Search Google Drive
Output Style
Progress updates: Brief status for each step
- "Creating notebook 'Research: AI'..."
- "Adding source: https://example.com..."
- "Starting audio generation... (task ID: abc123)"
Fire-and-forget for long operations:
- Start generation, return artifact ID immediately
- Do NOT poll or wait in main conversation - generation takes 5-45 minutes (see timing table)
- User checks status manually, OR use subagent with
JSON output: Use
flag for machine-readable output:
bash
notebooklm list --json
notebooklm auth check --json
notebooklm source list --json
notebooklm artifact list --json
JSON schemas (key fields):
json
{"notebooks": [{"id": "...", "title": "...", "created_at": "..."}]}
notebooklm auth check --json
:
json
{"checks": {"storage_exists": true, "json_valid": true, "cookies_present": true, "sid_cookie": true, "token_fetch": true}, "details": {"storage_path": "...", "auth_source": "file", "cookies_found": ["SID", "HSID", "..."], "cookie_domains": [".google.com"]}}
notebooklm source list --json
:
json
{"sources": [{"id": "...", "title": "...", "status": "ready|processing|error"}]}
notebooklm artifact list --json
:
json
{"artifacts": [{"id": "...", "title": "...", "type": "Audio Overview", "status": "in_progress|pending|completed|unknown"}]}
Status values:
- Sources: → (or )
- Artifacts: or → (or )
Error Handling
On failure, offer the user a choice:
- Retry the operation
- Skip and continue with something else
- Investigate the error
Error decision tree:
| Error | Cause | Action |
|---|
| Auth/cookie error | Session expired | Run then |
| "No notebook context" | Context not set | Use or flag (parallel), or (single-agent) |
| "No result found for RPC ID" | Rate limiting | Wait 5-10 min, retry |
| Google rate limit | Wait and retry later |
| Download fails | Generation incomplete | Check for status |
| Invalid notebook/source ID | Wrong ID | Run to verify |
| RPC protocol error | Google changed APIs | May need CLI update |
Exit Codes
All commands use consistent exit codes:
| Code | Meaning | Action |
|---|
| 0 | Success | Continue |
| 1 | Error (not found, processing failed) | Check stderr, see Error Handling |
| 2 | Timeout (wait commands only) | Extend timeout or check status manually |
Examples:
- returns 1 if source not found or processing failed
- returns 2 if timeout reached before completion
- returns 1 if rate limited (check stderr for details)
Known Limitations
Rate limiting: Audio, video, quiz, flashcards, infographic, and slide deck generation may fail due to Google's rate limits. This is an API limitation, not a bug.
Reliable operations: These always work:
- Notebooks (list, create, delete, rename)
- Sources (add, list, delete)
- Chat/queries
- Mind-map, study-guide, report, data-table generation
Unreliable operations: These may fail with rate limiting:
- Audio (podcast) generation
- Video generation
- Quiz and flashcard generation
- Infographic and slide deck generation
Workaround: If generation fails:
- Check status:
- Retry after 5-10 minutes
- Use the NotebookLM web UI as fallback
Processing times vary significantly. Use the subagent pattern for long operations:
| Operation | Typical time | Suggested timeout |
|---|
| Source processing | 30s - 10 min | 600s |
| Research (fast) | 30s - 2 min | 180s |
| Research (deep) | 15 - 30+ min | 1800s |
| Notes | instant | n/a |
| Mind-map | instant (sync) | n/a |
| Quiz, flashcards | 5 - 15 min | 900s |
| Report, data-table | 5 - 15 min | 900s |
| Audio generation | 10 - 20 min | 1200s |
| Video generation | 15 - 45 min | 2700s |
Polling intervals: When checking status manually, poll every 15-30 seconds to avoid excessive API calls.
Language Configuration
Language setting controls the output language for generated artifacts (audio, video, etc.).
Important: Language is a GLOBAL setting that affects all notebooks in your account.
bash
# List all 80+ supported languages with native names
notebooklm language list
# Show current language setting
notebooklm language get
# Set language for artifact generation
notebooklm language set zh_Hans # Simplified Chinese
notebooklm language set ja # Japanese
notebooklm language set en # English (default)
Common language codes:
| Code | Language |
|---|
| English |
| 中文(简体) - Simplified Chinese |
| 中文(繁體) - Traditional Chinese |
| 日本語 - Japanese |
| 한국어 - Korean |
| Español - Spanish |
| Français - French |
| Deutsch - German |
| Português (Brasil) |
Override per command: Use
flag on generate commands:
bash
notebooklm generate audio --language ja # Japanese podcast
notebooklm generate video --language zh_Hans # Chinese video
Offline mode: Use
flag to skip server sync:
bash
notebooklm language set zh_Hans --local # Save locally only
notebooklm language get --local # Read local config only
Troubleshooting
bash
notebooklm --help # Main commands
notebooklm auth check # Diagnose auth issues
notebooklm auth check --test # Full auth validation with network test
notebooklm notebook --help # Notebook management
notebooklm source --help # Source management
notebooklm research --help # Research status/wait
notebooklm generate --help # Content generation
notebooklm artifact --help # Artifact management
notebooklm download --help # Download content
notebooklm language --help # Language settings
Diagnose auth: - shows cookie domains, storage path, validation status
Re-authenticate:
Check version:
Update skill: