Gemini-Deep-Research
Gemini-Deep-Research is a live-fetch agent operated by Google. It does not crawl the web on a schedule. It hits your site only when an end-user asks the underlying AI a question that requires fresh information from a specific page.
Traffic is bursty and unpredictable. A single trending topic can send hundreds of Gemini-Deep-Research requests in an hour, then nothing for days. Each request typically reads one or two pages, not your whole site.
Allowing Gemini-Deep-Research is how your content becomes part of Google's answers. Blocking it means users asking that AI about your topic will be answered using someone else's content instead.
See Gemini-Deep-Research on your own site
Match the User-Agent header on incoming requests against the pattern below.
regex
Verify by IP
For higher confidence, also verify the source IP against the operator's published ranges. UA strings can be spoofed; IP ownership is harder to fake.
Renders JavaScript
Sometimes
IP verification
Published IP ranges
Crawl frequency
Burst, user-driven
Honors robots.txt
Yes
Honors Crawl-delay
Varies
Google runs 103 bots in total. Each one is a separate user-agent so you can allow or block them independently.
Link Unfurler
25- Google Feed Fetcher
- Google Image Proxy
- Google Publisher Center
- Google Web Preview
- Google-AdWords-Express
- Google-PageRenderer
- Google-Read-Aloud
- GoogleDocs
- GoogleImageProxy
- GoogleProducer
- Gmail Image Proxy
- Google Calendar Importer
- Google Page Renderer
- Google Web Snippet
- Google API
- Google Cloud Scheduler
- Google-Document-Conversion
- Google Sheets
- Google Slides
- Google Docs
- Google Area 120 Privacy Policy Fetcher
- Chrome Privacy Preserving Prefetch Proxy
- Google Cloud Function
- GoogleApps-DocumentScanner
- GoogleStackdriverMonitoring
Search Engine
14DevOps & Monitoring
13- Chrome-Lighthouse
- Google Inspection Tool
- Google Schema Markup Testing Tool
- Google Trust Services (DCV Check)
- Google-Structured-Data-Testing-Tool
- GoogleAssociationService
- ProjectShield Url Check
- Google Read Aloud
- Google Search Console
- Google Page Speed Insights
- Google Partner Monitoring
- Google Structured Data Testing Tool
- Google Stackdriver Monitoring
Ads Network Bot
11Training Crawler
9Agentic Browser
8Security Scanner
7SEO Crawler
4Brand Intelligence
3Shopping Bot
3Live-Fetch AI
2- Gemini-Deep-ResearchYou are here
- Google-NotebookLM
Ad Verification
2AI Coding Tool
1Agentic Commerce
1Should I let Gemini-Deep-Research through?
In most cases, yes. Live-fetch agents drive citations inside AI answers. Allowing keeps your content in the conversation. If volume gets noisy, rate-limit it before you block it outright.
Does blocking Gemini-Deep-Research affect my Google rankings?
No. Gemini-Deep-Research fetches a page only when a user is actively asking Google a question. It has nothing to do with how Google or Bing rank you. The cost of blocking is that Google can't quote your content in its answer.
How do I confirm a request is really from Gemini-Deep-Research?
Two checks. The User-Agent header should match a known Gemini-Deep-Research string, and the request's source IP should fall inside Google's published ranges. The User-Agent alone is trivially spoofable, so the IP check is what gives you confidence. Google publishes the ranges so you can validate at the CDN or edge.
Does a Gemini-Deep-Research visit count as a real user visit?
Sort of. There is a human asking Google a question on the other end, but they never load your page in their own browser. They see whatever Google quotes back, usually a snippet plus a citation link. Count it as upstream attention rather than as a session.
How is Gemini-Deep-Research different from Google's other bots?
Google splits work across multiple user-agents so site owners can decide on each one independently. Training crawlers, live-fetch agents, search indexers, and agentic browsers each get their own name. Worth scanning the rest of the Google family above to see which ones actually matter for your site.
What's the cleanest way to control Gemini-Deep-Research?
Two layers. Robots.txt for the polite crawlers that read it, and rules at your CDN or edge for the ones that don't. Rankly's Agent Experience handles both from a single config, so you can allow, block, rate-limit, or serve a stripped-down version per bot. Agent Analytics handles the observation half so you know which bots are actually worth a rule.
Verify everything above against the operator's own documentation.