透過集合功能整理內容
你可以依據偏好儲存及分類內容。
Googlebot
Googlebot 是 Google 搜尋兩種網路檢索器的通稱:
您可以從要求中的 HTTP user-agent
要求標頭來辨別是哪一種檢索器,不過這兩種檢索器都會遵循 robots.txt 中的同一個產品權杖 (使用者代理程式權杖),因此您無法在 robots.txt 中特意指定 Googlebot Smartphone 或 Googlebot Desktop。
Google 搜尋主要是將大多數網站的行動版內容編入索引。因此,大部分 Googlebot 檢索要求都會來自行動裝置檢索器,只有一小部分會來自電腦檢索器。
Googlebot 如何存取您的網站
一般而言,Googlebot 並不會在幾秒的時間內連續存取您的網站好幾次。不過,如果出現延遲的情況,短時間內的檢索頻率確實可能會稍微偏高。 如果您的網站無法應付 Google 的檢索要求,可以降低檢索頻率。
Googlebot 可以檢索 HTML 檔案或支援的文字檔的前 15 MB。系統會個別擷取 HTML 中參照的各項資源 (例如 CSS 和 JavaScript),且每次擷取都會受到相同檔案大小上限的限制。在檢索完檔案的前 15 MB 之後,Googlebot 就會停止檢索,而且只會傳送檔案的前 15 MB 以供建立索引。檔案大小限制會套用至未壓縮資料。其他 Google 檢索器 (例如 Googlebot 影片和 Googlebot 圖片) 可能會有不同的限制。
從美國的 IP 位址進行檢索作業時,Googlebot 的時區為太平洋時間。
如要瞭解其他 Googlebot 技術性資源,請參閱 Google 檢索器總覽。
禁止 Googlebot 瀏覽您的網站
Googlebot 主要是透過先前檢索網頁中嵌入的連結,發掘要檢索的新網址。在大部分情況下,即使您不公開網站的連結,也無法讓網站保持隱藏。這是因為您的「秘密」網站還是會連結到其他網站,一旦有人點選這個連結,您的「秘密」網址就會出現在參照標記中,其他網站也可以將該網址儲存在其參照記錄中並加以發布。
如要禁止 Googlebot 檢索您網站上的內容,您可以採取多種方式。請記住,「檢索」和「建立索引」是兩種不同的作業。禁止 Googlebot 檢索網頁並無法禁止 Googlebot 在搜尋結果中顯示該網頁的網址:
封鎖 Googlebot 會影響 Google 搜尋 (包括探索和所有 Google 搜尋功能) 和其他產品 (例如 Google 圖片、Google 影片和 Google 新聞)。
驗證 Googlebot
決定封鎖 Googlebot 前,請留意其他檢索器經常假冒 Googlebot 採用的 HTTP user-agent
要求標頭。因此請務必驗證有問題的要求,確認是否真的由 Google 提出。如要確認要求是否來自 Googlebot,最好的做法是針對要求的來源 IP 使用反向 DNS 查詢,或是比對來源 IP 與 Googlebot IP 範圍。
除非另有註明,否則本頁面中的內容是採用創用 CC 姓名標示 4.0 授權,程式碼範例則為阿帕契 2.0 授權。詳情請參閱《Google Developers 網站政策》。Java 是 Oracle 和/或其關聯企業的註冊商標。
上次更新時間:2025-08-04 (世界標準時間)。
[null,null,["上次更新時間:2025-08-04 (世界標準時間)。"],[[["\u003cp\u003eGooglebot is the name for Google's web crawlers, which include Googlebot Smartphone and Googlebot Desktop, used to understand and index website content.\u003c/p\u003e\n"],["\u003cp\u003eGooglebot primarily crawls and indexes the mobile version of websites, reflecting the dominance of mobile browsing.\u003c/p\u003e\n"],["\u003cp\u003eWebsite owners can control Googlebot's access by using robots.txt to manage crawl rate and prevent crawling of specific content.\u003c/p\u003e\n"],["\u003cp\u003eWhile blocking Googlebot prevents crawling, it doesn't automatically remove a page from Google Search results; \u003ccode\u003enoindex\u003c/code\u003e should be used for that purpose.\u003c/p\u003e\n"],["\u003cp\u003eIt's crucial to verify the authenticity of Googlebot requests as its user agent is frequently imitated by other crawlers.\u003c/p\u003e\n"]]],["Googlebot, comprising Desktop and Smartphone crawlers, indexes web content, primarily favoring the mobile version. It crawls most sites at a rate of once every few seconds, fetching up to 15MB of HTML or text-based files and their resources. To manage Googlebot's access, sites can use `robots.txt` to block crawling or `noindex` to prevent indexing. Blocking crawling affects Google Search and related products. Verify Googlebot requests via reverse DNS lookup or by checking the IP range.\n"],null,["# What Is Googlebot | Google Search Central\n\nGooglebot\n=========\n\n\nGooglebot is the generic name for two types of\n[web crawlers](/search/docs/fundamentals/how-search-works) used by Google Search:\n\n- [**Googlebot Smartphone**](/search/docs/crawling-indexing/google-common-crawlers#googlebot-smartphone): a mobile crawler that simulates a user on a mobile device.\n- [**Googlebot Desktop**](/search/docs/crawling-indexing/google-common-crawlers#googlebot-desktop): a desktop crawler that simulates a user on desktop.\n\n\nYou can identify the subtype of Googlebot by looking at the\n[HTTP `user-agent` request header](/search/docs/crawling-indexing/overview-google-crawlers)\nin the request. However, both crawler types obey the same product token (user agent token) in\nrobots.txt, and so you cannot selectively target either Googlebot Smartphone or Googlebot\nDesktop using robots.txt.\n\n\nFor most sites Google Search primarily\n[indexes the mobile version](/search/docs/crawling-indexing/mobile/mobile-sites-mobile-first-indexing)\nof the content. As such the majority of Googlebot crawl requests will be made using the mobile\ncrawler, and a minority using the desktop crawler.\n\nHow Googlebot accesses your site\n--------------------------------\n\n\nFor most sites, Googlebot shouldn't access your site more than once every few seconds on\naverage. However, due to delays it's possible that the rate will appear to be slightly higher\nover short periods. If your site is having trouble keeping up with Google's crawling requests, you\ncan [reduce the crawl rate.](/search/docs/crawling-indexing/reduce-crawl-rate)\n\n\nGooglebot can crawl the first 15MB of an HTML file or\n[supported text-based file](/search/docs/crawling-indexing/indexable-file-types).\nEach resource referenced in the HTML such as CSS and JavaScript is fetched separately, and\neach fetch is bound by the same file size limit. After the first 15MB of the file, Googlebot\nstops crawling and only sends the first 15MB of the file for indexing consideration. The file size\nlimit is applied on the uncompressed data. Other Google crawlers, for example Googlebot Video and\nGooglebot Image, may have different limits.\n\n\nWhen crawling from IP addresses in the US, the timezone of Googlebot is\n[Pacific Time](https://g.co/kgs/WSf8oR).\n\n\nOther\n[technical properties of Googlebot](/search/docs/crawling-indexing/overview-google-crawlers#crawl-technical-props)\nare described in the overview of Google's crawlers.\n\nBlocking Googlebot from visiting your site\n------------------------------------------\n\n\nGooglebot discovers new URLs to crawl primarily from links embedded in previously crawled pages.\nIt's almost impossible to keep a site secret by not publishing links to it. For example, as soon\nas someone clicks a link from your \"secret\" site to another site, your \"secret\" site URL may\nappear in the referrer tag and can be stored and published by the other site in its referrer log.\n\n\nIf you want to prevent Googlebot from crawling content on your site, you have a\n[number of options](/search/docs/crawling-indexing/control-what-you-share). Remember\nthere's a difference between *crawling* and *indexing*; blocking Googlebot from crawling\na page doesn't prevent the URL of the page from appearing in search results:\n\n- **Prevent Googlebot from crawling a page?** Use a [robots.txt file](/search/docs/crawling-indexing/robots/intro).\n- **Don't want Google to index a page?** Use [`noindex`](/search/docs/crawling-indexing/block-indexing).\n- **Prevent a page from being accessible at all by both crawlers or users?** Use [another method, such as password protection](/search/docs/crawling-indexing/control-what-you-share).\n\n\nBlocking Googlebot affects Google Search (including Discover and all Google Search features), as\nwell as other products such as Google Images, Google Video, and Google News.\n\nVerifying Googlebot\n-------------------\n\n\nBefore you decide to block Googlebot, be aware that the HTTP `user-agent` request\nheader used by Googlebot is often spoofed by other crawlers. It's important to verify that a\nproblematic request actually comes from Google. The best way to verify that a request actually\ncomes from Googlebot is to\n[use a reverse DNS lookup](/search/docs/crawling-indexing/verifying-googlebot#manual)\non the source IP of the request, or to match the source IP against the\n[Googlebot IP ranges](/search/docs/crawling-indexing/verifying-googlebot#use-automatic-solutions)."]]