透過集合功能整理內容
你可以依據偏好儲存及分類內容。
詳盡指南:Google 搜尋的運作方式
Google 搜尋是全自動化的搜尋引擎,會運用一種稱為「網路檢索器」的軟體定期探索網路,找出能加入索引的網頁。實際上,我們搜尋結果列出的大多數網頁都不是由人為提交,而是網路檢索器在探索網路時找到,然後自動加入搜尋結果。本文件說明 Google 搜尋在您網站中運作的各個階段。掌握這項基本知識有助於修正檢索問題、讓網頁編入索引,以及瞭解如何最佳化網站在 Google 搜尋中的呈現方式。
開始前的注意事項
深入瞭解 Google 搜尋的運作方式之前,請務必注意,Google 不接受以付費的方式增加檢索某網站的頻率或提高其排名。如果有人告訴您可以這樣做,請注意這並非事實。
即使您的網頁符合 Google 搜尋基礎入門的做法,Google 也無法保證一定會檢索您的網頁、建立索引或提供網頁。
Google 搜尋三階段簡介
Google 搜尋分為三個階段運作,並非所有網頁都會經歷各個階段:
-
檢索:Google 會透過稱為檢索器的自動化程式,從網際網路上找到的網頁下載文字、圖片和影片。
-
建立索引:Google 會分析網頁上的文字、圖片和影片檔案,並將相關資訊儲存在 Google 索引 (一個大型資料庫) 中。
-
提供搜尋結果:當使用者利用 Google 搜尋資訊時,Google 會傳回與使用者查詢內容相關的資訊。
檢索
第一個階段是找出網路上有哪些網頁。由於世界上並沒有統一登錄網頁的地方,因此 Google 必須持續尋找新網頁,並將這些網頁加入已知網頁清單,這項程序稱為「網址搜尋」。清單中都是 Google 造訪過的網頁。Google 會擷取已知網頁的連結並連到新網頁,以此發現其他未知的網頁,例如類別網頁這種中心頁面,會連結到新的網誌文章。此外,當您提交網頁清單 (Sitemap) 供 Google 檢索時,我們也會找到其他網頁。
Google 找到網頁網址後,可能會造訪 (或「檢索」) 該網頁,瞭解網頁內容。我們使用大量電腦來檢索網路上數十億的網頁。用來擷取網頁的程式名為 Googlebot,又稱為檢索器、漫遊器或自動尋檢程式。Googlebot 會使用特定的演算程序來決定要檢索的網站、檢索頻率,以及要從每個網站擷取的網頁數量。此外,Google 檢索器也經過設計,會避免檢索網站速度過快,以免網站超載。這個機制是以網站的回應為基礎 (例如 HTTP 500 錯誤代表「放慢速度」)。
但是,Googlebot 不會檢索所有找到的網頁。網站擁有者可能會禁止 Googlebot 檢索部分網頁,有些網頁則需要登入才能存取。
在檢索期間,Google 會轉譯網頁,並使用最新版本的 Chrome 執行找到的任何 JavaScript,這與瀏覽器轉譯您造訪的網頁類似。轉譯作業非常重要,因為網站通常會利用 JavaScript 將內容投放到網頁,而如果沒有經過轉譯,Google 可能無法看到這類內容。
檢索作業取決於 Google 檢索器是否能存取網站,以下列舉 Googlebot 存取網站碰到的部分常見問題:
建立索引
網頁經過檢索之後,Google 會嘗試解讀網頁內容,這個階段稱為「索引」,其中包括處理及分析文字內容和關鍵內容標記和屬性,例如 <title>
元素和 alt 屬性、圖片、影片等等。
在建立索引的過程中,Google 會判斷網頁是否為網際網路上其他網頁的重複版本或標準網頁。
標準網頁是指可能出現在搜尋結果中的網頁,為了選取標準網頁,我們首先會將網際網路上存在類似內容的網頁分為一組 (又稱為分群法),然後再選取其中最具代表性的網頁。群組中的其他網頁是在不同情境下提供的替代版本,例如使用者透過行動裝置進行搜尋,或是使用者在尋找網頁叢集中的特定網頁。
此外,Google 也會收集標準網頁及其內容的相關信號,以便在下個階段用於在搜尋結果中提供網頁。例如網頁的語言、內容所在國家/地區與網頁可用性,都屬於這類信號。
系統收集到的標準網頁及其叢集相關資訊可能會儲存在 Google 索引中,這是由數千台電腦代管的大型資料庫。我們不保證一定會建立索引;並非 Google 處理過的每個網頁都會建立索引。
索引也會取決於網頁內容及其中繼資料。常見的索引問題包括:
提供搜尋結果
使用者輸入查詢字詞時,我們的系統會搜尋索引資料庫,並找出符合的網頁,再傳回我們認為與使用者的查詢內容最相關且品質最佳的結果。關聯性取決於數百種因素,使用者的所在位置、語言和裝置 (電腦或手機) 等資訊都可能包含在內。舉例來說,位於巴黎和香港的使用者搜尋「單車維修店」時看到的結果就不一樣。
根據使用者的查詢,搜尋結果網頁所顯示的搜尋功能也會有所變動。舉例來說,搜尋「單車維修店」時,系統可能會顯示本地搜尋結果,而且不會顯示圖片搜尋結果;不過,如果搜尋「現代單車」,則較有可能顯示圖片搜尋結果,但不太可能顯示本地搜尋結果。您可以前往視覺元素庫,瀏覽 Google 網頁搜尋中最常見的 UI 元素。
Search Console 可能會告知您網頁已建立索引,但不會顯示在搜尋結果中。可能原因如下:
雖然本指南說明了 Google 搜尋的運作方式,但我們會不斷改進演算法。
您可以追蹤 Google 搜尋中心網誌,隨時掌握這些變更。
除非另有註明,否則本頁面中的內容是採用創用 CC 姓名標示 4.0 授權,程式碼範例則為阿帕契 2.0 授權。詳情請參閱《Google Developers 網站政策》。Java 是 Oracle 和/或其關聯企業的註冊商標。
上次更新時間:2025-08-04 (世界標準時間)。
[null,null,["上次更新時間:2025-08-04 (世界標準時間)。"],[[["\u003cp\u003eGoogle Search discovers, analyzes, and ranks web pages to deliver relevant search results to users.\u003c/p\u003e\n"],["\u003cp\u003eThe three stages of Google Search are crawling, indexing, and serving search results.\u003c/p\u003e\n"],["\u003cp\u003eCrawling involves discovering and fetching web pages using automated programs called crawlers.\u003c/p\u003e\n"],["\u003cp\u003eIndexing involves analyzing the content and metadata of web pages to understand their topic and relevance.\u003c/p\u003e\n"],["\u003cp\u003eServing search results involves selecting and ranking relevant pages from the index based on user queries and various factors.\u003c/p\u003e\n"]]],["Google Search operates in three stages: crawling, indexing, and serving. Crawling involves automated web crawlers (Googlebot) discovering and downloading content (text, images, videos) from web pages. Indexing analyzes this content, determining its relevance and canonical status, storing it in Google's database. Serving involves matching user queries with indexed pages and displaying the most relevant results, considering factors like user location and device. Google does not accept payment for crawling, indexing or ranking and can't guarantee that the content will be crawled, indexed or served.\n"],null,["# In-Depth Guide to How Google Search Works | Google Search Central\n\nIn-depth guide to how Google Search works\n=========================================\n\n\nGoogle Search is a fully-automated search engine that uses software known as web crawlers that\nexplore the web regularly to find pages to add to our index. In fact, the vast majority of\npages listed in our results aren't manually submitted for inclusion, but are found and added\nautomatically when our web crawlers explore the web. This document explains the stages of how\nSearch works in the context of your website. Having this base knowledge can help you fix\ncrawling issues, get your pages indexed, and learn how to optimize how your site appears in\nGoogle Search.\n| Looking for something less technical? Check out our [How Search Works site](https://www.google.com/search/howsearchworks/), which explains how Search works from a searcher's perspective.\n\nA few notes before we get started\n---------------------------------\n\n\nBefore we get into the details of how Search works, it's important to note that Google doesn't\naccept payment to crawl a site more frequently, or rank it higher. If anyone tells you\notherwise, they're wrong.\n\n\nGoogle doesn't guarantee that it will crawl, index, or serve your page, even if your page\nfollows the [Google Search Essentials](/search/docs/essentials).\n\nIntroducing the three stages of Google Search\n---------------------------------------------\n\nGoogle Search works in three stages, and not all pages make it through each stage:\n\n1. [**Crawling:**](#crawling) Google downloads text, images, and videos from pages it found on the internet with automated programs called crawlers.\n2. [**Indexing:**](#indexing) Google analyzes the text, images, and video files on the page, and stores the information in the Google index, which is a large database.\n3. [**Serving search results:**](#serving) When a user searches on Google, Google returns information that's relevant to the user's query.\n\nCrawling\n--------\n\n\nThe first stage is finding out what pages exist on the web. There isn't a central registry of\nall web pages, so Google must constantly look for new and updated pages and add them to its\nlist of known pages. This process is called \"URL discovery\". Some pages are known because\nGoogle has already visited them. Other pages are discovered when Google extracts a link from a\nknown page to a new page: for example, a hub page, such as a category page, links to a new\nblog post. Still other pages are discovered when you submit a list of pages (a\n[sitemap](/search/docs/crawling-indexing/sitemaps/overview)) for Google to crawl. \n\n\nOnce Google discovers a page's URL, it may visit (or \"crawl\") the page to find out what's on\nit. We use a huge set of computers to crawl billions of pages on the web. The program that\ndoes the fetching is called [Googlebot](/search/docs/crawling-indexing/googlebot)\n(also known as a crawler, robot, bot, or spider). Googlebot uses an algorithmic process to\ndetermine which sites to crawl, how often, and how many pages to fetch from each site.\n[Google's crawlers](/search/docs/crawling-indexing/overview-google-crawlers)\nare also programmed such that they try not to crawl the site too fast to avoid overloading it.\nThis mechanism is based on the responses of the site (for example,\n[HTTP 500 errors mean \"slow down\"](/search/docs/crawling-indexing/http-network-errors#http-status-codes)).\n\n\nHowever, Googlebot doesn't crawl all the pages it discovered. Some pages may be\n[disallowed for crawling](/search/docs/crawling-indexing/robots/robots_txt#disallow) by the\nsite owner, other pages may not be accessible without logging in to the site.\n\n\nDuring the crawl, Google renders the page and\n[runs any JavaScript it finds](/search/docs/crawling-indexing/javascript/javascript-seo-basics#how-googlebot-processes-javascript)\nusing a recent version of\n[Chrome](https://www.google.com/chrome/), similar to how your\nbrowser renders pages you visit. Rendering is important because websites often rely on\nJavaScript to bring content to the page, and without rendering Google might not see that\ncontent.\n\n\nCrawling depends on whether Google's crawlers can access the site. Some common issues with\nGooglebot accessing sites include:\n\n- [Problems with the server handling the site](/search/docs/crawling-indexing/http-network-errors#http-status-codes)\n- [Network issues](/search/docs/crawling-indexing/http-network-errors#network-and-dns-errors)\n- [robots.txt rules preventing Googlebot's access to the page](/search/docs/crawling-indexing/robots/intro)\n\nIndexing\n--------\n\n\nAfter a page is crawled, Google tries to understand what the page is about. This stage is\ncalled indexing and it includes processing and analyzing the textual content and key content\ntags and attributes, such as\n[`\u003ctitle\u003e` elements](/search/docs/appearance/title-link)\nand alt attributes,\n[images](/search/docs/appearance/google-images),\n[videos](/search/docs/appearance/video), and\nmore. \n\n\nDuring the indexing process, Google determines if a page is a\n[duplicate of another page on the internet or canonical](/search/docs/crawling-indexing/consolidate-duplicate-urls).\nThe canonical is the page that may be shown in search results. To select the canonical, we\nfirst group together (also known as clustering) the pages that we found on the internet that\nhave similar content, and then we select the one that's most representative of the group. The\nother pages in the group are alternate versions that may be served in different contexts, like\nif the user is searching from a mobile device or they're looking for a very specific page from\nthat cluster.\n\n\nGoogle also collects signals about the canonical page and its contents, which may be used in\nthe next stage, where we serve the page in search results. Some signals include the language\nof the page, the country the content is local to, and the usability of the page.\n\n\nThe collected information about the canonical page and its cluster may be stored in the Google\nindex, a large database hosted on thousands of computers. Indexing isn't guaranteed; not every\npage that Google processes will be indexed.\n\n\nIndexing also depends on the content of the page and its metadata. Some common indexing issues\ncan include:\n\n- [The quality of the content on page is low](/search/docs/essentials)\n- [Robots `meta` rules disallow indexing](/search/docs/crawling-indexing/block-indexing)\n- [The design of the website might make indexing difficult](/search/docs/crawling-indexing/javascript/javascript-seo-basics)\n\nServing search results\n----------------------\n\n| Google doesn't accept payment to rank pages higher, and ranking is done programmatically. [Learn more about ads on Google Search](https://www.google.com/search/howsearchworks/our-approach/ads-on-search/).\n\n\nWhen a user enters a query, our machines search the index for matching pages and return the\nresults we believe are the highest quality and most relevant to the user's query. Relevancy is\ndetermined by hundreds of factors, which could include information such as the user's\nlocation, language, and device (desktop or phone). For example, searching for \"bicycle repair\nshops\" would show different results to a user in Paris than it would to a user in Hong Kong. \n\n\nBased on the user's query the search features that appear on the search results page also\nchange. For example, searching for \"bicycle repair shops\" will likely show local results and\nno [image results](/search/docs/appearance/visual-elements-gallery#image-result),\nhowever searching for \"modern bicycle\" is more likely to show image results, but not local\nresults. You can explore the most common UI elements of Google web search in our\n[Visual Element gallery](/search/docs/appearance/visual-elements-gallery).\n\n\nSearch Console might tell you that a page is indexed, but you don't see it in search results.\nThis might be because:\n\n- [The content on the page is irrelevant to users' queries](/search/docs/fundamentals/seo-starter-guide#expect-search-terms)\n- [The quality of the content is low](/search/docs/essentials)\n- [Robots `meta` rules prevent serving](/search/docs/crawling-indexing/block-indexing)\n\n\nWhile this guide explains how Search works, we are always working on improving our algorithms.\nYou can keep track of these changes by following the\n[Google Search Central blog](/search/blog)."]]