By Matthew Edgar · Last Updated: April 21, 2023
Before you can do anything else with SEO, Google and Bing need to find your pages. This process of finding pages is called “discoverability”, during which a bot needs to find all the various pages contained on your website. There are a few different methods you can use to encourage bots to find all the pages—and a few methods that will prevent bots from finding pages.
Internal Links: The Importance of Hyperlinks
The hyperlink is the most basic means of helping robots find the pages that exist on your website. These links are included in the href attribute contained in the standard HTML anchor tag (<a>). In this example, the HREF attribute references Elementive’s home page (https://www.elementive.com). The anchor text is “Elementive” and describes what page is being linked to.
This tag + attribute combination was part of the original HTML specification. All these years later, this tag is still the main way that bots (and humans) discover the pages on a website. When robots find this link within a web page’s HTML, they extract the URL within the HREF attribute and add that URL to their database. New URLs found within an HREF attribute will be added to a crawl queue.
To increase the discoverability of your website’s pages, the first and most basic step is adding more links to the pages on your website. Links from one page of your website to another page of your website are referred to as internal links. The more internal links there are referencing a page on your website, the better the chances robots will discover the pages being linked to.
Be careful to not put internal links just anywhere. You want to make sure internal links are relevant for users. Search robots are sophisticated enough to determine which links are being placed just for their benefit.
Also, be mindful of where links are placed on the website. Putting links in the navigation or the footer won’t help discoverability nearly as much as putting links within the main content of your pages.
You can also add a rel attribute to hyperlinks. This allows for greater control in explaining what the link does. These types of qualifiers allow Google’s bots to better understand your links and the relationship between the pages being linked from and to. With that information, Google can handle the links appropriately after discovering them. With the attribute rel=”sponsored”, you are communicating that the link in question has some type of monetary relationship attached. With rel=”ugc”, you are communicating the link in question was added by users and not the site owners.
OnClick Links in Buttons or Other Tags
However, you can use the onclick attribute to create a link. By adding an onclick attribute to HTML elements like <p> or <button> (or other tags), you can make any element a link. For example, this button would take somebody to Elementive’s home page:
<button onclick="window.location.href = 'https://www.elementive.com';">Elementive</button>
Unfortunately, this does not work as effectively as a link because robots do not trigger the onclick event. If a button is your only link pointing to a page on your website, that page will have trouble being discovered.
If you do use or have to use the button onclick (or a p onclick, etc.) for some links on your website, make sure you also have regular hyperlinks with an HREF attribute linking to those pages as well.
Alternatively, consider restyling a link to look like it is a button. One route is you can style a traditional link like a button using CSS. Or, you can wrap the <button> in an <a>:
The final method we’ll discuss here is that you can support page discoverability with the XML sitemap. An XML sitemap is a list of all pages contained on your website that you want Google to find and crawl. While beneficial for nearly every website, XML sitemaps are a particularly helpful solution for websites that can’t easily add internal links to every page. Maybe there are too many pages to link to or maybe there are technical constraints prohibiting the inclusion of links within the website’s main content. The XML sitemap ensures Google’s bots are able to see all the pages.
When it comes to the XML sitemap, you want to make sure the pages listed in the sitemap are an accurate portrayal of your website’s pages. You don’t want to list error pages, duplicate pages, or old URLs that redirect elsewhere since those pages aren’t ones you want Google to discover. In the same way, you don’t want to list pages that you’ve disallowed or noindexed because those aren’t pages that Google needs to discover.
Check If Google Has Discovered A New Page
The best way to tell if Google has discovered a URL is by inspecting the URL in Google Search Console. The inspection result will tell you if Google has not already discovered the URL. From the inspection page, you can request indexing on a URL to encourage Google to crawl the page. For more details, check out my video about using the URL inspection tool.
How Many Pages Is Google Discovering?
You can also monitor discoverability in Google Search Console using Google’s Crawl Stats report. In the “By purpose” table included in the crawl stats data, Google states how many crawls were made for the purpose of discovering new content (you can also see how many crawls were for the purpose of refreshing content). If you click on “Discovery” in this table, you will see a list of pages that were discovered in Google’s recent crawls of the website.