Advanced SEO Techniques

Controlling the Robots

Overview
Structured Data
Structured Data Overview
10:18
501
Adding JSON-LD Structured Data
16:29
502
Site Name JSON-LD
503
Articles, News & Blog Post JSON-LD
504
Site Search JSON-LD
506
Client Reviews JSON-LD
506
Breadcrumbs JSON-LD
508
FAQ Page JSON-LD
509
Q&A Page JSON-LD
510
Twitter Cards
601
Embedding Rich Text in JSON
702
Validating & Debugging JSON-LD
703
Technical SEO
Controlling the Robots
801
Sitemap.xml
802
robots.txt
803
Removing Pages from Google
804
Removing Your WEBFLOW.IO Staged Site from Google
805
More Advanced SEO Techniques
810
Voice Search
810
Partially-Dynamic Headings
3:35
811
Full Site SEO Text Search
3:35
812
Influencing Google Search Appearance
Influencing Google's Search Appearance
901
No items found.
Published
Updated
November 4, 2022
in lightbox

"Robots" is an affectionate name given to the search engine site crawlers that explore and index the Internet to create search engine results.

Sometimes, there are pages you don't want Google to index, and there are mechanisms that allow you to do this.

Hiding Static Pages from Google

Webflow has added a static page setting called Sitemap Indexing. When this is toggled off, Webflow;

  • Excludes the page from your sitemap.xml
  • Adds a META noindex tag to the page, so robots know to ignore it for indexing

This feature is found just beneath your title and description settings.

You must have an active site plan on your site in order to toggle this setting. Therefore, you'll want to configure this aspect of your SEO after the site plan has been purchased.

Be careful not to confuse this with Site Search settings...

Don't confuse the Sitemap Indexing feature with Site search settings, which affects Webflow's internal Site Search indexing only.

Hiding Collection Pages from Google

Unfortunately, Collection Pages do not have this setting available.

There you can hide either a specific Collection Page, or all pages in a template, using these approaches.

Hiding all Collection Pages from a Template

Let's suppose you have a News collection, and you want to hide all pages that are generated from Google.

To do that, you can place a special META tag in the <head> custom code of that template page.

This is the tag you need;

<meta name="robots" content="noindex">

Hiding an Individual Collection Page

This same idea can be extended to allow you to hide individual collection items.

  • Add an option field to your CMS collection
  • Give it two values, index and noindex
  • Populate your items with the values you want
    • index means it will appear in SERPS
    • noindex means it will be suppressed from SERPS
  • Make that field required

Then, in your collection page template's HEAD custom code area, drop in the META tag-

<meta name="robots" content="">

Inside of the content attribute, between the double-quotes, insert your new option field.

You can now easily control each page's Google indexing individually.

What about robots.txt?

People often imagine that robots.txt is the answer to their Googlebot-exclusion needs, but it's typically not the right answer. Here are a few reasons why;

  • It's easy to mess up robots.txt and break your site's SEO entirely
  • Robots.txt tells bots what they are allowed to look at, and not what should be indexed. You may well see that page in SERPS, but with no title or description, just a URL.
  • If the page has already been indexed, and you add it to robots.txt, the bot will be prevented from re-visiting it, and will never de-index it even if you add the META noindex to the page. It's not allowed to look at the page, therefore it can't see your META, therefore it doesn't act on it.

FAQs

Answers to frequently asked questions.

Videos
No items found.
Table of Contents
Comments
Did we just make your life better?
Passion drives our long hours and late nights supporting the Webflow community. Click the button to show your love.