7 On-Page SEO Checks That You Need To Conduct On A Periodic Basis

on-page seo checkup
  • Save

Regarding SEO, we know that link building is an ongoing process, but more often than not, we tend to neglect the on-page SEO aspects.

Site updates, theme updates/changes, plugin updates, adding a new plugin/functionality, and other changes like updating a file using FTP can cause accidental errors that could lead to on-page SEO issues. Unless you proactively look for these errors, they will go unnoticed and will negatively influence your organic rankings.

For instance, I recently realized that I had been blocking out images in one of my blogs for almost 6 months because of an old and neglected Robots.txt file. Imagine the impact such a mistake could have on your rankings!

On-Page SEO Checkup

Keeping the importance of SEO in mind, here are 7 important checks you need to conduct periodically to ensure that your on-page SEO is on point.

Note: Even though these checks are for people running a WordPress blog, they can be used for any blogger on any platform.

Pages with broken links (be it internal or external links) can potentially lose rankings in search results. Even if you do have control over internal links, you do not have control over external links.

There is a huge possibility that a webpage or resource that you linked to no longer exists or has been moved to a different URL, resulting in a broken link.

This is why it is recommended to check for broken links periodically.

There are many ways to check for broken links, but one of the easiest and most efficient ways is with the ScreamingFrog SEO Software.

To find broken links on your site using ScreamingFrog, enter your domain URL in the space provided and click the “Start” button. Once the crawling is complete, select the Response Codes tab and filter your results based on “Client Error (4xx)”. You should now be able to see all links that are broken.

Click on each broken link and then select the Inlinks tab to see which page(s) actually contain this broken link. (Refer to image below.)

Screamingfrog links checker
  • Save

If you are using WordPress, you can also use a plugin like the Broken Link Checker. This plugin will find and fix all broken links.

Another way to check for broken links is through the Google Search Console. Log in and go to Crawl > Crawl Errors and check for “404” and “not found” errors under the URL Errors section.

If you do find 404 URLs, click on the URL and then go to the Linked From tab to see which page(s) contain this broken URL.

2. Find and delete low-value content

The command operator site:sitename.com” displays all pages on your site indexed by Google.

  • Save

By roughly scanning through these results, you should be able to check if all pages indexed are of good quality or if there are some low-value pages present.

Quick Tip: If your site has a lot of pages, change the Google Search settings to display 100 results at a time. This way you can easily scan through all results quickly.

An example of a low-value page would be the ‘search result’ page. You might have a search box on your site, and there is a possibility that all search result pages are being crawled and indexed. All these pages contain nothing but links, and hence are of little to no value. It is best to keep these pages from getting indexed.

Another example would be the presence of multiple versions of the same page in the index. This can happen if you run an online store and your search results have the option of being sorted.

Here’s an example of multiple versions of the same search page:

  • http://sitename.com/products/search?q=chairs
  • http://sitename.com/products/search?q=chairs&sort=price&dir=asc
  • http://sitename.com/products/search?q=chairs&sort=price&dir=desc
  • http://sitename.com/products/search?q=chairs&sort=latest&dir=asc
  • http://sitename.com/products/search?q=chairs&sort=latest&dir=desc

You can easily exclude such pages from being indexed by disallowing them in Robots.txt, or by using the Robots meta tag. You can also block certain URL parameters from getting crawled using the Google Search Console by going to Crawl > URL Parameters.

3. Check Robots.txt to see if you are blocking important resources.

When using a CMS like WordPress, it is easy to accidentally block out important content like images, javascript, CSS, and other resources that can actually help the Google bots better access/analyze your website.

For example, blocking out the wp-content folder in your Robots.txt would mean blocking out images from getting crawled. If the Google bots cannot access the images on your site, your potential to rank higher because of these images reduces. Similarly, your images will not be accessible through Google Image Search, further reducing your organic traffic.

In the same way, if Google bots cannot access the javascript or CSS on your site, they cannot determine if your site is responsive or not. So even if your site is responsive, Google will think it is not, and as a result, your site will not rank well in mobile search results.

To find out if you are blocking out important resources, log in to your Google Search Console and go to Google Index > Blocked Resources. Here you should be able to see all the resources that you are blocking. You can then unblock these resources using Robots.txt (or through .htaccess if need be).

For example, let’s say you are blocking the following two resources:

  • /wp-content/uploads/2022/01/image.jpg
  • /wp-includes/js/wp-embed.min.js

You can unblock these resources by adding the following to your Robots.txt file:

  • Allow: /wp-includes/js/
  • Allow: /wp-content/uploads/

To double check if these resources are now crawlable, go to Crawl > Robots.txt tester in your Google Search console, then enter the URL in the space provided and click “Test”.

  • Save

4. Check the HTML source of your important posts and pages

It’s one thing to use SEO plugins to optimize your site, and it’s another thing to ensure they are working properly. The HTML source is the best way to ensure that all of your SEO-based meta tags are being added to the right pages. It’s also the best way to check for errors that need to be fixed.

If you are using a WordPress blog, you only need to check the following pages (in most cases):

  1. Homepage/Frontpage (+ one paginated page if homepage pagination is present)
  2. Any single posts page
  3. One of each archive pages (first page and a few paginated pages)
  4. Media attachment page
  5. Other pages – if you have custom post pages

As indicated, you only need to check the source of one or two of each of these pages to make sure everything is right.

To check the source, do the following:

  1. Open the page that needs to be checked in your browser window.
  2. Press CTRL + U on your keyboard to bring up the page source, or right-click on the page and select “View Source”.
  3. Now check the content within the ‘head’ tags ( <head> </head> ) to ensure everything is right.
  • Save

Here are a few checks that you can perform:

  1. Check to see if the pages have multiple instances of the same meta tag, like the title or meta description tag. This can happen when a plugin and theme both insert the same meta tag into the header.
  2. Check to see if the page has a meta robots tag, and ensure that it is set up properly. In other words, check to make sure that the robots tag is not accidentally set to Noindex or Nofollow for important pages. And make sure that it is indeed set to Noindex for low value pages.
  3. If it is a paginated page, check if you have proper rel=”next” and rel=”prev” meta tags.
  4. Check to see if pages (especially single post pages and the homepage) have proper OG tags (especially the “OG Image” tag), Twitter cards, other social media meta tags, and other tags like Schema.org tags (if you are using them).
  5. Check to see if the page has a rel=”canonical” tag and make sure that it is showing the proper canonical URL.
  6. Check if the pages have a viewport meta tag. (This tag is important for mobile responsiveness.)

5. Check for mobile usability errors.

Sites that are not responsive do not rank well in Google’s mobile search results. Even though your site is responsive, there is no saying what Google bots will think. Even a small change like blocking a resource can make your responsive site look unresponsive in Google’s view.

  • Save

So even if you think your site is responsive, make it a practice to check if your pages are mobile friendly or if they have mobile usability errors.

To do this, log in to your Google Search Console and go to Search Traffic > Mobile Usability to check if any of these pages show mobile usability errors.

You can also use the Google mobile friendly test to check individual pages.

6. Check for render blocking scripts.

You might have added a new plugin or functionality to your blog which could have added calls to many javascript and CSS files on all pages of your site. The plugin’s functionality might be for a single page, yet calls to its javascript and CSS are on all pages.

For example, you might have added a contact form plugin that only works on one page – your contact page. But the plugin might have added its Javascript files on every page.

The more javascript and CSS references a page has, the longer it takes to load. This reduces your page speed which can negatively impact your search engine rankings.

The best way to ensure this does not happen is to check your site’s article pages using Google’s PageSpeed Insights tool on a regular basis. Check to see if there are render-blocking Javascript files and figure out if these scripts are needed for the page to function properly.

  • Save

If you find unwanted scripts, restrict these scripts only to pages that require them so they don’t load where they are not wanted. You can also consider adding a defer or async attribute to Javascript files.

7. Check and monitor site downtimes.

Frequent downtimes not only drive visitors away, they also hurt your SEO. This is why it is imperative to monitor your site’s uptime on a constant basis.

There are a host of free and paid services like Uptime Robot, Jetpack Monitor, Pingdom, Montastic, AreMySitesUp, and Site24x7 that can help you do just that. Most of these services will send you an email or even a mobile notification to inform you of site downtimes. Some services also send you a monthly report of how your site performed.

If you find that your site experiences frequent downtimes, it is time to consider changing your web host.

Things To Monitor For Proper On-Page SEO

These are some very important things to check for so that your on-page SEO stays highly optimized.

Conducting these checks on a regular basis will ensure that your on-site SEO is on point and that your rankings are not getting hurt without your knowledge.

Let me know what kinds of things you check for when doing an SEO site audit. What do you do to make sure your on-page SEO stays optimized? Let me know in the comments below!

Here are some more SEO articles to check out:

Like this post? Don’t forget to share it!

Subscribe on YouTube

  • Save
Authored By
Mukesh is a Blogger, Wordpress enthusiast, SEO professional and a Nethead. Feel free to visit my blog OrbitingWeb.com to connect with me.

42 thoughts on “7 On-Page SEO Checks That You Need To Conduct On A Periodic Basis”

  1. Gagan Kamboj

    thanks for the article as a new blogger i didnt know about the robots.txt blocking. Now i know the issue and will solve it.

    Keep Posting

  2. Rakhi

    Hello, I am new to SEO and learning new things about onpage seo and offpage seo. This post is really very informative for me. I learn so many new things from you. thank you so much for this.

  3. Chibuzor Aguwa

    Thanks for this information, i will definitely run a check on my site today.

  4. praveenrajarao

    Hello Mukesh,

    Those are some awesome tips and you have also explained in detail how to go about fixing issues that might affect the page performance and indexing.
    I will bookmark this article for further study.

  5. Nestor Paul Demoni

    Great post Mukesh,

    Really nice and well appreciated. As a beginner in SEO. This is very usefuk who wanted to optimized their search engine visibility.

  6. arvind

    i thinks broken link checker is the best plugin to keep your wordpress blog away from from dead links…i have been using both blogger and wordpress but this plugin takes care of all the broken links in wordpress which decreases lot of my work on SEO as compared to blogger….nice information for beginners in blogging..thanks for sharing.!!!

  7. TAP

    Thanks for sharing this article.. It will help my website to improve SEO score.

  8. Shrikanth

    Hi Mukesh,

    Really nice article. As a beginner in SEO it helped me lot to learn new things. Thanks a lot and keep it up.

  9. Pawan Sahani

    I use Broken Link Checker Plugin to find broken links and also I check Crawl error everyday.
    To check downtime I use Jetpack plugin.
    But, my blog speed is slower on mobile. I’m trying to improve the blog speed.
    Thanks for sharing such a wonderful article.

  10. Shafi Khan

    Thanks for the amazing on-page SEO checks. My site has got slower over the period, and this article has helped me boost the loading time.

    I’ll make sure I keep making these changes periodically.

  11. vijay

    How can i fix the below error of my blog to improve the speed

    Eliminate render-blocking JavaScript and CSS in above-the-fold content
    Leverage browser caching
    Enable compression
    Prioritize visible content

    1. Mukesh Mani

      Hi Vijay, reduce/deactivate plugins that you are not using and use a cache plugin like WP Super Cache. This will solve most of your issues.

  12. Ajay Malik

    Also Add this, Check for Short, long or missing description

    1. Mukesh Mani

      That’s a good point. You can check for short/missing/duplicate title and meta tags from your Google search console. Just go to Search Appearance > HTML Improvements

  13. Akshay S Manoj

    I make the meta robots noindex in categories and tags ,but it is Indexed in my site link.Some images also indexed in site link.Please help me to fix this issue.

    1. Mukesh Mani

      No worries Akshay. Google will eventually remove them from index.

      1. Akshay S Manoj

        Thank You For Your Reply

  14. anand manni

    I took a WAG and deleted the 404 log since I was fairly confident that the real urls needing redirecting had been done. I find this plugin confusing and it’s possibly because I don’t have a deep understanding yet of the entire process.

    I use Analytics, GTMetrix, Screaming Frog, Yoast SEO plugin and other reliable programs to test various aspects of my site. While I certainly want and need to make sure any incorrect urls are redirected, I certainly don’t want to redirect unnecessarily.

    Do you have any thoughts and/or suggestions about this?

    Also, since Broken Link Checker seems to be a resource guzzler, can it be used on a periodic basis to check links vs. being activated full-time?

    Thanks so much for your assistance.

    1. Harsh Agrawal

      You can use WP link status plugin as an alternative to WP broken link checker. This one is friendly for shared hosting & doesn’t consume lot of resources.

  15. Uchchash

    I always use site:example.com to find the unwanted indexed pages! As my blog covers only a single topic, I don’t have tons of posts! So, its easier for me to find if something is wrong!

  16. Ngan

    For point 2, the correct method is to use the meta NoIndex. Blocking the parameter in Robots.txt will not de-index the page from Google. It merely block Google Bot from crawling it, but the page is still part of the index.

    1. Mukesh Mani

      That’s a good point NGAN.

  17. Pronob

    This was new to me.It will help me definitely in near future.Thanks in Advance. 🙂

  18. Mohan Nayak

    Excellent Tips Mukesh,
    When it comes to SEO, most bloggers just focus on Content creation and link building, but you have shown that how broken link and an unresponsive page will hurt SEO.
    I would recommend this to everyone to check at least once a while.
    Nice work, thanks for sharing.

    1. Mukesh Mani

      Thank you Mohan, Glad you found it useful.

  19. Vishwajeet Kumar

    Hello Mukesh.

    Thanks for sharing these checklist of On page SEO. I pay close attention to my broken links and site down times. If something went wrong, I will try to resolve it as soon as possible. For site downtime monitor I am using Jetpack monitors. Works great for me.

  20. Lucky

    Hello Mukesh, My site is taking too much time to load as i have checked in the Google PageSpeed Insight. And i found the same as you have mentioned in the subtitle 6. What will happen if i remove render blocking javaScript and css page?? will my site appearance will change or what.?? I am new to SEO so i am confused. How can i remove the specific javaScript and CSS.?? Help me out.

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top