One of my assignments this week was to read through the Google Webmaster Guidelines and summarize their general recommendations.
-
Make sure your pages can get links from other pages. You will want to make sure you use crawlable links when you do that, which means using anchor tags (<a> with a href attribute).
-
Create a sitemap and ensure that it has links to all the important pages on the site. They also recommend having this as a “human-readable” list of links for those important pages.
-
Keep the number of links on a page to a “reasonable number”. They recommend “a few thousand at most”, which, to me, seems excessive.
-
Ensure that the hosting server supports the “If-Modified-Since” HTTP header, which is what informs the Googlebots when content has been changed since they last visited the page. It’s important as it saves bandwidth and, thus, networking overhead costs.
-
Use the robots.txt file well. It is important to ensure that the web crawler bots are not crawling non-important pages. Besides ensuring the robots.txt file has the right information, you also want to make sure that it is kept up-to-date. Doing these things ensures that your “crawling budget” is utilized well.
-
Another thing that you can do is manually submit your site to Google’s crawlers. This triggers them to head over to your site immediately, as opposed to waiting for it to be discovered crawling other sites.
-
Lastly, you should make sure that administrators of other relevant sites know about yours. Emailing an announcement, or posting such on social media is a great tool towards creating awareness of your site and garnering backlinks.