Tallest Tree SEO Podcast
Cord & Einar discuss how to optimize your website for web crawlers. How do things like robots.txt, site speed, HTTP codes, URLs, links, redirects, XML sitemaps, PDFs vs web pages, robots meta tags, and canonical URLs all affect web crawlers? How can you optimize these aspects of your site to make sure all the content you want robots to discover can be crawled and indexed? Sources Cited: * Google Robots Testing Tool [https://www.google.com/webmasters/tools/robots-testing-tool] * Merkle robots.txt Validator [https://technicalseo.com/tools/robots-txt/] * PageSpeed Insights [https://pagespeed.web.dev/] * Lighthouse [https://developer.chrome.com/docs/lighthouse/overview/] * Pingdom Speed Test [https://tools.pingdom.com/] * Robots Meta Tags [https://developers.google.com/search/docs/crawling-indexing/robots-meta-tag]
30 episodes
Comments
0Be the first to comment
Sign up now and become a member of the Tallest Tree SEO Podcast community!