SEO Tools - robots, sitemap, rss
In SEO, there are several additional tools used to effectively introduce a website to search engines.
In this lesson, we will explore commonly used SEO tools: robots.txt
, sitemap
, and rss
.
robots.txt
robots.txt
is a text file located on your website that informs search engines which pages should or should not be crawled.
User-agent: *
Disallow: /private/
Allow: /public/
The above robots.txt
instructs all search engines not to access the /private/
directory while allowing access to the /public/
directory.
Sitemap
A sitemap
is a file that lists a website's pages and shows the relationships between them.
Sitemaps function like a map by informing search engines about the site's structure and the links within web pages.
<?xml version="1.0" encoding="UTF-8"?>
<urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9">
<url>
<loc>https://example.com/page1</loc>
<lastmod>2023-10-15</lastmod>
<changefreq>daily</changefreq>
</url>
<url>
<loc>https://example.com/page2</loc>
</url>
</urlset>
The sitemap example above lists two web pages, with page1
showing both the last modification date and change frequency.
RSS
rss
is a file that automatically informs subscribers about content updates on a website.
It is mostly used for blogs, newsletters, and web services.
<?xml version="1.0"?>
<rss version="2.0">
<channel>
<title>My Website News</title>
<link>https://example.com</link>
<description>Latest updates from My Website</description>
<item>
<title>New Article</title>
<link>https://example.com/new-article</link>
<description>This is a new article.</description>
</item>
</channel>
</rss>
In this example, we introduce updates under the title My Website News
, including content like New Article
.
Want to learn more?
Join CodeFriends Plus membership or enroll in a course to start your journey.