Screaming fast SSD-based web hosting platform for your website.
Speed, security and expert support for WordPress powered websites.
Manage multiple hosting accounts on our web hosting platform.
Simple, Reliable, and Fast Cloud VPS Hosting Environment with High Performace.
Simple, Reliable, and Fast Cloud VPS Hosting Environment
Whatever your website or application needs, we’ve got you covered. Enjoy 100% in-house support, guaranteed performance and uptime, 1-click installs, and a super-intuitive control panel to make managing your websites and projects easy.
If you’ve ever built your own website, you may have heard of a robotx.txt file and wondered, what is this file for? Well, you’re in the right place! Below, we will review what this file is and why it’s important.
First of all, the robots.txt is a nothing more than a plain text file (ASCII or UTF-8) located in your domain root directory, which blocks (or allows) search engines to access certain areas of your site. The robots.txt contains a simple set of commands (or directives) and it’s typically applied in order to restrict crawler traffic onto your server, thus preventing unwanted resource usage.
Search engines use so called crawlers (or bots) in order to index parts of a website and return those as search results. You might want certain sensitive data stored on your server to be inaccessible for web searches. The robots.txt file helps you do just that.
Note: Files or pages on your website are not completely cut off from crawlers in case these files are indexed/referenced from other websites. To properly protect your URL from appearing in Google search engines, you can password-protect the files directly from your server.
In order to create your robots.txt file (if not already existent), simply follow the following steps:
1. Log into your cPanel account
2. Navigate to FILES section and click on File Manager
3. Browse File Manager to the website directory ( e.g public_html ) then Click on “New File” >> Type in “robots.txt” >> Click on “Create New File”.
4. Now you are free to edit the content of this file by double clicking on it.
Note: you can create only one robots.txt file for each domain. Duplicates are not allowed on the same root path. Each domain or sub-domain must contain its own robots.txt file.
Usually, a robots.txt file contains one or more rules, each on its own separate line. Each rule blocks or allows access to a given crawler to a specified file path or the entire website.
Block all crawlers (user-agents) from accessing the logs and ssl directories:
Block all crawlers to index the whole site:
Allow all user agents to access the entire site:
Block indexation for the whole site from a specific crawler:
Allow index to a specific web crawler and prevents indexation from others:
Under User-agent: you can type in the specific crawler name. You can also include all crawlers simply by typing in the star (*) symbol. With this command, you can filter out all crawlers except AdBot crawlers, which you need to enumerate explicitly. You can find a list of all crawlers on the internet.
Additionally, in order for the Allow and Disallow commands to work only for a specific file or folder, you must always include their names between “/”.
Notice how both commands are case-sensitive? It is especially relevant to know, that the crawler agents default setting is so that they can access any page or directory if not blocked by a Disallow: rule.
We are one of the few privately owned, independent web hosting companies nowadays. From a humble beginning we evolved into the first choice for web hosting services of more than 20,000 personal, small-business and large website owners.
Copyright © BrickellHost. All Rights Reserved.