I was wondering what is the standard practice for a rails app robots.txt file. i.e. which folders are generally prevented from the robots tracking.
My current robots.txt file is:
# User-agent: *
# Disallow: /
Disallow: /public/uploads/
Sitemap: www.mysite.co.za/sitemap.xml
My question is, do most people disallow /public/uploads from being tracked?
I also have a number of models/pages in which only the Admin user can access and perform the CRUD operations. This is protected by devise. I was wondering if it is necessary for these to be disallowed in the robots.txt file, and whether the spider is even able to index these pages (because they are protected for admin use only).
e.g. I have a category model to which only the admin can do CRUD on. Should I add:
Disallow: /categories/
(or is it with the *)
Disallow: /categories/*
These are all my queries around robots.txt usage in rails. Does this make sense?
Thanks,
Matt