Author |
Message |
hinksta
Worker


Joined: Dec 23, 2005
Posts: 226
Location: UK
|
Posted:
Mon Mar 12, 2007 7:09 am |
|
First of all Well Done to Raven and the team for all the hard work dragging phpnuke into the 21st century.
Just a couple of questions regarding robots.txt
What is Crawl-delay: 5 ?
and should ShortLinks be included? |
|
|
|
 |
jakec
Site Admin

Joined: Feb 06, 2006
Posts: 3048
Location: United Kingdom
|
Posted:
Mon Mar 12, 2007 7:20 am |
|
Crawl delay should slow down the crawling of bots on your site. Although I do recall reading somewhere that Google ignores this option, but if you have your site registered with Google you can specify a delay there instead.
I'm at work at the moment so I do have any files to look at, but I guess bots don't necessarily need to crawl the shortlinks directory so you could include if you wanted to. Maybe Montego can confirm/clarify this?
I don't think is really a bug as such, so I have moved the post.
Jakec |
|
|
|
 |
kguske
Site Admin

Joined: Jun 04, 2004
Posts: 6437
|
Posted:
Mon Mar 12, 2007 10:56 am |
|
They should not need to crawl the shortlinks directory, and it would be better if this is added to the robots.txt file. |
_________________ I search, therefore I exist...
Only registered users can see links on this board! Get registered or login! |
|
|
 |
montego
Site Admin

Joined: Aug 29, 2004
Posts: 9457
Location: Arizona
|
Posted:
Tue Mar 13, 2007 6:22 am |
|
I agree. My oversight. Sorry about that. |
_________________ Only registered users can see links on this board! Get registered or login!
Only registered users can see links on this board! Get registered or login! |
|
|
 |
kguske

|
Posted:
Tue Mar 13, 2007 6:32 am |
|
Not a big deal. I don't think it will really hurt anything. |
|
|
|
 |
Susann
Moderator

Joined: Dec 19, 2004
Posts: 3191
Location: Germany:Moderator German NukeSentinel Support
|
Posted:
Tue Mar 20, 2007 11:57 am |
|
Did we ever added Audioslaves Google Tap folder into the robots.txt ?
There isnĀ“t any instruction where ist says you should add this folder to the robots.txt.
In my opionion its not required because I never saw a bot in the past years indexing the Google Tap folder from my website. |
|
|
|
 |
montego

|
Posted:
Wed Mar 21, 2007 5:38 am |
|
Do not know Susann. Just thought that it could not hurt is all. |
|
|
|
 |
bugsTHoR
Involved


Joined: Apr 05, 2006
Posts: 263
|
Posted:
Wed Apr 04, 2007 12:36 pm |
|
i believe you have to enter this to stop googlebot indexing your site
in the robots file at the /Root
User-agent: Googlebot
Disallow: / |
_________________ LUV RAVEN DISTROBUTION BEBE
Clanthemes.com are great (free advertisements for now until i get to 20,000 posts LoL) |
|
|
 |
jakec

|
Posted:
Wed Apr 04, 2007 12:47 pm |
|
Why would you want to stop Googlebot from indexing your site, unless you never want it appear in the Google search?  |
|
|
|
 |
bugsTHoR

|
Posted:
Wed Apr 04, 2007 12:55 pm |
|
jakec wrote
Quote: | Although I do recall reading somewhere that Google ignores this option | in reply to first post
so thought i would add that you need this to stop googlebot
someone might want it. never know |
|
|
|
 |
jakec

|
Posted:
Wed Apr 04, 2007 2:06 pm |
|
OK I see what you are saying.
My point was that although Googlebot doesn't recognise the crawl delay in the robots.txt files, you can apply a delay if your site is registered with Google.
If you block Googlebot it is unlikely that people will find your website. Maybe that is what you want, if it is a private site. |
|
|
|
 |
bugsTHoR

|
Posted:
Wed Apr 04, 2007 5:26 pm |
|
i have seen some guys /ladies on here have private sites (friends/ family)so it would be usefull to them.
i can see if you have a really busy site it would also be very usfull to with other indexing robots or what not on the net.
i do think its usefull as the google site i seen, on googlebot it has more than 10,000,000 robots out there. |
|
|
|
 |
jakec

|
Posted:
Thu Apr 05, 2007 2:44 am |
|
Like I said, I understand what you are saying, and I too have some personal sites that I don't really want indexed, but also I'm not bothered if people did visit them. So blocking good bots in this way is one solution.
...but the original question was:
hinksta wrote: | What is Crawl-delay: 5 ?
and should ShortLinks be included? |
|
|
|
|
 |
Guardian2003
Site Admin

Joined: Aug 28, 2003
Posts: 6799
Location: Ha Noi, Viet Nam
|
Posted:
Thu Apr 05, 2007 2:06 pm |
|
Going off topic slightly but if you want to stop all bots (that adhere to the robots.txt) then you can simply useCode:
User-agent: *
Disallow: /
|
As far as I am aware, the Crawl-delay only works with Slurp (at least Yahoo claim its a directive specifically for their bots).
http://help.yahoo.com/help/us/ysearch/slurp/slurp-03.html |
|
|
|
 |
|