Koozali.org: home of the SME Server
Legacy Forums => Experienced User Forum => Topic started by: Craig Jensen on October 31, 2003, 02:29:46 AM
-
Hi.
I have been studying the options available through my SME 5.6 server to 'control' those who insist on downloading my entire site with their slurping applications. I have noted a few that offer limited help.
I have lately been stopping httpd for a given time and then re-starting, just so I can use a bit of MY OWN bandwidth for awhile :-) (BTW, does anyone else find that this seems to be more and more of a problem?)
Please if you will, enlighten me on methods you have used that efficiently 'choke' these bandwidth hogs...
Thank you for your replies.
Craig Jensen
-
Craig:
For those automatic grabbing tools you may want to make a robots.txt file...
for more info have a look at http://www.searchengineworld.com/robots/robots_tutorial.htm
-
Not a problem here, because I'm not making anything available that anybody would want much of ... no binaries!
-
Hello. We're using a .htaccess in the primary, something like is discussed at this site:
http://www.webmasterworld.com/forum13/687-3-15.htm
We copied 90e-smithAccess10primary to /etc/e-smith/templates-custom/etc/httpd/conf/httpd.conf/90e-smithAccess10primary
and added :
in
Options +FollowSymLinks
and changed:
AllowOverride None to :
AllowOverride All
Expanded and restarted httpd
This allows you to choose which spiders you want to allow, and to redirect those you don't want to a 404 page or another site.
-
Thank you for the responses. I can see use for each option mentioned.
Thanks again
Craig Jensen
-
Kirk Ferguson wrote:
> Hello. We're using a .htaccess in the primary, something
> like is discussed at this site:
There's never any need to use .htaccess, since you can use template fragments.
> This allows you to choose which spiders you want to allow,
> and to redirect those you don't want to a 404 page or another
> site.
You can do that with a custom template fragment. Using .htaccess means that you have your spider information in two places rather than one, and if you are not careful, might open up a security hole.
Charlie
-
Charlie,
Thanks for the help. So I should try to include the commands (or something like them) in a custom template fragment rather than using that .htaccess file?
Should this be part of the httpd.conf section for the primary?
I sure don't need any more security worries, but the server this runs on is a a slow line and seems to attract more spiders than other servers I work with for some reason.
Kirk