SLiCK_303 Posted March 26, 2018 Author Posted March 26, 2018 the rest of robots.txt has it like: Disallow: */classes/ do we do it like that, with the *, or like you said: Disallow: /blackhole/
movieseals Posted March 26, 2018 Posted March 26, 2018 @slick_303 On my version, I did this: Disallow: */blackhole/ Seems to work.
SLiCK_303 Posted March 26, 2018 Author Posted March 26, 2018 @movieseals thats what i did too, since the rest of the disallows were that way
movieseals Posted March 26, 2018 Posted March 26, 2018 @datakick Very nice implementation! Great addition to the TB arsenal!
datakick Posted March 26, 2018 Posted March 26, 2018 @SLiCK_303 @movieseals - robots.txt standard technically does not recognize asterisk as a wildchar. So, if some bot implemented robots.txt handling strictly according to specification, they wouldn't understand your directive and could actually follow the link. Google, Bing, and other major players support this kind of extension to the standard, so they will process it properly. But I think it's safer to use Disallow: /blackhole/, or even Disallow: /blackhole to be sure no bot with good intentions will be caught in this trap. Basically, this directive means that any url starting with /blackhole/ is prohibited from browsing. Your (nonstandard) directive says that any url containing /blackhole/ is prohibited.
movieseals Posted March 26, 2018 Posted March 26, 2018 @datakick said in Spam shopping carts: Disallow: /blackhole/ Thank you for the clarification. Makes sense!
datakick Posted April 4, 2018 Posted April 4, 2018 just FYI, I've added this module to the store. I'm not going to create a separate forum thread for this module, but if anyone has feature requests feel free to create one
Recommended Posts
Create an account or sign in to comment
You need to be a member in order to leave a comment
Create an account
Sign up for a new account in our community. It's easy!
Register a new accountSign in
Already have an account? Sign in here.
Sign In Now