I need to get a pattern matching rule to get this results.
allow /dir/path_name.htm/something
disallow /dir/path_name/something
and disallow /dir/path_name.htm
Actually those two disallows are typos accumulated all along. Those pages never exist. How to stop google crawling them never again?
I tested here: http://www.frobee.com/robots-txt-check/ with the following, but seems nothing working.
Allow: /dir/*.htm/?*
Disallow: /dir/*
What went wrong? Thank you.