
Originally Posted by
djr33
It doesn't matter what files your server actually has. To a visitor (including any bots) the URL is correct, even if, technically, those files don't exist.
My point is something else:
If you have a link to 1/2/3/ and that works, then the bot will index it.
1/2/ may not exist. But that is NOT a problem unless the bot actually tries to go there.
As I said, bots don't guess about URLs. So the bot will NEVER try to go to 1/2/ UNLESS you have a link to it.
You won't need a site map for this. Search engines now are smarter than that, although adding a site map also won't hurt anything (and might help visitors).
Here's the basic summary:
Bots search by following all links. Any page that is linked to on your site will eventually be visited by bots.
Bots will never visit pages that are never linked to. This includes pages that don't exist.
If there is a page that does not exist, but it IS linked to, then you might have a problem.
All of that is unrelated to .htaccess, which can do fancy things like that with the URL, and it can be helpful, if confusing to manage. But as far as SEO is relevant, those pages DO exist (just as much as they would if they were real folders and files).
You can think of a much simpler situation:
1. Create a new directory called "/test/" on your web host.
2. Create a file called "/test/test.htm"
Now, if a search engine visits that page, it will index it. Remember that it would need to be linked to from somewhere (for example, your site's homepage, or a different website).
But the search engine will not try to access "/test/" which does "not exist" in the sense that there is nothing to load there and you will get some sort of error (either 403 "forbidden" or 404 "not found").
There's nothing different from the SEO point of view about the current situation, it's just more complicated.
Bookmarks