Website architecture is one of the most important aspects of creating a search engine friendly website. Below are just a few questions I was asked recently on the topic of navigation, site structure, site maps and pages site.
If I have a relatively small [site] with a flat/linear file structure and each page has links on it to every other page that the spiders can follow does it benefit me at all to have a site map?
If you only have a five to ten page site where every page is of equal value to the rest then a site map is unnecessary. However once you get beyond that, or begin having pages that are sub-pages of a section of your site it's a good idea to create a site map. Even if all the pages are represented in the navigation the site map will help the user know what information your site contains without having to go look for it.
One primary benefit of the site map is that it gives the visitor a birds-eye view of the entire site. In most instances, the navigation itself won't accomplish this, though it can depending on how it's laid out. But even still, if the visitor looking for a site map does so because they perceive an inability to find something that your site should have. It provides them with a complete picture. They know if they cannot find it on the site map then there is no need to keep looking through the rest of the site.
If I have a relatively small site is there any benefit for ranking by having a silo or pyramid structure rather than a flat structure? Do the spiders always prefer silo structures or does it only make a difference if the site if it is over a certain number of pages and if so what is that number?
I think in this case you defer to the visitor over the search engine. Organize your pages in a way that makes sense and will help the visitor easily find the information they are looking for. There is nothing wrong with silo-ing your site structure, even for small sites. Just make sure you're not adding additional layers for the sake of adding layers. Create a structure that makes sense from the visitor perspective. If you do that then it will make sense from the search engine perspective as well.
Do the robots really produce a 404 error if they can't find a robot text file on my site and should I make one just for that reason even if its a small site that otherwise wouldn't need it?
Anytime a search engine comes to your site (at least those who honor the robots.txt protocol) will search for the robots.txt file. Generally, not having a robots.txt file won't hurt you in any way but I do recommend adding one anyway.
Without the file in place the search engines can interpret that any way they want. Mostly it will tell them that your site is either completely open or forbidden. If you have the file in place you are letting the engines know that you have made a definitive declaration of what they can and cannot do in regards to spidering the site.
I have heard that pages should not be more that 100 or 150 k. Is that true? Does it affect ranking besides obviously affecting download times? Is k the same as kb (I know I could probably figure that one out)? How can I tell how big my pages are?
Its always best to keep your code as streamlined as possible. I don't think there is a hard and fast rule anymore as to how big a page can be, however if the engines are finding pages that have significantly more code than content then it very well may have a negative effect. Large pages with lots of content generally won't be a problem, but if the content to code ration is skewed too far toward code, then that represents an issue that's worth fixing. If too many of these pages exist on your site then its highly likely that many of your pages won't get fully spidered and/or many pages of your site could be left out of the index.
Typically when referring to file size someone might say 150k when they mean 150 kb. However if they say it's 150k kb then that suggests its 150 thousand kb. Just about any web file manager such as dreamweaver can tell you how big the page is. If the file is saved on your computer you can also look at the file size through windows explorer. When considering page size, search engines generally are just looking at the code, not necessarily at any additional downloadable items on the page (images, flash files, etc.)
I use iweb to make my site and it is limited in some things that it can do. It uses iframes to inset html content (or you can mess with it yourself). The way I understand it the frames are actually an entirely separate page that is imposed on top of the page you put it on. I have heard frames are bad for search rankings are these frames the same as iframes.
I don't have any experience with iweb so I can't answer any question in regards to that. Traditionally frames are not a good way to go for a site, but there are many ways to have the benefit of frames, without actually having frames. For one, if you need separate scrolling areas, this can be accomplished with CSS. For another, if you want to have one file for navigation instead of putting the same navigation code on every page, the use of server side includes (SSI) is the way to go. Includes allow you to create one file then globally include that file throughout the site. Update that one file and every page that pulls the include file in shows the updated information.
These questions are just a drop in the bucket of what you need to know regarding site architecture and navigation. But I do hope you they provided you with some valuable insight. I have answered a number of other website architecture questions in the past that will give you even more information on this topic. You can also check out my architecture checklist.
0 comments:
Post a Comment