A great website design is important to any business trying to find success online, but if a website doesn’t have a chance at good search rankings, very few people may ever get to see it. To avoid having a website that is not search engine friendly, you simply need to take into consideration some basic SEO principles and good content development practices.
As a web designer, I put a lot of thought into what issues I have encountered and issues I have heard others deal with when building websites with SEO in mind. Here are 10 great SEO tips that can lead you to a more search engine friendly web design without sacrificing your creativity and style.
1. Make Sure Your Site Navigation Is Search Engine Friendly
Using Flash for navigation on your website can be bad news if you aren’t aware of how to make Flash objects accessible and web-crawler-friendly. Search engines have a really tough time crawling a website that uses Flash.
2. Place Scripts Outside of the HTML Document
3. Use Content That Search Engine Spiders Can Read
Content is the life force of a website, and it is what the search engines feed on. When designing a website, makes sure you take into account good structure for content (headings, paragraphs, and links).
Sites with very little content tend to struggle in the search results and, in most cases, this can be avoided if there is proper planning in the design stages. For example, don’t use images for text unless you use a CSS background image text replacement technique.
4. Design Your URLs for Search Friendliness
Search friendly URLs are not URLs that are hard to crawl, such as query strings. The best URLs contain keywords that help describe the content of the page. Be careful of some CMS’s that use automatically generated numbers and special code for page URLs. Good content management systems will give you the ability to customize and “prettify” your website’s URLs.
5. Block Pages You Don’t Want Search Engines Index
There could be pages on your site that you don’t want search engines to index. These pages could be pages that add no value to your content, such as server-side scripts. These web pages could even be pages you are using to test your designs as you are building the new website (which is not advised, yet many of us still do).
Don’t expose these web pages to web robots. You could run into duplicate content issues with search engines as well as dilute your real content’s density, and these things could have a negative effect on your website’s search positions.
The best way to prevent certain web pages from being indexed by search engine spiders is to use a robots.txt file, one of the five web files that will improve your website.
If you have a section of your website that is being used as a testing environment, make it password-protected or, better yet, use a local web development environment such as XAMPP or WampServer.
6. Don’t Neglect Image Alt Attributes
Make sure that all of your image
alt attributes are descriptive. All images need alt attributes to be 100% W3C-compliant, but many comply to this requirement by adding just any text. No
alt attribute is better than inaccurate
Search engines will read
alt attributes and may take them into consideration when determining the relevancy of the page to the keywords a searcher queries. It is probably also used in ranking image-based search engines like Google Images.
Outside of the SEO angle, image
alt attributes help users who cannot see images.
7. Update Pages with Fresh Content
If your website has a blog, you may want to consider making room for some excerpts of the latest posts to be placed on all of your web pages. Search engines love to see content of web pages changing from time to time as it indicates that the site is still alive and well.
With changing content, comes greater crawling frequency by search engines as well.
You won’t want to show full posts because this could cause duplicate content issues.
8. Use Unique Meta Data
Page titles, descriptions, and keywords should all be different. Many times, web designers will create a template for a website and forget to change out the meta data, and what ends up happening is that several pages will use the original placeholder information.
Every page should have its own set of meta data; it is just one of the things that helps search engines get a better grasp of how the structure of the website is constructed.
9. Use Heading Tags Properly
Make good use of heading tags in your web page content; they provide search engines with information on the structure of the HTML document, and they often place higher value on these tags relative to other text on the web page (except perhaps hyperlinks).
<h1> tag for the main topic of the page. Make good use of
<h6> tags to indicate content hierarchy and to delineate blocks of similar content.
I don’t recommend using multiple
<h1> tags on a single page so that your key topic is not diluted.
10. Follow W3C Standards
Search engines love well-formed, clean code (who doesn’t?). Clean code makes the site easier to index, and can be an indicative factor of how well a website is constructed.
Following W3C standards also almost forces you to write semantic markup, which can only be a good thing for SEO.