Thursday, July 31, 2008

SEO Basics For A Rookiee

The Search Engine Optimization process is briefly described by the following stages:

1. Keyword Research:

Proper research assures that your website is optimized around the keywords most often used by your visitors. Without proper research, Optimization is likely to yield poor results. The research phase produces a finalized set of keywords that may be anywhere from 1-150 keywords. You can use the free tool available at http://freekeywords.wordtracker.com. Prepare a set of keywords for every page of the website depending on the content that is on the webpage.

2. Identifying your Competition:

After selecting the best keywords, it is important to know which websites rank highest in the major Search Engines. Websites that appear the most often are referred to as your ‘Competition.' Comparing your website to your Competition helps to give an idea of where the initial efforts of Optimization should be focused. Identify and study the keywords being used. Check the link popularity and PR of the competitors and make a note of it. Identify the traffic the competitors’ are getting. You can use these free tools available at – http://www.marketleap.com, http://www.trafficestimate.com, http://www.alexa.com.

A detailed analysis of your website and the competition provides the details necessary to plan a successful Optimization.

3. Meta Tags: Meta-tags & Title tag, Hyperlinks, Image Alt tags:

Optimizing the HTML code includes creating Title tags, Meta description tags, Meta keyword tags, hyperlinks, and possibly headings & image alt attributes. All of these are factors in helping the search engine spiders to properly classify your site, with some being more important than others. Rather than go into detail on how to perform all of these tasks, I'll direct you to some of my previous SEO articles.

When your code is optimized, it's finally time to upload your new pages to your server. In the past, at this point we would then resubmit them to the search engines. However, these days, this is an unnecessary step. As long as you're dealing with an existing site with some links already pointing to it, the search spiders should visit and index at least your home page fairly quickly. It may take a bit longer for them to re-index the inner pages, but you can rest assured that they will.

Do not spam the description or keyword meta tag by stuffing meaningless keywords or even spend too much time on this tag. SEO pros all agree that these tags are not as important today as they once were.

4. Content Optimization:

Once the keyword phrases are chosen for each page of the site, it's time to get down to the nitty-gritty and start utilizing them within the copy. If you're rewriting from scratch, be sure your copywriter understands which phrases need to be used, what the goals of the site are, and who the target market is. Obviously this information will affect how they write the copy, so the more they know; the more accurate your copy will turn out. If you're lucky enough to be able to work in the necessary keywords, that's your next step. Once your copy is finished and approved, you should now have a number of pages focusing on 2 or 3 keyword phrases each.

5. HTML Site Map:

Create the html sitemaps providing the links of all the pages. Have the keywords in the body text and the anchor links if possible.

6. XML Site Map:

Create xml sitemaps and submit to Google and Yahoo. Free tools are available on - http://www.xml-sitemaps.com/, www.google.com/webmasters/ and http://siteexplorer.search.yahoo.com/.

7. Website Structure Analysis:

Website Structure Analysis actually means website architecture. It is an approach to the design and planning of websites which involves certain criteria like technical, aesthetic and functional. The main focus is on the user and user requirements in designing a structure of the website. Attention is primarily given to the web content, usability, interaction design, information and web design.

In case of search engine optimization the navigation schemes of the website should be created most importantly for use by visitors. Through usability testing we can easily determine what type of navigation schemes visitors prefer. Like for example a text based or graphics based one.

For a website to be successful first and foremost, it should satisfy your target audience. At the same time by making sites simple and easily accessible we will be providing the easiest path for the search engine robots to index our site. This strategy helps in the long run success of search engine optimization. No matter how often the search engines algorithm change the good content and simple navigation is always rewarded.

Search engine friendly design: The easier the navigation and the more text on the page the better it is not only for the visitor but also for the search engine robots. Certain points to be considered while designing a search engine friendly site are not to include long javascript in the source code, dynamically generated pages, frames and so on as the search engine robots have difficulty in crawling them.

However there are techniques to overcome the problems with indexing but the best way to assure that the pages will be indexed is to keep them simple and ensure return visits both from the search engines and the visitors as well.

8. Robot.txt file:

Many search engines use programs called robots to gather web pages for indexing. These programs are not limited to a pre-defined list of web pages; they can follow links on pages they find, which makes them a form of intelligent agent. The process of following links is called spidering, wandering, or gathering.

The robots.txt file,

• is a text file created in notepad or any text editor
• should be placed in the top level directory or root of the website or server space
• should include all lower case letters

Through the robots.txt file we can,

• exclude all robots from visiting the server
• allow complete access to all robots
• exclude robots from accessing a portion of the server
• exclude a specific robot
• exclude certain type of files from accessing by specifying the file extensions.

Here are few examples of robot.txt file. Check both good and bad practices.

Entry Meaning
User-agent: *
Disallow:
The asterisk (*) in the User-agent field is shorthand for "all robots". Because nothing is disallowed, everything is allowed.
User-agent: *
Disallow: /cgi-bin/
Disallow: /tmp/
Disallow: /private/

In this example, all robots can visit every directory except the three mentioned.

More examples are available in my previous post - http://seocolumn.blogspot.com/2005/12/search-engine-robots-and-robottxt.html

9. One-way links and Directory links:

Some SEOs recommend that you do these before all the other work so that you can start getting rankings right away. However, I prefer to wait until the copy changes have been made and uploaded. The newly focused copy helps the directory reviewers to more easily understand what your site is all about, and they'll be less apt to edit your submitted description. If you submit before your site's copy is using your specific keyword phrases within the copy, the reviewer may feel these keywords don't belong within your directory description.

When link building, think quality, not quantity. One single, good, authoritative link can do a lot more for you than a dozen poor quality links, which can actually hurt you.

Where as directory submission to Yahoo, DMOZ… can happen immediately. You just need to make sure you have the details asked by these website forms are ready with you. And make sure to have the home page to have some relevant content prior to submission to these directories.

10. Reciprocal links:

As with directory submissions, I prefer to wait until the site is in perfect condition before starting to request links. You can certainly get started researching link partners before the SEO is complete; however, the better your site is, the more likely others will be willing to link to it. Link building can be done in a quick burst, but should also be an ongoing process. You should always be on the lookout for sites that are a good fit with yours and contact them about exchanging links.

In some cases, additional content pages may also need to be created as per the need of the keywords you want to rank for.

11. Article Submissions:

The fundamental base of the internet is not the page, but the link. A great way to accumulate links is to write articles and submit them to places interested in quality articles.

When you write articles for free on certain websites, you are typically given a by-line or author bio which you can link to your website. Few websites also accept links within the body text using html tags.

Few good Article resources are listed below.

http://goarticles.com/
http://ezinearticles.com/add_url.html
http://amazines.com
http://www.articlecity.com/
http://bpubs.com/
http://businessknowhow.com/
http://www.certificate.net/wwio/ideas.shtml
http://www.promotionworld.com/
http://promotenewz.com/

12. Adopt Web 2.0 Technologies:

Do not think that the traditional tactics of SEO – optimizing using meta tags, keyword density, prominence of keywords…more has lost the importance. They are still important and one needs to follow these for better rankings.

Here are few easy steps…http://seocolumn.blogspot.com/search/label/web 2.0

No comments: