Seo stands for search engine optimization. This page is dedicated to helping improve your page rank on search engines.
When selecting a domain name, try to choose a short domain name, omitting words like "the" the longer the url, the harder it is to remember, and this affects page rank.
Domain registration duration impacts SEO, if you intend to have the site a long time, register several years rather than a short period of time, as this will improve SEO.
your sites urls should have 1 or the other url re-writing, and 301 permanent redirecting so search engines do not see duplicate content. You are not penalized for url length by adding a www. subdomain.
# curl -I http://funtoo.org/Welcome
# curl -I http://www.funtoo.org/Welcome
Question marks are bad. Most search engines freak out about indexing anything past a ? so use url rewrites to remove index.php? and leave just site.com/Main_Page
You should prefer - over _ to represent spaces. MediaWiki is flawed in design preferring _ over -. This is an older SEO problem, yet it still persists in affecting your page ranking. web-server-stack is an example of an SEO friendly url with spaces.
How fast your page loads significantly impacts seo. Installing a caching reverse proxy, and testing that it hits close to 100% of the time is a good idea. It prevents your server from rebuilding pages, hitting the processor, and also speeds up page delivery to move the bottle neck from the php building to the internet connection speed. see: Varnish
Meta information is parsed directly by search engines. with these in place they will be displayed rather than the initial text of your canonical landing page.
Dublin core is a system that generates meta information dynamically.
Posting links around is ok so long as it is organic, and not spammy. addthis share widgets improve SEO as they produce many backlinks that are not a bot dumping anywhere, and everywhere. addthis produces twitter/facebook/g+/redit etc share links, and has an analytic service.
SiteMaps & robots.txt
Sitemaps generate xml pages that tell crawlers about your sites content pages, and robots.txt tell crawlers about pages they're not allowed to index. In a few iterations ill look up how to make a robots that allows everything, as if its not even there, and an external site map generating service.
testing, and evaluating
- http://www.seomastering.com/ (shows estimated page value in USD)
- http://seositecheckup.com/ (throttled to 1 check every 30 minutes)
free with required registration
pay with free trial
free 1 site per week testing