Professional, No Monthly Fees, No Third Party Ads, Advanced Personalization, Confirmed Opt-in, Broadcast, HTML Form Generator, & much more.
Newsletter
Subscribe to our newsletter and get all of the latest tips and tricks sent directly to your email! Get a Free Gift!
Name
E-Mail
RSS Feed
Get the most recent posts and comments sent to you directly by subscribing to our RSS feeds!
Subscribe to RSS! Subscribe to RSS Comments!
Nov
6

A Tribute to Steve Jobs

MiloPerson

R.I.P. Steve Jobs

(Feb, 24th 1955 – Oct, 5th 2011)

Mar
4

Beyond Positive Thinking

MiloPersonal Mastery

 

Larry King Live – Beyond Positive Thinking – Part 1

 

Larry King Live – Beyond Positive Thinking – Part 2

Mar
3

The Crisis of Credit

MiloEconomics

The layman explanation on the Sub Prime Crisis …


The Crisis of Credit Visualized from Jonathan Jarvis on Vimeo.

Jun
19

Search Engine Optimization (SEO) – Spiders

MiloSEO

About Spiders and Internal Links

Spiders representing the Internet’s search engines regularly crawl the World Wide Web to look for Web sites and other content to add to the search engines’ proprietary databases. They do so by following the hyperlinks that connect Web sites on the Internet. Thus, when crawling the Internet, a spider theoretically might visit any of the billions of pages on the Web. — Excluding, of course, sites that have no links pointing to them; unless those sites are submitted directly to the engines.

When visiting specific Web sites — whether submitted directly or located via a regular crawl — search engines similarly rely on their spiders to crawl the sites’ contents. The spiders do so by following the site’s internal hyperlinks. Although some limit the crawl to two or three directory layers, spiders generally will attempt to visit and evaluate all pages and elements of a visited site. This means that even though a user only submits the URL of his/her site’s home page, the search engine’s spider may visit and index several pages. Many engines may even index FTP, image or multimedia files that they come across while spidering a Web site. — Provided that the spider has internal links to follow.

Due to their very nature, spiders are crippled when they have no links to follow. Thus, a Web site with broken or no internal links on it will prevent a visiting spider from venturing beyond its point of entry. If the site contains functional external links, the spider will follow those and leave the site behind. If the site does not contain any external links, the spider de facto has reached a dead end. Unless the site in question is a one-page Web site, broken or lacking hyperlinks inevitably will force the spider to bypass pages and other site elements that otherwise might have been indexed individually and contributed to the Web site’s overall traffic influx. okay

If, however, all the Web pages on a site are linked to each other with functional hyperlinks, a visiting spider is able to peruse every nook and cranny of the site.

 

About Link Analysis

Search engines Google, Yahoo!, MSN, and Ask all consider the amount and — more important — the quality of “back links” (i.e., inbound page links) perhaps the single most important factor in their ranking calculations for indexed Web content. In essence, the search engines consider the presence of an inbound link to a given Web page an indication that the page from which the link originates considers the page it is pointing to important enough to actually link to. Additionally, the anchor text of the back link will help the search engine understand why the page is important. Based on various parameters, including the importance of the originating Web site, the search engines will evaluate each link pointing to a given Web site. This means that the nature of each link is actually more important than the overall number of links pointing to a site. This is particularly true in the case of Google whose PageRank system largely depends on link analysis. Indeed, the very root of Google’s immense success lies in link analysis: When Google entered the search engine market, most search engines relied on keyword usage and Meta data when analyzing and ranking Web content. Google’s focus on back links as a means to determine the relevance of a given Web page thus broke new ground and has led directly to Google’s overwhelming market dominance.

Sheer link quantity will have little impact on a page’s ranking. A limited number of quality links likely will have a bigger impact on a site’s search engine ranking than a large number of low-importance links. A high number of more or less relevant links to a site might increase the chances of random traffic stopping by the site, but will not make a dent in an otherwise lackluster search engine ranking. In other words: The key to building links to a Web site is to obtain links from the right places, rather than from as many links as possible. Note that link analysis is only one of several pieces of the search engine ranking puzzle. Therefore, to achieve the best results quality links should go hand-in-hand with well-composed Meta and Title tag content and site copy.

 

Not All Links Are Created Equal

In their earlier incarnations, the Internet’s search engines were more prone to include link quantity in their ranking calculations. But with that concept all but gone, link quality is one of the keys to boosting search engine rankings.

When analyzing links to a Web site, search engines will attempt to determine the relative importance of the originating sites. This means that if an originating site offers similar or related content to the one it is pointing to, the link will receive a higher rating than if the originating site’s content was entirely unrelated. As well, an originating site may be ranked based on the links that are pointing to it, and hence the relative importance of those links and their origins. The highest-quality links are those that appear to occur organically: Organic linking occurs when a Webmaster comes across a Web page that he/she finds particularly interesting or important or for some other reason wants to direct his/her site visitors’ attention to and thus decides to link to that page from his/her site. Organic linking, thus, is not the product of any form of link exchange. Obviously, no search engine can determine with absolute certainty the genesis of a given back link. However, one-way linking may serve as an indication of organic linking, as link exchanges generally involve reciprocal linking. Reciprocal linking is not necessarily bad — indeed, it can be a necessary element in any Web site owner’s link building efforts —, but advanced search engines do not rate them as highly as organic ones.

A search engine’s link quality analysis further examines the relative importance of the page that contains the outbound link. In short: If, for example, a page that holds a Google PageRank of 8 points to a given Web page, then — theoretically — that link will have a more positive influence on the linked-to page’s search engine ranking than it would had the link originated from a page with a PageRank of 3. This, of course, assumes that the search engines trust that the back link is bona fide and not part of a link buying/link farming scheme.

It is important to keep in mind that the exact link analysis method varies between the Internet’s search engines, and that the same link therefore will not necessarily have the same significance with all search engines. Nevertheless, meticulously built links to and from carefully selected sites are bound to be more effective than random reciprocal links obtained via free-for-all link exchange services and farms. Still, reciprocal linking can be useful, and is often offered as common courtesy when requesting links from another Web site.

Search engines in some cases consider internal links — i.e., links between a Web site’s pages — bona fide links and thus incorporate them into their ranking algorithms. In most cases, however, internal links will have no effect on the ranking system.

 

Link Context

Some search engines, including Google and Ask’s Teoma, when performing link analysis zoom in on the link text and the context in which the link is found. This is generally done by examining words in or in the vicinity of the link. If the context fits the link, then the link likely will be considered more important than if it seemed out of context. In a nutshell: If a link or a lead-in to a link mentions music CDs and the link points to an online music store, that link will receive a relatively high ranking because the context supports the link destination. Conversely, if the link text/adjacent text makes no mentioning of music or any related terms, search engines will not credit the link for its context. In short: If a back link indicates that the originating Web site considers the targeted site important, then the link text and link context may indicate why it considers it important.

 

Finding and Acquiring Good Links

How, then, can the Web developer determine which link originators will create the desired results?

Using the Internet’s search engines is the easiest and most obvious method to find out which Web sites will provide relevant and hence high-scoring links. By visiting a search engine and typing in a search phrase that ideally would lead to the type of site the developer is creating the search engine will return a list of matching sites. Generally, the best-ranked matches are the ones most likely to generate the most effective links. Having perused the results, the next step is to visit the sites and convince their administrators to provide the desired links.

There is, of course, no guarantee that a particular Web site will add a requested link. There can be many reasons that a Web site might decline to do so. For example, the site might refuse to link to an obviously competing online location. Or, the site’s owner may not be interested in being affiliated with other sites. Another reason could be that the site is already providing several outbound links and is not interested in adding to the count. Thus, sometimes the Web developer will have to look for links someplace else.

In any case, to request a link to a site, the Web site owner should compose and send a formal request to the targeted Web site’s owner or administrator and explain why providing a link would be mutually beneficial. If applicable, reciprocal linking can be offered as an incentive. But if the request is turned down it is better to move on to the next Web site on the list and request a link from there instead.

 

Sandboxing

Lately, the top search engines, notably Google and Yahoo!, have employed so-called “sandboxing” measures to ensure that no new Web page’s ranking is allowed to skyrocket until its links and content have been examined carefully. The sandboxing period, generally a few months, applies to most newly-indexed pages and allows the search engines to ascertain that a page’s inbound links do not originate from a link farm or other form of search engine spamming. The sandboxing concept can be frustrating for Web site owners, but it further increases the importance of link quality in relation to search engine ranking.

 

Frame-Related Problems and Solutions

Because spider-based search engines cannot crawl the content of Web sites that use frames — and some older browsers cannot display them — Web developers are generally advised to abstain from using frames when building their Web sites.

Frames are usually used to separate site content and facilitate navigation. However, because search engines cannot read and crawl the body contents and links of sites that are built with frames, framed sites usually don’t score high in search engine rankings. If frame usage is required, the page should include a Noframes tag that will provide alternate content for search spiders and browsers that do not support frames.

 

The Problem With Frame Usage

A framed Web page consists of a home page — or frameset —, which contains HTML code that defines and links to the frames themselves. The frameset page is the master Head tag. This tag defines all other pages used in the framed site.

When visiting a framed Web site, search engines can view the master frameset page, but cannot put the framed pages together. This means that search engines only see the HTML code on the frameset page, not the primary page body contents, which are located in the frame elements. Thus, when submitting a framed Web page to an Internet search engine, generally, the frameset page is the one submitted, but because the actual content resides in the frame elements, the search engine won’t see any potentially rank-boosting elements located in the frame elements. Therefore, only very few, if any, framed Web sites possess top-ten search engine rankings.

See below for an example of frameset content. In this case, the frameset points to three frames: Header, FrameOne, FrameTwo:

<html>
<head>
<title>Site Title</title>
<meta http-equiv=”content-style-type” content=”text/css”>
<meta http-equiv=”content-type” content=”text/html”>
<meta name=”description” content=”Site description;”>
<meta name=”keywords” content=”keywords”>
<frameset> border=”0″ rows=”100,*”>
<frame name=”header” noresize src=”Header.htm”>
<frameset cols=”100,8*”>
<frame name=”navigation” noresize src=”FrameOne.htm”>
<frame name=”content” noresize src=”FrameTwo.htm”>
</frameset>
</frameset>
</html>

Obviously, the above content will not give an algorithmic search engine much to work with. — And because all of the site’s body content, including in- and external links, is absent, the search engine’s ability to crawl, evaluate and rank the site is severely limited.

Naturally, getting rid of the frames entirely would solve the problem of lackluster search engine rankings for framed Web sites. However, ridding a Web page of frames does limit the site-design options. The Web site developer therefore should carefully weigh the pros and cons of using frames on the site before deciding whether to stick with or abandon frame usage.

 

The Noframes Solution

An alternative to omitting frames is to add a Noframes tag to the master frameset page. While spiders cannot crawl framed Web sites, they can read the contents of a Noframes tag. The Noframes tag enables the Web site developer to create an alternate no-frame version of a framed Web site. When a search engine — or browser that cannot read frames — comes across a framed site, it will instead read or display the contents of the Noframes tag.

The Noframes tag may contain a simple message and a link to a no-frame alternate site. Or, the tag can contain the entire HTML code, including links and body content, for the alternate site. The latter option secures that visiting search spiders can crawl the site in its entirety, including all in- and external site links.

 

Placing the Noframes Tag

The Noframes tag should be placed inside the Frameset tag. See example below.

<html>
<head>
<title>Site Title </title>
<meta http-equiv=”content-style-type” content=”text/css”>
<meta http-equiv=”content-type” content=”text/html”>
<meta name=”description” content=”Site description”>
<meta name=”keywords” content=”keywords”>
<frameset> border=”0″ rows=”100,*”>
<frame name=”header” noresize src=”Header.htm”>
<frameset cols=”100,8*”>
<frame name=”navigation” noresize src=”FrameOne.htm”>
<frame name=”content” noresize src=”FrameTwo.htm”>
<noframes>
<body>
<h1>Page headline </h1>
<p> Page body </p>
</body>
</noframes>
</frameset>
</frameset>
</html>

 

Using Flash Animation

Well-composed Flash animation can turn a Web site into a virtual work of art. Unfortunately, most spider-based search engines do not read Flash. Although Google’s spider currently has the capability of reading and indexing Flash animation, it is generally recommended that site elements that are critical to the page’s message, subject, and navigation are created in a format that can be read by all spiders. Flash animation on a Web site will not help improve the site’s search engine ranking. — No matter how aesthetically pleasing the animation makes the site. Although Flash animation will not improve a Web site’s ranking, it could, if used unwisely, damage it. Note that Macromedia’s Flash Search Engine SDK provides a set of object and source code designed to convert a Flash file’s text and links into HTML, thus enabling search engines to read the content.

On home pages, Flash animation is sometimes used to create visually stunning lead-ins. Occasionally, home pages are built entirely with Flash. In such cases some or all of the home page content is unreadable and -crawlable for Internet search spiders. Therefore, pages that feature substantial amounts of Flash animation might only score mediocre rankings with Internet search engines, while entirely Flash-based pages probably won’t be indexed at all. In other words, submitting entirely Flash-animated pages to Internet search engines is usually futile. If submitting pages that are partly Flash-animated, decent rankings might be achieved by optimizing the page’s Meta and Title tags, and by adding well-composed and keyword-rich page copy.

 

Flash- and HTML-Based Hyperlinks

Because search engine spiders cannot read Flash, crawler-based search engines will detect no links created with Flash. For search engine positioning purposes, regular HTML hyperlinks therefore are considerably more effective than Flash-based links. However, by incorporating both solutions — an HTML and a Flash version of each page link — on the same page, Flash animation can be retained without crippling visiting spiders.

 

Redirections and Google

n the context of Google optimization your Web site’s internal links mainly serve to enable the Google to spider to access all corners of the site. All internal links, therefore, must be functional.

Generally, Google advises against using HTTP redirections on submitted URLs. Instead, you should submit the destination URL itself. If Google visits and indexes a Web site without the site’s URL being submitted via Google’s “Add URL” link (or through SEO services like Traffic Blazer), then the actual site will be indexed. Not the redirected URL.

In its Quality Guidelines Google very clearly advises Web masters against using “sneaky redirects.” That phrasing refers to redirections that serve the single purpose of tricking spiders into believing that the destination is more important than it really is.

Google does, however, recommend using a 301 permanent redirect HTTP status code if you have moved an existing Google-indexed site to a new URL:

Once your new site is live, you may wish to place a permanent redirect (using a “301” code in HTTP headers) on your old site to inform visitors and search engines that your site has moved.

In short: If you have moved your content, use a 301 redirection. Otherwise, do not attempt to redirect the Googlebot.

 

Understanding the Robots.txt Standard

 

In some cases, you may want to take control of how Internet search engines crawl and index your Web site. There can be a number of reasons to instruct visiting spiders to steer clear of entire Web sites, or certain site elements. For example, you might wish to prevent pages that are under construction from being indexed by search engines. Or you might simply want to limit your bandwidth consumption by preventing certain spiders from crawling your site.

Regardless of the reason for limiting spider access to your site, the standard method of doing so is to create a “robots.txt” file and place it in your Web site’s root directory. The Robots.txt file — whose directions practically all robots/spiders comply with — can prevent certain or all search engines from visiting and indexing particular pages within a Web site. A robots.txt file can also be used to entirely block spiders from crawling a site. The Traffic Blazer Robots.txt Generator provides and easy means of creating a robots.txt file for your Web site.

 

Composing the robots.txt file

The robots.txt file can either block spiders entirely, or it can prevent the search engines from accessing specific directories, files or entire Web pages.

The robots.txt file consists of two defining elements:

  • user-agent:
  • disallow: /

The first element— “user-agent:” — specifies which agents, spiders or browsers should read and obey the commands in the file. The second element — “disallow:” — defines which files and directories should be blocked from the search engines. An asterisk (“*”) denotes “everything.” By following this concept, you can thus define which spider(s) should steer clear of the site and which ones should crawl all or only selected parts of its contents.

 

The Robots Meta tag

A less versatile alternative to the robots.txt file is to create a Robots Meta tag for each applicable page on your Web site. A Robots Meta tag can be used to instruct search engine spiders not to index a page and/or not to follow the hyperlinks on the page. This tag must be included in the <head> tag portion of every Web page you want excluded. Note that the Robots Meta tag applies to all search engine spiders/robots; the Robots Meta standard does not allow you to specify particular spider names.

Apr
15

Free CMS – Joomla & Drupal

MiloCMS, Drupal, Joomla

If you are wondering if there are good alternatives to the WordPress, I came across 2 recently. They are Joomla and Drupal. Similar to WordPress, they are free open source Content Management System (CMS), web-based program which allows users to manage and organise content, documents and multimedia files on websites and to collaborate and interact. To give an idea of the power of CMS, think facebook, MySpace & Youtube.

Joomla

I took a spin on the demo provided by Joomla on its website and became a convert. The interface is simple and user friendly. Users do not need to know anything about coding to get started. I put in some advertisements on the side bar and contents section without much of a hassle. There are also plugins, templates (themes), language selection and polls available.

Drupal

Visiting the Drupal and going to some of the sites using Drupal as a platform gives me an impression that it is more poerful than Joomla. One website I visited was a discussion forum, yep … created out of Drupal, here it is  http://www.fastcompany.com/ . Afterall, Drupal did won  the best overall 2007 Open Source CMS Award given by Packt Publishing. Joomla came in second. WordPress was first in the category for Best Open Source Social Networking Content Management System, in which Drupal came in second. Drupal is also better in community features and search engine friendliness. One drawback is that you will need to have a bit of web programming know-how to get around it.

Setting up websites in Drupal or Joomla platforms are definitely worth considering.

Apr
7

Free Graphic Tools

MiloGraphic Tools

Graphic editing tools are indispensable for any online business. There are so many free graphics and image enhancing tools out there that you don’t really have to burn cash to get one. Here’s a run down of some of these free graphics tools:

1.  Paint.NET

This is my current favourite. Paint.NET is free image and photo editing software for computers that run Windows. It features an intuitive and innovative user interface with support for layers, unlimited undo, special effects, and a wide variety of useful and powerful tools. An active and growing online community provides friendly help, tutorials, and plugins. Specify the pictures size (ht & width) in pixels for your banners, logos or advertisement graphics to fit them into your web page.

2.  Adobe Photoshop Web Edition

As you may know, Photoshop has a huge followings and most professionals use them. This one is web based and its free but offers only some very basic features. Although user friendly, it has its limitations:

– not able to accept .png format pics.
– does not allow creation of new file. Images to upload, edit, then download.
– does not have an aspect change feature although cropping and reduction of resolution is available. 
– images store on the website is not private, but you can host your pictures up to 2 GB, with easy link creation. Pictures may be hosted here if you’re selling on eBay to save cost.

3.  Picnik

Web based, fast, easy & fun. Works much like the Photoshop Web Edition. Upload, edit, then save the pics to a folder on your PC. Editing is made easy by the dragging of sliders. Allows for text with flexibility of moving it around the picture.

4.  Phixr

Web based. The operation windows for Phixr are too small for my comfort, can’t see the changes clearly. The entered text are fixed to a few pre-determined poistions, eg. top-left, centre, etc. and not able to move it around in the pictures. Can accept 5 formats, jpg, png, pdf, gif & OCR.

5. Splashup

At first look, you could mistake this for the desk top version of Photoshop. The tools icons available at the sidebar is almost similar to Photoshop CS3. Formerly known as Fauxto, it is a powerful editing tool and photo manager. With all the features professionals use and novices want, it’s easy to use, works in real-time and allows you to edit many images at once. Splashup runs in all browsers, integrates seamlessly with top photosharing sites. Guess what it allows you to not just edit an existing photo but also allows you to create new graphics, of specific pixel size, to fit into your web pages. To me, this is the best of the best as far as web based free photo editor is concern. You got to give it a spin to really know its power.

There you go, free and powerful resources for your online activities

 

Feb
14

Acquiring Products & Building Subscribers List

MiloAutoresponder, List Building, Product, Traffic

Since the setting up of my autoresponder, I had been pretty busy looking at ways to acquire products and building a list. These 2 activities usually go together cause without a product to offer as an “ethical bribe” it would not be possible to build the list. With a reasonable size list, a marketing campaign can then be set up to sell the products.

I have so far garnered about 1000 odd products and more than 100 subscribers to my list. That took up most of my internet time over the last one month.

A few other things I did recently:

1.  Put up a subscriber opt-in form at the top left corner of my site.
2.  Set up the ultimate Multi-Media Autoresponder sales pages. You can access it by clicking on the image from the right side bar.
3.  Added some more useful codes to my “HTML Codes” page, codes which I use pretty often in setting up web pages.
4.  Went into a few Joint Ventures (JV) with some internet marketers to build a list.
5.  Of course celebrating the Chinese New Year, meeting up & building bonds with old friends and relatives.

Guess what, the MM Autoresponder is starting to rake in an income even before I launch any serious marketing campaign. Couldn’t imagine the result once I start promoting it. The MM Autoresponder is a professional and ultimate cost saving autoresponder and I know it will sell. As the saying goes … The Money is In The List, and to build a list, an autoresponder is indispensable.

My next major tasks are:

1.  To identify more opportunities to increase my subscriber base.
2.  To capitalize on the natural traffic of eBay and hopefully become a Power Seller. In fact I have already starting to create the sales pages.
3.  To set up a marketing campaign for the products I had acquired.

That will be quite a handful over the next one to two month.

Dec
17

Product Rights De-mystified

MiloProduct

If you have subscribed to my newsletter, you will probably be exposed to hundreds of free products that come with certain rights at your disposal and you can turn them into profits if desired. Here is a run down on the different types of rights to products:

Normal ReSell Rights (RR)
When you acquire a product with normal Resell Rights, you will be allowed to distribute it or sell it to your customers at 100% profit. However, there are restrictions such as you can only sell the product for a specific price and you will also not be allowed to change the contents. This allows the author to propagate his purpose and is thus a win-win situation.

Master ReSell Rights (MRR)
When you acquire master resell rights, not only can you sell it to your customers at 100% profit, your customers also have the rights to sell it. You can also sell master resell rights to your customers for an additional price, so they too can sell resell rights and master resell rights.

Private Label Rights (PLR)
With private label resell rights, you may not only resell the product but you may also edit the content or product and claim it as your own work (meaning, you become its author). Basically you can call it your own and do whatever you want with the products, wholely or partially.

Rebranding
With rebranding resell rights, you may replace the author links with yours. That is in addition to your normal resell rights.

No Rights
You can only have it for your personal use and not distribute it in any way.

There you have it … Know Your Rights, Know Your Potential!

Dec
6

My Autoresponder is a GO!

MiloAutoresponder, List Building

I am elated!

I had been fully focused getting my autoresponder up and running over the past one month. It is definitely well worth the effort albeit the many errors encountered, sleepless nights and having had to change host in the process.

Here I would like to share with you how you can also save loads of money if you have the intention to set up your autoresponder. To me the autoresponder is the next most important thing to setup after you have got your site up and running.

Why is an autoresponder so important?

In short, an autoresponder helps builds a targeted mailing list for your niche. The list then allows you to do some marketing, be it selling your own products, selling affiliate products or simply bringing traffic to your site and be exposed to your other avenues for revenue. Most importantly it allows you to make money on autopilot. These are big topics and I will not elaborate too much over here at this moment.

In case you think it is spamming, it is NOT. Most people would have experienced receiving unsolicited emails on enlargement of you-know-what and sales of medicine. I hate spamming just like anyone else.

In above-board internet marketing, the subscriber likes what he sees and subscribes to it willingly and voluntarily because there is value for him. He also has the option to opt out of the list anytime by a single click.

Some gurus equate a good list to an ATM. With a good list the possibilities are unlimited. Every time when you need money, instead of going to the ATM, just send out a marketing campaign to your list and watch the money rolls in.

Put it simply, no autoresponder, no list. No list no marketing. Enough said.

What are your autoresponder options?

The 2 most popular autoresponder used by internet marketers I know of are:
1.  GetResponse (USD17.95/mth)
2.  Aweber (USD19.95/mth)

Problem is they both come with a monthly recurring cost. Quite hefty for a new startup if you ask me. I don’t like the idea of recurring cost. The only recurring costs I am willing to pay are Domain name and Host, nothing else.

Many a times I was tempted to subscribe to GetResponse because many experience internet marketers had recommended it. I held off for a while and got paid off.

I stumbled upon an Autoresponder with a one time payment going for a song. The price is less than one month the basic subscription of GetResponse.  Not only that, it has got all the features you will ever need in an autoresponder … and I mean all, no joke. Crazy, but its true. That’s what I call Stylo.

The Pro Internet Marketer, Paul Johnston, who sold it to me was with me every step of the way setting it up. Thank you Paul.

If you would like to know more about this autoresponder, let me know by dropping a comment.

My next job is to create a list capture page, sometimes known as a squeeze page. You will also see a subscription box appearing on my site soon.

Go on and subscribe to it if you would like to be updated on what you see on my site and lots more from other sites.

Yes! I am back in business!

Nov
9

Step 8: Complete the 2 step registration, Customize & Insert Contents

MiloStartup, WordPress

Complete the 2-Step Registration 

Congratulations for successfully installed WordPress. Complete the 2 step registration process when prompted and you are good to go. The inputs may be modified if desired at the Options menu.

Customize

Next is to customise your site with desired  themes and plugins.

Again some links that you can refer to for help if you run into any trouble in the installation process:

WordPress Main Page
WordPress Support Forums

Create Contents 

Last, but not least, insert contents. Try to get up to about 12 quality posts ASAP. 12 is just a general guide. Without sufficient contents, Adsense or Blogrush may reject your application.

That completes our series of setting up a WordPress website.

Welcome to the WordPress blogging community! Happy blogging!

 PS: To read the entire story in starting up a website, select “Startup” under categories.

Google