If you decide to go into affiliate marketing, understand that you will need a lot of very targeted traffic if you want to make any real money. Those affiliate offers also need to provide a high commission amount to you on each sale. You also need to ensure that the returns or chargebacks for those products or services are low. The last thing you want to do is to sell a product or service that provides very little value and gets returned often.
To avoid undesirable content in the search indexes, webmasters can instruct spiders not to crawl certain files or directories through the standard robots.txt file in the root directory of the domain. Additionally, a page can be explicitly excluded from a search engine's database by using a meta tag specific to robots (usually ). When a search engine visits a site, the robots.txt located in the root directory is the first file crawled. The robots.txt file is then parsed and will instruct the robot as to which pages are not to be crawled. As a search engine crawler may keep a cached copy of this file, it may on occasion crawl pages a webmaster does not wish crawled. Pages typically prevented from being crawled include login specific pages such as shopping carts and user-specific content such as search results from internal searches. In March 2007, Google warned webmasters that they should prevent indexing of internal search results because those pages are considered search spam.[46]
Moreover, the PageRank mechanism is entirely general, so it can applied to any graph or network in any field. Currently, the PR formula is used in bibliometrics, social and information network analysis, and for link prediction and recommendation. It's even used for system analysis of road networks, as well as biology, chemistry, neuroscience, and physics.
I’m growing tired of this game between Google and the rest of the online community about how to “manipulate” my content and code to better rank in your system. It seems that you guys have completely over complicated the game. If I add a nofollow tag, why on earth would any page rank be added to that link. I just told you to NOT FOLLOW it! The fact that it receives any rank at all is absurd.

Black hat SEO attempts to improve rankings in ways that are disapproved of by the search engines, or involve deception. One black hat technique uses text that is hidden, either as text colored similar to the background, in an invisible div, or positioned off screen. Another method gives a different page depending on whether the page is being requested by a human visitor or a search engine, a technique known as cloaking. Another category sometimes used is grey hat SEO. This is in between black hat and white hat approaches, where the methods employed avoid the site being penalized, but do not act in producing the best content for users. Grey hat SEO is entirely focused on improving search engine rankings.
While Google never sells better ranking in our search results, several other search engines combine pay-per-click or pay-for-inclusion results with their regular web search results. Some SEOs will promise to rank you highly in search engines, but place you in the advertising section rather than in the search results. A few SEOs will even change their bid prices in real time to create the illusion that they "control" other search engines and can place themselves in the slot of their choice. This scam doesn't work with Google because our advertising is clearly labeled and separated from our search results, but be sure to ask any SEO you're considering which fees go toward permanent inclusion and which apply toward temporary advertising.
Internet Marketing Inc. is one of the fastest growing full service Internet marketing agencies in the country with offices in San Diego, and Las Vegas. We specialize in providing results driven integrated online marketing solutions for medium-sized and enterprise brands across the globe. Companies come to us because our team of well-respected industry experts has the talent and creativity to provide your business with a more sophisticated data-driven approach to digital marketing strategy. IMI works with some clients through IMI Ventures, and their first product is VitaCup.
If you build a new site and only used Domain Authority to create links, you could EASILY have got linked from the worst page possible, even if it was from the best domain, because of the INTERNAL LINKS of the other web pages! How on earth are you going to be able to see the strength of a link if that strength depends on the internal links on an entirely different website?!
Nashville Grant, here’s the mental model I’d employ: search engines want to return great content. If you make such a fantastic site that all the web has heard of you, search engines should normally reflect that fact and return your site. A lot of bad SEO happens because people say “I’ll force my way to the top of Google first, and then everyone will find out about my site.” Putting rankings before the creation of a great site is in many ways putting the cart before the horse. Often the search rankings follow from the fact that you’re getting to be well-known on the web completely outside the sphere of search. Think about sites like Twitter and Facebook–they succeed by chasing a vision of what users would want. In chasing after that ideal of user happiness and satisfaction, they became the sort of high-quality sites that search engines want to return, because we also want to return what searches will find useful and love. By chasing a great user experience above search rankings, many sites turn out to be what search engines would want to return anyway.
A variety of methods can increase the prominence of a webpage within the search results. Cross linking between pages of the same website to provide more links to important pages may improve its visibility.[47] Writing content that includes frequently searched keyword phrase, so as to be relevant to a wide variety of search queries will tend to increase traffic.[47] Updating content so as to keep search engines crawling back frequently can give additional weight to a site. Adding relevant keywords to a web page's meta data, including the title tag and meta description, will tend to improve the relevancy of a site's search listings, thus increasing traffic. URL canonicalization of web pages accessible via multiple urls, using the canonical link element[48] or via 301 redirects can help make sure links to different versions of the url all count towards the page's link popularity score.

Great article and writing in general. My company just published a 5,000 word Keyword targeting best practices guide for PPC and SEO, and we linked to your article “10 Reasons You Should Use Google Trends for More Than Just Keyword Research”. http://vabulous.com/keyword-research-targeting-for-ppc-and-seo-guide/ I would love if you checked it out and possibly shared it if you like it.


Assume a small universe of four web pages: A, B, C and D. Links from a page to itself are ignored. Multiple outbound links from one page to another page are treated as a single link. PageRank is initialized to the same value for all pages. In the original form of PageRank, the sum of PageRank over all pages was the total number of pages on the web at that time, so each page in this example would have an initial value of 1. However, later versions of PageRank, and the remainder of this section, assume a probability distribution between 0 and 1. Hence the initial value for each page in this example is 0.25.

Muratos – I’ve never nofollowed Amazon affiliate links on the theory that search engines probably recognize them for what they are anyway. I have a blog, though, that gets organic traffic from those Amazon products simply because people are looking for “Copenhagen ring DVD” and I hard-code the product names, musicians’ names, etc. on the page rather than use Amazon’s sexier links in iframes, etc.
All of the examples above and more could be used as anchor text for the same backlink. Google will index each differently. Not only that, Google will even examine the few words before and after the anchor text as well as take into account all of the text on the page. It will also attribute value to which backlink was first in the page and diminish the value for each following link.
In my example, if I am passing PR to a local eatery by having a do-follow link but there are 9 nofollow links on that page and I only had 10 points to begin with then that lowers the value I can give from my local foodie blog to that site. In that case would it actually be better to either disallow comments on that page or to disallow links associated with the comments on that page? I mean if my client is a food blogger (and some are) and they tell the restaurateur “when I write about you it will be good for your Google juice because I will place a link to you with my post” then they would really be diminishing the value they could give by having an increased number of links. Kinds of sucks for the blogger who wants a lot of comments, no?
A more intelligent surfer that probabilistically hops from page to page depending on the content of the pages and query terms the surfer that it is looking for. This model is based on a query-dependent PageRank score of a page which as the name suggests is also a function of query. When given a multiple-term query, Q={q1,q2,...}, the surfer selects a q according to some probability distribution, P(q) and uses that term to guide its behavior for a large number of steps. It then selects another term according to the distribution to determine its behavior, and so on. The resulting distribution over visited web pages is QD-PageRank.[41]
Black hat SEO is to be avoided. This is basically link spamming. You can pay somebody peanuts to do this on your behalf and, for a very short period, it brings results. Then Google sees what’s happened, and they delist your site permanently from search engine rankings. Now, you need a new website and new content, etc.—so, black hat SEO is a terrible idea.

The Truth? You don't often come across genuine individuals in this space. I could likely count on one hand who those genuine-minded marketers might be. Someone like Russel Brunson who's developed a career out of providing true value in the field and helping to educate the uneducated is one such name. However, while Brunson has built a colossal business, the story of David Sharpe and his journey to becoming an 8-figure earner really hits home for most people.
Google's strategy works well. By focusing on the links going to and from a Web page, the search engine can organize results in a useful way. While there are a few tricks webmasters can use to improve Google standings, the best way to get a top spot is to consistently provide top quality content, which gives other people the incentive to link back to their pages.
PageRank was once available for the verified site maintainers through the Google Webmaster Tools interface. However, on October 15, 2009, a Google employee confirmed that the company had removed PageRank from its Webmaster Tools section, saying that "We've been telling people for a long time that they shouldn't focus on PageRank so much. Many site owners seem to think it's the most important metric for them to track, which is simply not true."[67] In addition, The PageRank indicator is not available in Google's own Chrome browser.
It’s hard to believe that the Internet is now multiple decades old. Affiliate marketing has been around since the earliest days of online marketing. It’s a great solution for businesses that are risk-averse or don’t have the budget to spend on upfront marketing costs. Use affiliate marketing to build a new revenue stream for your ecommerce or B2B business.
i.e. the PageRank value for a page u is dependent on the PageRank values for each page v contained in the set Bu (the set containing all pages linking to page u), divided by the number L(v) of links from page v. The algorithm involves a damping factor for the calculation of the pagerank. It is like the income tax which the govt extracts from one despite paying him itself.
Also hadn’t thought about decreasing the rank value based on the spammyness of sites a page is linking into. My guess on how to do it would be determining the spammyness of individual pages based on multiple page and site factors, then some type of reverse pagerank calcuation starting with the those bad scores, then overlaying that on top of the “good” pagerank calculation as a penalty. This is another thing which would be interesting to play around with in the Nutch algorithm.

Brian, you are such an inspiration. I wonder how do you get all these hacks and then publish them for all of us. I have been reading your stuff from quite a time now, but I have a problem. Every time I read something you post I feel overwhelmed but I haven’t been really able to generate any fruitful results on any of my sites. I just don’t know where to start. Imagine I don’t even have an email list.
I started taking action right away on the “Best Of” Blog Posts” approach… I found some great blogs and left a relevant and useful comment. The first impression, sins a lot of the blogs see my as the competition it is not easy to get past the moderator. I made 6 or 7 comments the first day and will update this comment after I have a good number of post to measure results…
A decent article which encourages discussion and healthy debate. Reading some of the comments I see it also highlights some of the misunderstandings some people (including some SEOs) have of Google PageRank. Toolbar PageRank is not the same thing as PageRank. The little green bar (Toolbar PageRank) was never a very accurate metric and told you very little about the value of any particular web page. It may have been officially killed off earlier this year, but the truth is its been dead for many years. Real PageRank on the other hand, is at the core of Google’s algorithm and remains very important.
×