Hi Bill, Yes – thanks. I think I’ll have to do more of these. I couldn’t really go beyond Pagerank in an 18 minute Pubcon session. Although the random surfer model expired (and wasn’t even assigned to Google), it is still a precursor to understanding everything that has come after it. I think I would love to do more videos/presentations on both Reasonable surfer patent, Dangling Nodes and probably a lifetime of other videos in the future. To be able to demonstrate these concept without giving people headaches, though, the PageRank algorithm in Matrix form provides a good understanding of why you can’t "just get links" and expect everything to be at number 1.

A decent article which encourages discussion and healthy debate. Reading some of the comments I see it also highlights some of the misunderstandings some people (including some SEOs) have of Google PageRank. Toolbar PageRank is not the same thing as PageRank. The little green bar (Toolbar PageRank) was never a very accurate metric and told you very little about the value of any particular web page. It may have been officially killed off earlier this year, but the truth is its been dead for many years. Real PageRank on the other hand, is at the core of Google’s algorithm and remains very important.


Before Sharpe ever came into close proximity with the internet marketing field, he was a construction worker. Needing a way to make ends meet, like millions of other people around the world, he turned to a field that could hopefully pay the bills. But try as he might, he was never able to actually get ahead. Until one day, when Sharpe discovered the amount of money being made online by internet marketers, his entire mindset changed.
Before online marketing channels emerged, the cost to market products or services was often prohibitively expensive, and traditionally difficult to measure. Think of national television ad campaigns, which are measured through consumer focus groups to determine levels of brand awareness. These methods are also not well-suited to controlled experimentation. Today, anyone with an online business (as well as most offline businesses) can participate in online marketing by creating a website and building customer acquisition campaigns at little to no cost. Those marketing products and services also have the ability to experiment with optimization to fine-tune their campaigns’ efficiency and ROI.
Nice word is not enough for this. You show that Blogging is like Apple vs Samsung. You can create lot of post and drive traffic (which is Samsung like lot of phone every year) or you can create high quality post like apple (which is you) and force higher rank site to make content like you copy content from you blog. Now i will work hard on already publish post until they will not get traffic.

Heading tags. Always use H tags to optimize your content layout. Try and use variations on your keyphrases in some headings, too. Don’t repeat keyphrases in headings unless it’s absolutely necessary. (This doesn’t stop you from needing to repeat the keyphrase in the body of your content). H tags are HTML codes – you can find a link to HTML codes and how to use them at the end of this section.
In order to be a data driven agency, we foster a culture of inspired marketing entrepreneurs that collaborate, innovate, and are constantly pushing the threshold of marketing intelligence. Our analytics team is well versed in mathematics, business analytics, multi-channel attribution modeling, creating custom analytics reporting dashboards, and performing detailed analysis and reporting for each client.
A Web crawler may use PageRank as one of a number of importance metrics it uses to determine which URL to visit during a crawl of the web. One of the early working papers[56] that were used in the creation of Google is Efficient crawling through URL ordering,[57] which discusses the use of a number of different importance metrics to determine how deeply, and how much of a site Google will crawl. PageRank is presented as one of a number of these importance metrics, though there are others listed such as the number of inbound and outbound links for a URL, and the distance from the root directory on a site to the URL.
Also hadn’t thought about decreasing the rank value based on the spammyness of sites a page is linking into. My guess on how to do it would be determining the spammyness of individual pages based on multiple page and site factors, then some type of reverse pagerank calcuation starting with the those bad scores, then overlaying that on top of the “good” pagerank calculation as a penalty. This is another thing which would be interesting to play around with in the Nutch algorithm.
An omni-channel approach not only benefits consumers but also benefits business bottom line: Research suggests that customers spend more than double when purchasing through an omni-channel retailer as opposed to a single-channel retailer, and are often more loyal. This could be due to the ease of purchase and the wider availability of products.[24]
Targeting, viewability, brand safety and invalid traffic: Targeting, viewability, brand safety and invalid traffic all are aspects used by marketers to help advocate digital advertising. Cookies are a form of digital advertising, which are tracking tools within desktop devices; causing difficulty, with shortcomings including deletion by web browsers, the inability to sort between multiple users of a device, inaccurate estimates for unique visitors, overstating reach, understanding frequency, problems with ad servers, which cannot distinguish between when cookies have been deleted and when consumers have not previously been exposed to an ad. Due to the inaccuracies influenced by cookies, demographics in the target market are low and vary (Whiteside, 2016).[42] Another element, which is affected within digital marketing, is ‘viewabilty’ or whether the ad was actually seen by the consumer. Many ads are not seen by a consumer and may never reach the right demographic segment. Brand safety is another issue of whether or not the ad was produced in the context of being unethical or having offensive content. Recognizing fraud when an ad is exposed is another challenge marketers face. This relates to invalid traffic as premium sites are more effective at detecting fraudulent traffic, although non-premium sites are more so the problem (Whiteside, 2016).[42]
Hi Brian thank you for sharing this awesome backlinking techniques. My site is currently not ranking well. It used to be, sometime mid last year, but it suddenly got de-ranked. Not really sure why. I haven’t been participating in any blackhat techniques or anything at all. I’ll try a few of your tips and hopefully it will help my site back to its shape.
Moreover, the PageRank mechanism is entirely general, so it can applied to any graph or network in any field. Currently, the PR formula is used in bibliometrics, social and information network analysis, and for link prediction and recommendation. It's even used for system analysis of road networks, as well as biology, chemistry, neuroscience, and physics.
On a blog the page rank should go to the main article pages. Now it just gets “evaporated” if you use “nofollow” or scattered to all the far flung nooks and crannys which means google will not be able to see the wood for the trees. The vast majority of a site’s overall page rank will now reside in the long tail of useless pages such as commentors profile pages. This can only make it harder for google to serve up the most relevant pages.
Search engines want websites to have a level playing field, and look for natural links built slowly over time. While it is fairly easy to manipulate links on a web page to try to achieve a higher ranking, it is a lot harder to influence a search engine with external backlinks from other websites. This is also a reason why backlinks factor in so highly into a search engine's algorithm. Lately, however, a search engine's criteria for quality inbound links has gotten even tougher, thanks to unscrupulous webmasters trying to achieve these inbound links by deceptive or sneaky techniques, such as with hidden links, or automatically generated pages whose sole purpose is to provide inbound links to websites. These pages are called link farms, and they are not only disregarded by search engines, but linking to a link farm could get your site banned entirely.
A strategy that is linked into the effectiveness of digital marketing is content marketing.[39] Content marketing can be briefly described as "delivering the content that your audience is seeking in the places that they are searching for it".[39] It is found that content marketing is highly present in digital marketing and becomes highly successful when content marketing is involved. This is due to content marketing making your brand more relevant to the target consumers, as well as more visible to the target consumer.
Yes, the more links on a page the smaller the amount of page rank it can pass on to each, but that was as it was before. With regard to what happens to the ‘missing’ page rank, it seems that if this is the case all over the Internet, and it will be, the total amount of page rank flow is reduced the same all over so you don’t need as much page rank flow to your good links to maintain relative position.

Google's strategy works well. By focusing on the links going to and from a Web page, the search engine can organize results in a useful way. While there are a few tricks webmasters can use to improve Google standings, the best way to get a top spot is to consistently provide top quality content, which gives other people the incentive to link back to their pages.


I find it amazing that youtube.com has been “nofollowing” its featured videos for last 12 months (still doing it as I type) when it now seems that this means “i dont trust this content” and “i want to page rank to flow to this content”. In fact a quick glance at a youtube page tells you that youtube are currently flushing 50% of their page rank (very approx) down the toilet on every page.

Suppose instead that page B had a link to pages C and A, page C had a link to page A, and page D had links to all three pages. Thus, upon the first iteration, page B would transfer half of its existing value, or 0.125, to page A and the other half, or 0.125, to page C. Page C would transfer all of its existing value, 0.25, to the only page it links to, A. Since D had three outbound links, it would transfer one third of its existing value, or approximately 0.083, to A. At the completion of this iteration, page A will have a PageRank of approximately 0.458.


When running PPC ads, it's important that you keep careful track of the specific ads and keywords that you're targeting. You can do this by using the Google Analytics UTM builder to create campaign URLs that you can use to track the campaign source, the medium and any keywords or terms that you might be targeting. This way, you can determine the effectiveness of any campaign that you run and figure out the precise conversion rate.

If you are serious about improving web traffic to your website, we recommend you read Google Webmasters and Webmaster Guidelines. These contain the best practices to help Google (and other search engines) find, crawl, and index your website. After you have read them, you MUST try our Search Engine Optimization Tools to help you with Keyword Research, Link Building, Technical Optimization, Usability, Social Media Strategy and more.
In the beginning, it was rough for Sharpe. No one out there should think that it's going to be easy whatsoever. His journey took years and years to go from an absolute beginner, to a fluid and seasoned professional, able to clearly visualize and achieve his dreams, conveying his vast knowledge expertly to those hungry-minded individuals out there looking to learn how to generate a respectable income online.
By using Internet platforms, businesses can create competitive advantage through various means. To reach the maximum potential of digital marketing, firms use social media as its main tool to create a channel of information. Through this a business can create a system in which they are able to pinpoint behavioral patterns of clients and feedback on their needs.[30] This means of content has shown to have a larger impingement on those who have a long-standing relationship with the firm and with consumers who are relatively active social media users. Relative to this, creating a social media page will further increase relation quality between new consumers and existing consumers as well as consistent brand reinforcement therefore improving brand awareness resulting in a possible rise for consumers up the Brand Awareness Pyramid.[31] Although there may be inconstancy with product images;[32] maintaining a successful social media presence requires a business to be consistent in interactions through creating a two way feed of information; firms consider their content based on the feedback received through this channel, this is a result of the environment being dynamic due to the global nature of the internet.[29] Effective use of digital marketing can result in relatively lowered costs in relation to traditional means of marketing; Lowered external service costs, advertising costs, promotion costs, processing costs, interface design costs and control costs.[32]
It is interesting to read the article which however as others have commented leaves questions unanswered. The ideal internal page rank flow is clearly down the hierarchy and back up again evidently and no page to page links down the line. If this effect is big enough to have provoked an algorithm change then it must be substantial. Removing those related product links altogether would improve ranking and degrade the user experience of the site which surely is undesirable. I suspect the lesser of two evils was chosen.
But I also don’t wanna lose PageRank on every comment with a link… If I can give PageRank and lose none, I wanna let the comment there, even without nofollow. But if I lose PageRank on every link, even inside original post, EVEN MORE if nofollow also takes PageRank out of me, I may just start using JavaScript or simple text without anchor for links… I definetely don’t like this idea, but I dislike even more losing PageRank on each outlink on my site. I’d just link top quality sites that I actively wanna vote for Search Engines. 
×