It's key to understand that nobody really knows what goes into PageRank. Many believe that there are dozens if not hundreds of factors, but that the roots go back to the original concept of linking. It's not just volume of links either. Thousands of links by unauthoritative sites might be worth a handful of links from sites ranked as authoritative.

The original Random Surfer PageRank patent from Stanford has expired. The Reasonable Surfer version of PageRank (assigned to Google) is newer than that one, and has been updated via a continuation patent at least once. The version of PageRank based upon a trusted seed set of sites (assigned to Google) has also been updated via a continuation patent and differs in many ways from the Stanford version of PageRank. It is likely that Google may be using one of the versions of PageRank that they have control over (the exclusive license to use Stanford’s version of PageRank has expired along with that patent). The updated versions of PageRank (reasonable surfer and Trusted Seeds approach) both are protected under present day patents assigned to Google, and both have been updated to reflect modern processes in how they are implemented. Because of their existence, and the expiration of the original, I would suggest that it is unlikely that the random surfer model-base PageRank is still being used.


Backlinks can be time-consuming to earn. New sites or those expanding their keyword footprint may find it difficult to know where to start when it comes to link building. That's where competitive backlink research comes in: By examining the backlink profile (the collection of pages and domains linking to a website) to a competitor that's already ranking well for your target keywords, you can gain insight about the link building that may have helped them. A tool like Link Explorer can help uncover these links so you can and target those domains in your own link building campaigns.


By 2004, search engines had incorporated a wide range of undisclosed factors in their ranking algorithms to reduce the impact of link manipulation. In June 2007, The New York Times' Saul Hansell stated Google ranks sites using more than 200 different signals.[25] The leading search engines, Google, Bing, and Yahoo, do not disclose the algorithms they use to rank pages. Some SEO practitioners have studied different approaches to search engine optimization, and have shared their personal opinions.[26] Patents related to search engines can provide information to better understand search engines.[27] In 2005, Google began personalizing search results for each user. Depending on their history of previous searches, Google crafted results for logged in users.[28]
In 1998, two graduate students at Stanford University, Larry Page and Sergey Brin, developed "Backrub", a search engine that relied on a mathematical algorithm to rate the prominence of web pages. The number calculated by the algorithm, PageRank, is a function of the quantity and strength of inbound links.[21] PageRank estimates the likelihood that a given page will be reached by a web user who randomly surfs the web, and follows links from one page to another. In effect, this means that some links are stronger than others, as a higher PageRank page is more likely to be reached by the random web surfer.

If you are serious about improving search traffic and are unfamiliar with SEO, we recommend reading this guide front-to-back. We've tried to make it as concise as possible and easy to understand. There's a printable PDF version for those who'd prefer, and dozens of linked-to resources on other sites and pages that are also worthy of your attention.
Peter made a very good point in all of this, and Michael Martinez did in a backhanded way as well. Talking about a concept related PageRank sounds cool. It doesn’t actually have to be useful or practical, and it usually isn’t; but as long as the impression of something productive is given off, then that can be all that matters in the eyes of those who lack sufficient knowledge.
Thanks for the clarification, Matt. We were just wondering today when we would hear from you on the matter since it had been a couple of weeks since SMX. I think we’d all be interested to know the extent to which linking to “trusted sites,” helps PageRank. Does it really mitigate the losses incurred by increasing the number of links? I ask because it seems pretty conclusive that the total number of outbound links is now the deciding metric for passing PageRank and not the number of DoFollow links. Any thoughts from you or others?
Another excellent guide is Google’s “Search Engine Optimization Starter Guide.” This is a free PDF download that covers basic tips that Google provides to its own employees on how to get listed. You’ll find it here. Also well worth checking out is Moz’s “Beginner’s Guide To SEO,” which you’ll find here, and the SEO Success Pyramid from Small Business Search Marketing.
The Truth? You don't often come across genuine individuals in this space. I could likely count on one hand who those genuine-minded marketers might be. Someone like Russel Brunson who's developed a career out of providing true value in the field and helping to educate the uneducated is one such name. However, while Brunson has built a colossal business, the story of David Sharpe and his journey to becoming an 8-figure earner really hits home for most people.
Regarding nofollow on content that you don’t want indexed, you’re absolutely right that nofollow doesn’t prevent that, e.g. if someone else links to that content. In the case of the site that excluded user forums, quite a few high-quality pages on the site happened not to have links from other sites. In the case of my feed, it doesn’t matter much either way, but I chose not to throw any extra PageRank onto my feed url. The services that want to fetch my feed url (e.g. Google Reader or Bloglines) know how to find it just fine.
What seems to be happening is that the toolbar looks at the URL of the page the browser is displaying and strips off everything down the last “/” (i.e. it goes to the “parent” page in URL terms). If Google has a Toolbar PR for that parent then it subtracts 1 and shows that as the Toolbar PR for this page. If there’s no PR for the parent it goes to the parent’s parent’s page, but subtracting 2, and so on all the way up to the root of your site.  If it can’t find a Toolbar PR to display in this way, that is if it doesn’t find a page with a real calculated PR, then the bar is greyed out.

PageRank has been used to rank spaces or streets to predict how many people (pedestrians or vehicles) come to the individual spaces or streets.[51][52] In lexical semantics it has been used to perform Word Sense Disambiguation,[53] Semantic similarity,[54] and also to automatically rank WordNet synsets according to how strongly they possess a given semantic property, such as positivity or negativity.[55]


Google PageRank (Google PR) is one of the methods Google uses to determine a page's relevance or importance. Important pages receive a higher PageRank and are more likely to appear at the top of the search results. Google PageRank (PR) is a measure from 0 - 10. Google Pagerank is based on backlinks. The more quality backlinks the higher Google Pagerank. Improving your Google page rank (building QUALITY backlinks ) is very important if you want to improve your search engine rankings.
For instance, if you have an article called “How To Do Keyword Research,” you can help reinforce to Google the relevance of this page for the subject/phrase “keyword research” by linking from an article reviewing a keyword research tool to your How To Do Keyword Research article. This linking strategy is part of effective siloing, which helps clarify your main website themes.
A PageRank results from a mathematical algorithm based on the webgraph, created by all World Wide Web pages as nodes and hyperlinks as edges, taking into consideration authority hubs such as cnn.com or usa.gov. The rank value indicates an importance of a particular page. A hyperlink to a page counts as a vote of support. The PageRank of a page is defined recursively and depends on the number and PageRank metric of all pages that link to it ("incoming links"). A page that is linked to by many pages with high PageRank receives a high rank itself.
The original Random Surfer PageRank patent from Stanford has expired. The Reasonable Surfer version of PageRank (assigned to Google) is newer than that one, and has been updated via a continuation patent at least once. The version of PageRank based upon a trusted seed set of sites (assigned to Google) has also been updated via a continuation patent and differs in many ways from the Stanford version of PageRank. It is likely that Google may be using one of the versions of PageRank that they have control over (the exclusive license to use Stanford’s version of PageRank has expired along with that patent). The updated versions of PageRank (reasonable surfer and Trusted Seeds approach) both are protected under present day patents assigned to Google, and both have been updated to reflect modern processes in how they are implemented. Because of their existence, and the expiration of the original, I would suggest that it is unlikely that the random surfer model-base PageRank is still being used.
SEO is a marketing discipline focused on growing visibility in organic (non-paid) search engine results. SEO encompasses both the technical and creative elements required to improve rankings, drive traffic, and increase awareness in search engines. There are many aspects to SEO, from the words on your page to the way other sites link to you on the web. Sometimes SEO is simply a matter of making sure your site is structured in a way that search engines understand.
In the 2000s, with more and more Internet users and the birth of iPhone, customers started searching products and making decisions about their needs online first, instead of consulting a salesperson, which created a new problem for the marketing department of a company. In addition, a survey in 2000 in the United Kingdom found that most retailers had not registered their own domain address.[12] These problems made marketers find the digital ways for market development.

Organic SEO's flip-side offers up a paid method for marketing on search engines like Google. SEM provides an avenue for displaying ads through networks such as Google's Adwords and other paid search platforms that exist across the web throughout social media sites like Facebook, Instagram and even video sites like YouTube, which, invariably, is the world's second largest search engine.
For instance, you might use Facebook’s Lookalike Audiences to get your message in front of an audience similar to your core demographic. Or, you could pay a social media influencer to share images of your products to her already well-established community. Paid social media can attract new customers to your brand or product, but you’ll want to conduct market research and A/B testing before investing too much in one social media channel.
If you want to concentrate the PR into one, or a few, pages then hierarchical linking will do that. If you want to average out the PR amongst the pages then "fully meshing" the site (lots of evenly distributed links) will do that - examples 5, 6, and 7 in my above. (NB. this is where Ridings’ goes wrong, in his MiniRank model feedback loops will increase PR - indefinitely!)
Content is king. It always has been and it always will be. Creating insightful, engaging and unique content should be at the heart of any online marketing strategy. Too often, people simply don't obey this rule. The problem? This takes an extraordinary amount of work. However, anyone that tells you that content isn't important, is not being fully transparent with you. You cannot excel in marketing anything on the internet without having quality content.
SEO is a marketing discipline focused on growing visibility in organic (non-paid) search engine results. SEO encompasses both the technical and creative elements required to improve rankings, drive traffic, and increase awareness in search engines. There are many aspects to SEO, from the words on your page to the way other sites link to you on the web. Sometimes SEO is simply a matter of making sure your site is structured in a way that search engines understand.

I won’t blame MC. Google, knows what they does. These are things that webmasters need not worry about. Well, it won’t make much difference as far as I think. I don’t use no follow tags specifically – I use WP for blogging purposes and it does rest of the things for me other than writing content which I do. I think it is the content and the external links that sites point to – which should be considered. I mean, if a computer blog owner posts a really fantastic computer article about something related to computer, and also puts some links to external pages (which are really useful for the readers), then that post, should be ranked high in gooogle – And I think google does this well – So, webmasters, just concentrate on yur website/blogs etc and leave rest of the things to Big G.
Google's founders, in their original paper,[18] reported that the PageRank algorithm for a network consisting of 322 million links (in-edges and out-edges) converges to within a tolerable limit in 52 iterations. The convergence in a network of half the above size took approximately 45 iterations. Through this data, they concluded the algorithm can be scaled very well and that the scaling factor for extremely large networks would be roughly linear in {\displaystyle \log n} , where n is the size of the network.
I have a question. My site is in download/free resources niche, and I’m having some troubles figuring out link building opportunities. Most of the webmasters from other websites in this niche don’t really actively respond to my email. I guess it’s because this is a niche that people often build a high traffic website and leave it there as a passive income. Do you have any suggestion?
What an article… thank you so much for the priceless information, we will be changing our pages around to make sure we get the highest page rank available to us, we are trying to get high page rank sites to link to us, hopefully there is more information out there to gather as we want to compete within our market to gain as much market-share as possible.
The internet was the little guy savior, simple sites could rank well locally. Sadly your company is in the process of destroying that. In this economy small business with zero page rank that are listed on page 22 of results, need to be found in order to survive. My customers are really suffering because of the work that is coming out of Google, it keeps getting worse. Their conversions are still good coming out of Yahoo and MSN and now Bing. They do not have the resources to produce blogs, forums, or $5,000 websites let alone pay for Adwords when they are just trying to pay rent and not a lot of people can do their own web production.
Another reason to achieve quality backlinks is to entice visitors to come to your website. You can't build a website, and then expect that people will find your website without pointing the way. You will probably have to get the word out there about your site. One way webmasters got the word out used to be through reciprocal linking. Let's talk about reciprocal linking for a moment.

Totally agree — more does not always equal better. Google takes a sort of ‘Birds of a Feather’ approach when analyzing inbound links, so it’s really all about associating yourself (via inbound links) with websites Google deems high quality and trustworthy so that Google deems YOUR web page high quality and trustworthy. As you mentioned, trying to cut corners, buy links, do one-for-one trades, or otherwise game/manipulate the system never works. The algorithm is too smart. 

There is another way to gain quality backlinks to your site, in addition to related site themes: anchor text. When a link incorporates a keyword into the text of the hyperlink, we call this quality anchor text. A link's anchor text may be one of the under-estimated resources a webmaster has. Instead of using words like "click here" which probably won't relate in any way to your website, using the words "Please visit our tips page for how to nurse an orphaned kitten" is a far better way to utilize a hyperlink. A good tool for helping you find your backlinks and what text is being used to link to your site is the Backlink Anchor Text Analysis Tool. If you find that your site is being linked to from another website, but the anchor text is not being utilized properly, you should request that the website change the anchor text to something incorporating relevant keywords. This will also help boost your quality backlinks score.

Organic SEO's flip-side offers up a paid method for marketing on search engines like Google. SEM provides an avenue for displaying ads through networks such as Google's Adwords and other paid search platforms that exist across the web throughout social media sites like Facebook, Instagram and even video sites like YouTube, which, invariably, is the world's second largest search engine.
An SEO technique is considered white hat if it conforms to the search engines' guidelines and involves no deception. As the search engine guidelines[18][19][51] are not written as a series of rules or commandments, this is an important distinction to note. White hat SEO is not just about following guidelines, but is about ensuring that the content a search engine indexes and subsequently ranks is the same content a user will see. White hat advice is generally summed up as creating content for users, not for search engines, and then making that content easily accessible to the online "spider" algorithms, rather than attempting to trick the algorithm from its intended purpose. White hat SEO is in many ways similar to web development that promotes accessibility,[52] although the two are not identical.
Backlinks take place across the Internet when one website mentions another website and links to it. Also, referred to as “incoming links,” backlinks make their connection through external websites. These links from outside domains point to pages on your own domain. Whenever backlinks occur, it is like receiving a vote for a webpage. The more votes you get from the authoritative sites creates a positive effect on a site’s ranking and search visibility.
In the past, the PageRank shown in the Toolbar was easily manipulated. Redirection from one page to another, either via a HTTP 302 response or a "Refresh" meta tag, caused the source page to acquire the PageRank of the destination page. Hence, a new page with PR 0 and no incoming links could have acquired PR 10 by redirecting to the Google home page. This spoofing technique was a known vulnerability. Spoofing can generally be detected by performing a Google search for a source URL; if the URL of an entirely different site is displayed in the results, the latter URL may represent the destination of a redirection.
Also, by means of the iterative calculation, the sum of all pages' PageRanks still converges to the total number of web pages. So the average PageRank of a web page is 1. The minimum PageRank of a page is given by (1-d). Therefore, there is a maximum PageRank for a page which is given by dN+(1-d), where N is total number of web pages. This maximum can theoretically occur, if all web pages solely link to one page, and this page also solely links to itself.
Despite this many people seem to get it wrong! In particular “Chris Ridings of www.searchenginesystems.net” has written a paper entitled “PageRank Explained: Everything you’ve always wanted to know about PageRank”, pointed to by many people, that contains a fundamental mistake early on in the explanation! Unfortunately this means some of the recommendations in the paper are not quite accurate.
But how do you get quoted in news articles? Websites such as HARO and ProfNet can help you to connect with journalists who have specific needs, and there are other tools that allow you to send interesting pitches to writers. Even monitoring Twitter for relevant conversations between journalists can yield opportunities to connect with writers working on pieces involving your industry.
Was reviewing some competitive data and thought this was pretty interesting. I ran a batch analysis on Ahrefs of competitors. See attached screenshot. With just 603 backlinks, Our site is ranking up there with sites with 2x, 3x, 10x the number of backlinks/unique ips. Guessing some of this authority is coming from the backlinks program and general good quality of those links. Hard to speculate but nice to see. Ben R.
It doesn’t mean than you have to advertise on these social media platforms. It means that they belong to that pyramid which will function better thanks to their support. Just secure them and decide which of them will suit your goal better. For example, you can choose Instagram because its audience is the most suitable for mobile devices and bits of advice of their exploitation distribution.
Could the nofollow change could be interpreted as a form of usability guidance? For instance, I’ve recently removed drop-down menus from a handful of sites because of internal link and keyword density issues. This wasn’t done randomly. Tests were done to measure usage and value of this form of navigation that made it easy to make the change – allowing usability and SEO to dovetail nicely.

Mega-sites, like http://news.bbc.co.uk have tens or hundreds of editors writing new content – i.e. new pages - all day long! Each one of those pages has rich, worthwile content of its own and a link back to its parent or the home page! That’s why the Home page Toolbar PR of these sites is 9/10 and the rest of us just get pushed lower and lower by comparison…

As I was telling Norman above, these days what we’ve come to call content marketing is really a big part of “link building.” You can’t buy links, and “you link to me I’ll link to you” requests often land on deaf ears. Its really all about creating high quality content (videos, images, written blog posts) that appeals to the needs/wants of your target market, and then naturally earning inbound links from sources that truly find what you have to offer worth referencing.
In December 2009, Google announced it would be using the web search history of all its users in order to populate search results.[32] On June 8, 2010 a new web indexing system called Google Caffeine was announced. Designed to allow users to find news results, forum posts and other content much sooner after publishing than before, Google caffeine was a change to the way Google updated its index in order to make things show up quicker on Google than before. According to Carrie Grimes, the software engineer who announced Caffeine for Google, "Caffeine provides 50 percent fresher results for web searches than our last index..."[33] Google Instant, real-time-search, was introduced in late 2010 in an attempt to make search results more timely and relevant. Historically site administrators have spent months or even years optimizing a website to increase search rankings. With the growth in popularity of social media sites and blogs the leading engines made changes to their algorithms to allow fresh content to rank quickly within the search results.[34]
Hey brian, it is extremely fantastic stuff, i am not getting words to appreciate your work..brilliant. No one dares to share their business secrets with others but you are awesome and thank you so much. Iam a beginner in digital marketing, iam learning consistently by following your posts, tips and tricks. eventually i became an intermediate person thanks for your help.
We must be careful with our reciprocal links. There is a Google patent in the works that will deal with not only the popularity of the sites being linked to, but also how trustworthy a site is that you link to from your own website. This will mean that you could get into trouble with the search engine just for linking to a bad apple. We could begin preparing for this future change in the search engine algorithm by being choosier with which we exchange links right now. By choosing only relevant sites to link with, and sites that don't have tons of outbound links on a page, or sites that don't practice black-hat SEO techniques, we will have a better chance that our reciprocal links won't be discounted.
Place strategic search phrases on pages. Integrate selected keywords into your website source code and existing content on designated pages. Make sure to apply a sug­gested guideline of one to three keywords/phrases per content page and add more pages to complete the list. Ensure that related words are used as a natural inclu­sion of your keywords. It helps the search engines quickly determine what the page is about. A natural approach to this works best. In the past, 100 to 300 words on a page was recommended. Many tests show that pages with 800 to 2,000 words can outperform shorter ones. In the end, the users, the marketplace, content and links will determine the popularity and ranking numbers.
Cross-platform measurement: The number of marketing channels continues to expand, as measurement practices are growing in complexity. A cross-platform view must be used to unify audience measurement and media planning. Market researchers need to understand how the Omni-channel affects consumer's behaviour, although when advertisements are on a consumer's device this does not get measured. Significant aspects to cross-platform measurement involves de-duplication and understanding that you have reached an incremental level with another platform, rather than delivering more impressions against people that have previously been reached (Whiteside, 2016).[42] An example is ‘ESPN and comScore partnered on Project Blueprint discovering the sports broadcaster achieved a 21% increase in unduplicated daily reach thanks to digital advertising’ (Whiteside, 2016).[42] Television and radio industries are the electronic media, which competes with digital and other technological advertising. Yet television advertising is not directly competing with online digital advertising due to being able to cross platform with digital technology. Radio also gains power through cross platforms, in online streaming content. Television and radio continue to persuade and affect the audience, across multiple platforms (Fill, Hughes, & De Franceso, 2013).[45]
But if you do it properly, it can be worth your money. Also, press releases can be much more than just a block of text. In December 2018, we ran a press release through Business Wire that had multiple backlinks, stylized call outs, and even a video! If you put effort into them, press releases can be not just a source of backlinks, but also serve as a great marketing piece as well. 
×