Is very telling and an important thing to consider. Taking the model of a university paper on a particular subject as an example, you would expect the paper to cite (link to) other respected papers in the same field in order to demonstrate that it is couched in some authority. As PageRank is based on the citation model used in university work, it makes perfect sense to incorporate a “pages linked to” factor into the equation.
So, when you find a relevant forum, be sure that you have written an authorized profile description and toss in your main concept or word of great significance. Then study the forum, its rules, and the way it operates. Examine the forum to know whether its members share links in threads. Become a reliable person making more and more friends and placing posts interesting for the forum participants. Thanks to that you may get more internal linkage to your profile and gain authority. And, of course, threads will build your credibility.Why do you need all that?
Search engines find and catalog web pages through spidering (also known as webcrawling) software. Spidering software "crawls" through the internet and grabs information from websites which is used to build search engine indexes. Unfortunately, not all search engine spidering software works the same way, so what gives a page a high ranking on one search engine may not necessarily give it a high ranking on another. Note that rather than waiting for a search engine to discover a newly created page, web designers can submit the page directly to search engines for cataloging.
Some search engines have also reached out to the SEO industry, and are frequent sponsors and guests at SEO conferences, webchats, and seminars. Major search engines provide information and guidelines to help with website optimization.[18][19] Google has a Sitemaps program to help webmasters learn if Google is having any problems indexing their website and also provides data on Google traffic to the website.[20] Bing Webmaster Tools provides a way for webmasters to submit a sitemap and web feeds, allows users to determine the "crawl rate", and track the web pages index status. 

Also hadn’t thought about decreasing the rank value based on the spammyness of sites a page is linking into. My guess on how to do it would be determining the spammyness of individual pages based on multiple page and site factors, then some type of reverse pagerank calcuation starting with the those bad scores, then overlaying that on top of the “good” pagerank calculation as a penalty. This is another thing which would be interesting to play around with in the Nutch algorithm.
Companies that employ overly aggressive techniques can get their client websites banned from the search results. In 2005, the Wall Street Journal reported on a company, Traffic Power, which allegedly used high-risk techniques and failed to disclose those risks to its clients.[15] Wired magazine reported that the same company sued blogger and SEO Aaron Wall for writing about the ban.[16] Google's Matt Cutts later confirmed that Google did in fact ban Traffic Power and some of its clients.[17]

Some search engines have also reached out to the SEO industry, and are frequent sponsors and guests at SEO conferences, webchats, and seminars. Major search engines provide information and guidelines to help with website optimization.[18][19] Google has a Sitemaps program to help webmasters learn if Google is having any problems indexing their website and also provides data on Google traffic to the website.[20] Bing Webmaster Tools provides a way for webmasters to submit a sitemap and web feeds, allows users to determine the "crawl rate", and track the web pages index status.


Matt, you don’t mention the use of disallow pages via robots.txt. I’ve read that PageRank can be better utilised by disallowing pages that probably don’t add value to users searching on engines. For example, Privacy Policy and Terms of Use pages. These often appear in the footer of a website and are required by EU law on every page of the site. Will it boost the other pages of the site if these pages are added to robots.txt like so?
I would like to know how Google is handling relevancy with so many websites now jumping on the “no follow” wagon? Seems like just about every major website has no follow links, so with the Panda updates this year what’s happening to all that lost link power? Seem’s like this tactic will stagnate the growth of up-and-coming websites on the internet to me. Am I right here?
Just think about any relationship for a moment. How long you've known a person is incredibly important. It's not the be-all-end-all, but it is fundamental to trust. If you've known someone for years and years and other people that you know who you already trust can vouch for that person, then you're far more likely to trust them, right? But if you've just met someone, and haven't really vetted them so to speak, how can you possibly trust them?

One thing is certain: interlinking sites doesn't help you from a search engine standpoint. The only reason you may want to interlink your sites in the first place might be to provide your visitors with extra resources to visit. In this case, it would probably be okay to provide visitors with a link to another of your websites, but try to keep many instances of linking to the same IP address to a bare minimum. One or two links on a page here and there probably won't hurt you.


Once you understand how everything works, and your expectations are set the right way, decide what you want to do. Do you want to become an affiliate marketer? Do you want to be a network marketer? Do you want to become a blogger and sell your own products? Squeeze pages, which are glorified sales pages that attract people and direct their attention towards a single action of providing their email address, are created in a variety of methods. The better they are, the more likely they'll convert.
Some search engines have also reached out to the SEO industry, and are frequent sponsors and guests at SEO conferences, webchats, and seminars. Major search engines provide information and guidelines to help with website optimization.[18][19] Google has a Sitemaps program to help webmasters learn if Google is having any problems indexing their website and also provides data on Google traffic to the website.[20] Bing Webmaster Tools provides a way for webmasters to submit a sitemap and web feeds, allows users to determine the "crawl rate", and track the web pages index status. 

Another illicit practice is to place "doorway" pages loaded with keywords on the client's site somewhere. The SEO promises this will make the page more relevant for more queries. This is inherently false since individual pages are rarely relevant for a wide range of keywords. More insidious, however, is that these doorway pages often contain hidden links to the SEO's other clients as well. Such doorway pages drain away the link popularity of a site and route it to the SEO and its other clients, which may include sites with unsavory or illegal content.
Let’s say that I want to link to some popular search results on my catalog or directory site – you know, to give a new user an alternative way of sampling the site. Of course, following Google’s advice, I have to “avoid allowing search result-like pages to be crawled”. Now, I happen to think that these pages are great for the new user, but I accept Google’s advice and block them using robots.txt.
A decent article which encourages discussion and healthy debate. Reading some of the comments I see it also highlights some of the misunderstandings some people (including some SEOs) have of Google PageRank. Toolbar PageRank is not the same thing as PageRank. The little green bar (Toolbar PageRank) was never a very accurate metric and told you very little about the value of any particular web page. It may have been officially killed off earlier this year, but the truth is its been dead for many years. Real PageRank on the other hand, is at the core of Google’s algorithm and remains very important.

If I’m writing a page about the use of the vCard microformat on a page, it absolutely makes sense for me to link out to the definition where it was originally published, and improves user experience as well as lending authority to my arguments. Often as SEOs we get obsessed with the little things, claiming that its hard to get links on particular subjects, and that is pretty true, but its mainly our own selfishness in linking out to authority content that prevents other people giving us the same courtesy.

The Open Directory Project (ODP) is a Web directory maintained by a large staff of volunteers. Each volunteer oversees a category, and together volunteers list and categorize Web sites into a huge, comprehensive directory. Because a real person evaluates and categorizes each page within the directory, search engines like Google use the ODP as a database for search results. Getting a site listed on the ODP often means it will show up on Google.

Something a lot of people seem to have overlooked was hinted at in Greg Boser’s comment above. Greg identified that there is a major (and unfair) disparity with how authority sites such as Wikipedia disrupt the linkscape by run-of-site nofollows. Once Wikipedia implemented the no-follows, previously high-value links from Wikipedia were rendered worthless making the site less of a target for spammers. Increasingly large sites are following suit in order to cleanse their own pages of spam.

In the past, the PageRank shown in the Toolbar was easily manipulated. Redirection from one page to another, either via a HTTP 302 response or a "Refresh" meta tag, caused the source page to acquire the PageRank of the destination page. Hence, a new page with PR 0 and no incoming links could have acquired PR 10 by redirecting to the Google home page. This spoofing technique was a known vulnerability. Spoofing can generally be detected by performing a Google search for a source URL; if the URL of an entirely different site is displayed in the results, the latter URL may represent the destination of a redirection.
PageRank is one of many, many factors used to produce search rankings. Highlighting PageRank in search results doesn’t help the searcher. That’s because Google uses another system to show the most important pages for a particular search you do. It lists them in order of importance for what you searched on. Adding PageRank scores to search results would just confuse people. They’d wonder why pages with lower scores were outranking higher scored pages.

It is interesting to read the article which however as others have commented leaves questions unanswered. The ideal internal page rank flow is clearly down the hierarchy and back up again evidently and no page to page links down the line. If this effect is big enough to have provoked an algorithm change then it must be substantial. Removing those related product links altogether would improve ranking and degrade the user experience of the site which surely is undesirable. I suspect the lesser of two evils was chosen.
You can confer some of your site's reputation to another site when your site links to it. Sometimes users can take advantage of this by adding links to their own site in your comment sections or message boards. Or sometimes you might mention a site in a negative way and don't want to confer any of your reputation upon it. For example, imagine that you're writing a blog post on the topic of comment spamming and you want to call out a site that recently comment spammed your blog. You want to warn others of the site, so you include the link to it in your content; however, you certainly don't want to give the site some of your reputation from your link. This would be a good time to use nofollow.
Totally agree — more does not always equal better. Google takes a sort of ‘Birds of a Feather’ approach when analyzing inbound links, so it’s really all about associating yourself (via inbound links) with websites Google deems high quality and trustworthy so that Google deems YOUR web page high quality and trustworthy. As you mentioned, trying to cut corners, buy links, do one-for-one trades, or otherwise game/manipulate the system never works. The algorithm is too smart.
Links still matter as part of the algorithmic secret sauce. The influence of a site’s link profile is plain to see in its search engine rankings, whether for better or worse, and changes in that link profile cause noticeable movement up or down the SERP. An SEO’s emphasis today should be on attracting links to quality content naturally, not building them en masse. (For more on proper link building today, see http://bit.ly/1XIm3vf )
However, the biggest contributing factors to a backlink’s effect on your rank is the website it’s coming from, measured by the acronym ART: authority, a measure of a site’s prestige/reliability — .edu and .gov sites are particularly high-authority); relevance, a measure of how related the site hosting the link is to the content; and trust, which is not an official Google metric, but relates to how much a site plays by the rules of search (i.e. not selling links) and provides good content.
So what happens when you have a page with “ten PageRank points” and ten outgoing links, and five of those links are nofollowed? […] Originally, the five links without nofollow would have flowed two points of PageRank each […] More than a year ago, Google changed how the PageRank flows so that the five links without nofollow would flow one point of PageRank each.
More appropriately, blame Google for ever making the PageRank score visible. When Google first started, PageRank was something it talked about as part of its research papers, press releases and technology pages to promote itself as a smarter search engine than well-established and bigger rivals at the time — players like Yahoo, AltaVista and Lycos, to name a few.
Just think about any relationship for a moment. How long you've known a person is incredibly important. It's not the be-all-end-all, but it is fundamental to trust. If you've known someone for years and years and other people that you know who you already trust can vouch for that person, then you're far more likely to trust them, right? But if you've just met someone, and haven't really vetted them so to speak, how can you possibly trust them?

It’s not a secret that Google appreciates business citations and listings. They are a part of its search algorithm. It’s a strong fact that must make you choose business links for your SEO campaign. The other benefit is that because of them you can receive unoptimized and DoFollow links. These links can guarantee trustworthy neighboring of your site that will attract Internet users and clients. Google considers these platforms as trustworthy and knows that they attract other business clients. In other words, almost all of them are accepted as 100% relevant.
Brunson talks about this reverse engineering in his book called, Dot Com Secrets, a homage to the internet marketing industry, and quite possibly one of the best and most transparent books around in the field. Communication is what will bridge the divide between making no money and becoming a massive six or seven-figure earner. Be straight with people and learn to communicate effectively and understand every stage of the process and you'll prosper as an internet marketer.

I have a small service business called Eco Star Painting in Calgary and I do all of my own SEO. I’m having trouble getting good backlinks. How do you suggest a painting company get quality backlinks other than the typical local citation sites and social media platforms? I don’t know what I can offer another high domain site in terms of content. Do you have any suggestions?

In the past, the PageRank shown in the Toolbar was easily manipulated. Redirection from one page to another, either via a HTTP 302 response or a "Refresh" meta tag, caused the source page to acquire the PageRank of the destination page. Hence, a new page with PR 0 and no incoming links could have acquired PR 10 by redirecting to the Google home page. This spoofing technique was a known vulnerability. Spoofing can generally be detected by performing a Google search for a source URL; if the URL of an entirely different site is displayed in the results, the latter URL may represent the destination of a redirection.
All major crawler-based search engines leverage links from across of the web, but none of them report a static “importance” score in the way Google does via its Google Toolbar. That score, while a great resource for surfers, has also provided one of the few windows into how Google ranks web pages. Some webmasters, desperate to get inside Google, keep flying into that window like confused birds, smacking their heads and losing their orientation….
So, for example, a short-tail keyphrase might be “Logo design”. Putting that into Google will get you an awful lot of hits. There’s a lot of competition for that phrase, and it’s not particularly useful for your business, either. There are no buying signals in the phrase – so many people will use this phrase to learn about logo design or to examine other aspects of logo design work.
For instance, the Pew Internet & American Life Project has demographic data that suggests individuals between the ages of 18 and 33 are the most likely to use mobile Internet technology like smartphones and tablets, while the “Gen-X” demographic of individuals who are in their 30’s and 40’s are far more likely to seek out information through their laptop and desktop computers.(See also Targeted Marketing)
The world is mobile today. Most people are searching on Google using a mobile device. The desktop version of a site might be difficult to view and use on a mobile device. As a result, having a mobile ready site is critical to your online presence. In fact, starting in late 2016, Google has begun experiments to primarily use the mobile version of a site's content41 for ranking, parsing structured data, and generating snippets.
A disadvantage of digital advertising is the large amount of competing goods and services that are also using the same digital marketing strategies. For example, when someone searches for a specific product from a specific company online, if a similar company uses targeted advertising online then they can appear on the customer's home page, allowing the customer to look at alternative options for a cheaper price or better quality of the same product or a quicker way of finding what they want online.

From a customer experience perspective, we currently have three duplicate links to the same URL i.e. i.e. ????.com/abcde These links are helpful for the visitor to locate relevant pages on our website. However, my question is; does Google count all three of these links and pass all the value, or does Google only transfer the weight from one of these links. If it only transfers value from one of these links, does the link juice disappear from the two other links to the same page, or have these links never been given any value?
The Google algorithm's most important feature is arguably the PageRank system, a patented automated process that determines where each search result appears on Google's search engine return page. Most users tend to concentrate on the first few search results, so getting a spot at the top of the list usually means more user traffic. So how does Google determine search results standings? Many people have taken a stab at figuring out the exact formula, but Google keeps the official algorithm a secret. What we do know is this:
In an effort to manually control the flow of PageRank among pages within a website, many webmasters practice what is known as PageRank Sculpting[65]—which is the act of strategically placing the nofollow attribute on certain internal links of a website in order to funnel PageRank towards those pages the webmaster deemed most important. This tactic has been used since the inception of the nofollow attribute, but may no longer be effective since Google announced that blocking PageRank transfer with nofollow does not redirect that PageRank to other links.[66]
Finally, it’s critical you spend time and resources on your business’s website design. When these aforementioned customers find your website, they’ll likely feel deterred from trusting your brand and purchasing your product if they find your site confusing or unhelpful. For this reason, it’s important you take the time to create a user-friendly (and mobile-friendly) website.
Backlinks are a major ranking factor for most search engines, including Google. If you want to do SEO for your website and get relevant organic traffic, building backlinks is something you should be doing. The more backlinks your website has from authoritative domains, the higher reputation you’ll have in Google’s eyes. And you’ll dominate the SERPS.
Digital marketing planning is a term used in marketing management. It describes the first stage of forming a digital marketing strategy for the wider digital marketing system. The difference between digital and traditional marketing planning is that it uses digitally based communication tools and technology such as Social, Web, Mobile, Scannable Surface.[57][58] Nevertheless, both are aligned with the vision, the mission of the company and the overarching business strategy.[59]
Online reviews have become one of the most important components in purchasing decisions by consumers in North America. According to a survey conducted by Dimensional Research which included over 1000 participants, 90% of respondents said that positive online reviews influenced their buying decisions and 94% will use a business with at least four stars. Interestingly, negative reviews typically came from online review sites whereas Facebook was the main source of positive reviews. Forrester Research predicts that by 2020, 42% of in-store sales will be from customers who are influenced by web product research.

A generalization of PageRank for the case of ranking two interacting groups of objects was described in [32] In applications it may be necessary to model systems having objects of two kinds where a weighted relation is defined on object pairs. This leads to considering bipartite graphs. For such graphs two related positive or nonnegative irreducible matrices corresponding to vertex partition sets can be defined. One can compute rankings of objects in both groups as eigenvectors corresponding to the maximal positive eigenvalues of these matrices. Normed eigenvectors exist and are unique by the Perron or Perron-Frobenius theorem. Example: consumers and products. The relation weight is the product consumption rate.
Start Value (In this case) is the number of actual links to each “node”. Most people actually set this to 1 to start, but there are two great reasons for using link counts. First, it is a better approximation to start with than giving everything the same value, so the algorithm stabilizes in less iterations and it is so useful to check my spreadsheet in a second… so node A has one link in (from page C)
Keeping up with the latest trends is a must for any business, but ignoring technology trends in the digital world is the matter of staying in business. Unfortunately, those trends (while easy enough to find mentioned online) are rarely explained well. There seems to be this mistaken idea that anyone who has an interest or need in the practice will just magically get the jargon. As we all know, that is one superpower that doesn’t exist in the real world.
A breadcrumb is a row of internal links at the top or bottom of the page that allows visitors to quickly navigate back to a previous section or the root page. Many breadcrumbs have the most general page (usually the root page) as the first, leftmost link and list the more specific sections out to the right. We recommend using breadcrumb structured data markup28 when showing breadcrumbs.
Unfortunately, SEO is also a slow process. You can make “quick wins” in markets which are ill-established using SEO, but the truth is that the vast majority of useful keyphrases (including long-tail keyphrases) in competitive markets will already have been optimized for. It is likely to take a significant amount of time to get to a useful place in search results for these phrases. In some cases, it may take months or even years of concentrated effort to win the battle for highly competitive keyphrases.
After finding websites that have good metrics, you have to make sure the website is related to your site. For each competitor backlink, try to understand how your competitor got that link. If it was a guest article, send a request to become a contributor as well. If it was a product review by a blogger, contact the writer and offer them a good deal in exchange for a similar review.
The amount of link juice passed depends on two things: the number of PageRank points of the webpage housing the link, and the total number of links on the webpage that are passing PageRank. It’s worth noting here that while Google will give every website a public-facing PageRank score that is between 1 and 10, the “points” each page accumulates from the link juice passed by high-value inbound links can — and do — significantly surpass ten. For instance, webpages on the most powerful and significant websites can pass link juice points in the hundreds or thousands. To keep the rating system concise, Google uses a lot of math to correlate very large (and very small) PageRank values with a neat and clean 0 to 10 rating scale.

Regarding nofollow on content that you don’t want indexed, you’re absolutely right that nofollow doesn’t prevent that, e.g. if someone else links to that content. In the case of the site that excluded user forums, quite a few high-quality pages on the site happened not to have links from other sites. In the case of my feed, it doesn’t matter much either way, but I chose not to throw any extra PageRank onto my feed url. The services that want to fetch my feed url (e.g. Google Reader or Bloglines) know how to find it just fine.

Thanks a lot for all of those great tips you handed out here. I immediately went to work applying the strategies that you mentioned. I will keep you posted on my results. I have been offering free SEO services to all of my small business bookkeeping clients as a way of helping them to grow their businesses. Many of them just don’t have the resources required to hire an SEO guru to help them but they need SEO bad. I appreciate the fact that you share your knowledge and don’t try to make it seem like it’s nuclear science in order to pounce on the innocent. All the best to you my friend!

In my view there is nothing wrong with saying ‘hey Google, these pages are not important from a search engine perspective, let me not give them so much weight’. Regardless of how Google now views these type of pages from a weight perspective, doing the above as a webmaster should be logical and encouraged. You have said this yourself at least a few times in the past.
On October 17, 2002, SearchKing filed suit in the United States District Court, Western District of Oklahoma, against the search engine Google. SearchKing's claim was that Google's tactics to prevent spamdexing constituted a tortious interference with contractual relations. On May 27, 2003, the court granted Google's motion to dismiss the complaint because SearchKing "failed to state a claim upon which relief may be granted."[67][68]

Maintenance. Ongoing addition and modification of keywords and website con­tent are necessary to continually improve search engine rankings so growth doesn’t stall or decline from neglect. You also want to review your link strategy and ensure that your inbound and outbound links are relevant to your business. A blog can provide you the necessary structure and ease of content addition that you need. Your hosting company can typically help you with the setup/installation of a blog.


Two weeks ago I changed a few internal anchor text links for a HTML SELECT Label in order to save some space in the menu bar. Today, when I saw in Google the Cache (text-version) page of my site I realized that all the links in the HTML SELECT Label cannot be followed. So I understand that Googlebot doesn’t follow this links and obviously there’s no inbound ‘link juice’. Is that so?
Bob Dole (interesting name), you’re certainly welcome to use Bing if you prefer, but before you switch, you might check whether they do similar things. I know that Nate Buggia has strongly recommended not to bother with PageRank sculpting in the past, for example, or at least that was my perception from his comments at the last couple SMX Advanced conferences.
Matt, this is an excellent summary. I finally got around to reading “The Search” by John Battelle and it was very enlightening to understand much of the academia behind what led to the creation of Backrub.. er Google.Looking at how many times the project was almost shutdown due to bandwidth consumption (> 50% of what the university could offer at times) as well as webmasters being concerned that their pages would be stolen and recreated. It’s so interesting to see that issues we see today are some of the same ones that Larry and Sergey were dealing with back then. As always, thanks for the great read Matt!
Yes the links we have are found elsewhere but our focus is saving our users and clients time so we consolidated the links because it takes hours and hours and hours of searching to find them and some searchers are not very savvy when it comes to looking for, and finding, good quality information. I look at the links like a library, my library has these books, so do a bunch of other libraries. I think it is a shame that I have to hide my books from Google because I have to many really good ones because it is seen as a BAD thing in Google’s eyes. Darned if you dont create a good site, and darned if you do.
At IMI, we have the ability to incorporate Affiliate Marketing into any integrated digital marketing campaign in order to push product sales and drive revenue. We believe that this is just one piece of the digital puzzle however. Affiliate campaigns can incorporate many aspects of digital marketing including web design and development, display, paid search, conversion optimization, and SEO.
In my view, the Reasonable Surfer model would findamentally change the matrix values above, so that the same overall PageRank is apportioned out of each node, but each outbound link carres a different value. In this scenario, you can indeed make the case that three links will generate more traffic than one, although the placement of these links might increase OR DECREASE the amount of PageRank that is passed, since (ultimately) the outbound links from page A to Page B are dependent on the location of all other outbound links on Page A. But this is the subject of another presentation for the future I think.
A breadcrumb is a row of internal links at the top or bottom of the page that allows visitors to quickly navigate back to a previous section or the root page. Many breadcrumbs have the most general page (usually the root page) as the first, leftmost link and list the more specific sections out to the right. We recommend using breadcrumb structured data markup28 when showing breadcrumbs.

Search Engine Optimization

×