Matt, you don’t mention the use of disallow pages via robots.txt. I’ve read that PageRank can be better utilised by disallowing pages that probably don’t add value to users searching on engines. For example, Privacy Policy and Terms of Use pages. These often appear in the footer of a website and are required by EU law on every page of the site. Will it boost the other pages of the site if these pages are added to robots.txt like so?

nofollow is beyond a joke now. There is so much confusion (especially when other engines’ treatment is factored in), I don’t know how you expect a regular publisher to keep up. The expectation seems to have shifted from “Do it for humans and all else will follow” to “Hang on our every word, do what we say, if we change our minds then change everything” and nofollow lead the way. I could give other examples of this attitude (e.g. “We don’t follow JavaScript links so it’s ‘safe’ to use those for paid links”), but nofollow is surely the worst.
Assume a small universe of four web pages: A, B, C and D. Links from a page to itself, or multiple outbound links from one single page to another single page, are ignored. PageRank is initialized to the same value for all pages. In the original form of PageRank, the sum of PageRank over all pages was the total number of pages on the web at that time, so each page in this example would have an initial value of 1. However, later versions of PageRank, and the remainder of this section, assume a probability distribution between 0 and 1. Hence the initial value for each page in this example is 0.25.
Advertiser Disclosure: Some of the products that appear on this site are from companies from which QuinStreet receives compensation. This compensation may impact how and where products appear on this site including, for example, the order in which they appear. QuinStreet does not include all companies or all types of products available in the marketplace.
Internet Marketing Inc. is one of the fastest growing full service Internet marketing agencies in the country with offices in San Diego, and Las Vegas. We specialize in providing results driven integrated online marketing solutions for medium-sized and enterprise brands across the globe. Companies come to us because our team of well-respected industry experts has the talent and creativity to provide your business with a more sophisticated data-driven approach to digital marketing strategy. IMI works with some clients through IMI Ventures, and their first product is VitaCup.
Pay per click (PPC) advertising, commonly referred to as Search Engine Marketing, delivers targeted traffic and conversions and will yield results faster than organic search engine optimization. Successful PPC marketing programs offer incredible revenue and brand-building opportunities. However, without a thorough understanding of how PPC works, it is very easy to mismanage valuable advertising budgets. That’s where we come in!
An essential part of any Internet marketing campaign is the analysis of data gathered from not just the campaign as a whole, but each piece of it as well. An analyst can chart how many people have visited the product website since its launch, how people are interacting with the campaign's social networking pages, and whether sales have been affected by the campaign (See also Marketing Data Analyst). This information will not only indicate whether the marketing campaign is working, but it is also valuable data to determine what to keep and what to avoid in the next campaign.
Suppose instead that page B had a link to pages C and A, page C had a link to page A, and page D had links to all three pages. Thus, upon the first iteration, page B would transfer half of its existing value, or 0.125, to page A and the other half, or 0.125, to page C. Page C would transfer all of its existing value, 0.25, to the only page it links to, A. Since D had three outbound links, it would transfer one third of its existing value, or approximately 0.083, to A. At the completion of this iteration, page A will have a PageRank of approximately 0.458.
Black hat SEO attempts to improve rankings in ways that are disapproved of by the search engines, or involve deception. One black hat technique uses text that is hidden, either as text colored similar to the background, in an invisible div, or positioned off screen. Another method gives a different page depending on whether the page is being requested by a human visitor or a search engine, a technique known as cloaking. Another category sometimes used is grey hat SEO. This is in between black hat and white hat approaches, where the methods employed avoid the site being penalized, but do not act in producing the best content for users. Grey hat SEO is entirely focused on improving search engine rankings.
Brian Dean, an SEO expert and the creator of BackLinko, uses SEO tactics to rank #1 on YouTube for keywords like “on page SEO” and “video SEO”. Initially, Dean admits his YouTube account struggled to get any views. Employing SEO methods like keyword optimization has enabled Dean to rise to #1 on YouTube for search results related to his business. He published his full strategy on Backlinko.
Description meta tags are important because Google might use them as snippets for your pages. Note that we say "might" because Google may choose to use a relevant section of your page's visible text if it does a good job of matching up with a user's query. Adding description meta tags to each of your pages is always a good practice in case Google cannot find a good selection of text to use in the snippet. The Webmaster Central Blog has informative posts on improving snippets with better description meta tags18 and better snippets for your users19. We also have a handy Help Center article on how to create good titles and snippets20.

Social media marketing on its own is free to use, as is content marketing if you’re creating and managing the content yourself. Several email marketing platforms have free plans for your first few hundred or thousand subscribers, giving you time to get your revenue increasing before you have to start paying. There are even free keyword research tools like Google’s Keyword Planner that can help you optimize your site at no cost.
Companies often use email marketing to re-engage past customers, but a “Where’d You Go? Want To Buy This?” message can come across as aggressive, and you want to be careful with your wording to cultivate a long-term email subscriber. This is why JetBlue’s one year re-engagement email works so well -- it uses humor to convey a sense of friendliness and fun, while simultaneously reminding an old email subscriber they might want to check out some of JetBlue’s new flight deals.
For the purpose of their second paper, Brin, Page, and their coauthors took PageRank for a spin by incorporating it into an experimental search engine, and then compared its performance to AltaVista, one of the most popular search engines on the Web at that time. Their paper included a screenshot comparing the two engines’ results for the word “university.”
Ask for a technical and search audit for your site to learn what they think needs to be done, why, and what the expected outcome should be. You'll probably have to pay for this. You will probably have to give them read-only access to your site on Search Console. (At this stage, don't grant them write access.) Your prospective SEO should be able to give you realistic estimates of improvement, and an estimate of the work involved. If they guarantee you that their changes will give you first place in search results, find someone else.
Influencer marketing: Important nodes are identified within related communities, known as influencers. This is becoming an important concept in digital targeting. It is possible to reach influencers via paid advertising, such as Facebook Advertising or Google Adwords campaigns, or through sophisticated sCRM (social customer relationship management) software, such as SAP C4C, Microsoft Dynamics, Sage CRM and Salesforce CRM. Many universities now focus, at Masters level, on engagement strategies for influencers.
Hi Brian thank you for sharing this awesome backlinking techniques. My site is currently not ranking well. It used to be, sometime mid last year, but it suddenly got de-ranked. Not really sure why. I haven’t been participating in any blackhat techniques or anything at all. I’ll try a few of your tips and hopefully it will help my site back to its shape.
According to the U.S. Commerce Department, consumers spent $453.46 billion on the web for retail purchases in 2017, a 16.0% increase compared with $390.99 billion in 2016. That’s the highest growth rate since 2011, when online sales grew 17.5% over 2010. Forrester predicts that online sales will account for 17% of all US retail sales by 2022. And digital advertising is also growing strongly; According to Strategy Analytics, in 2017 digital advertising was up 12%, accounting for approximately 38% of overall spending on advertising, or $207.44 billion.
If you build a new site and only used Domain Authority to create links, you could EASILY have got linked from the worst page possible, even if it was from the best domain, because of the INTERNAL LINKS of the other web pages! How on earth are you going to be able to see the strength of a link if that strength depends on the internal links on an entirely different website?!
I liken this to a paradoxical Catch-22 scenario, because it seems like without one you can't have the other. It takes money to drive traffic, but it takes traffic to make money. So don't make the mistake that millions of other online marketers make around the world. Before you attempt to scale or send any semblance of traffic to your offers, be sure to split-test things to oblivion and determine your conversion rates before diving in headfirst.
Hi Matt, I have a question about PR: N/A. With the recent update I found many sites including mine went from PR: 3 to PR: N/A. I Googled for Site:mydomain.com to find it its banned, but I found its not banned, I posted this question on Google Webmaster forum and couple of other places but I didn’t get any help to fix it. I don’t know whom to ask, or how to figure this out. Could you please help me out please?
As mentioned earlier, technology and the internet allows for 24 hours a day, 7 days a week service for customers as well as enabling them to shop online at any hour of that day or night, not just when the shops are over and across the whole world. This is a huge advantage for retailers to use it and direct customers from the store to its online store. It has also opened up an opportunity for companies to only be online based rather than having an outlet or store due to the popularity and capabilities of digital marketing.
SEO experts have a really bad habit: They like to throw around strange words and industry jargon when they talk to customers without checking to make sure that their clients understand the topic at hand. Some do this intentionally to paper over the fact that they use black hat techniques that will ultimately hurt their customers. But for most, it’s simply a matter of failing to recognize that part of their job is to educate their clients.
Writing blog posts is especially effective for providing different opportunities to land on page one of search engines -- for instance, maybe your eyeglass store’s website is on page three of Google for “eyeglasses,” but your “Best Sunglasses of 2018” blog post is on page one, pulling in an impressive amount of traffic (over time, that blog post could also boost your overall website to page one).
Quite simply, a backlink is one website mentioning another website and linking to it. It is not merely referencing the website or it’s web address. It has to be a clickable link using an href attribute within the code. It is the difference between http://www.moz.com and Moz. Even though the first example displays a URL, the search engines do not register this as a backlink, whereas the word that has a link (often underlined and in a different color), is.

Search engines use complex mathematical algorithms to guess which websites a user seeks. In this diagram, if each bubble represents a website, programs sometimes called spiders examine which sites link to which other sites, with arrows representing these links. Websites getting more inbound links, or stronger links, are presumed to be more important and what the user is searching for. In this example, since website B is the recipient of numerous inbound links, it ranks more highly in a web search. And the links "carry through", such that website C, even though it only has one inbound link, has an inbound link from a highly popular site (B) while site E does not. Note: Percentages are rounded.
Two weeks ago I changed a few internal anchor text links for a HTML SELECT Label in order to save some space in the menu bar. Today, when I saw in Google the Cache (text-version) page of my site I realized that all the links in the HTML SELECT Label cannot be followed. So I understand that Googlebot doesn’t follow this links and obviously there’s no inbound ‘link juice’. Is that so?
Page and Brin's theory is that the most important pages on the Internet are the pages with the most links leading to them. PageRank thinks of links as votes, where a page linking to another page is casting a vote. The idea comes from academia, where citation counts are used to find the importance of researchers and research. The more often a particular paper is cited by other papers, the more important that paper is deemed. 
In 2005, in a pilot study in Pakistan, Structural Deep Democracy, SD2[61][62] was used for leadership selection in a sustainable agriculture group called Contact Youth. SD2 uses PageRank for the processing of the transitive proxy votes, with the additional constraints of mandating at least two initial proxies per voter, and all voters are proxy candidates. More complex variants can be built on top of SD2, such as adding specialist proxies and direct votes for specific issues, but SD2 as the underlying umbrella system, mandates that generalist proxies should always be used.
If you are using Responsive Web Design, use meta name="viewport" tag to tell the browser how to adjust the content. If you use Dynamic Serving, use the Vary HTTP header to signal your changes depending on the user-agent. If you are using separate URLs, signal the relationship between two URLs by tag with rel="canonical" and rel="alternate" elements.
This is what happens to the numbers after 15 iterations…. Look at how the 5 nodes are all stabilizing to the same numbers. If we had started with all pages being 1, by the way, which is what most people tell you to do, this would have taken many more iterations to get to a stable set of numbers (and in fact – in this model – would not have stabilized at all)
PageRank is a link analysis algorithm and it assigns a numerical weighting to each element of a hyperlinked set of documents, such as the World Wide Web, with the purpose of "measuring" its relative importance within the set. The algorithm may be applied to any collection of entities with reciprocal quotations and references. The numerical weight that it assigns to any given element E is referred to as the PageRank of E and denoted by {\displaystyle PR(E).} Other factors like Author Rank can contribute to the importance of an entity.

Search engines are smart, but they still need help. The major engines are always working to improve their technology to crawl the web more deeply and return better results to users. However, there is a limit to how search engines can operate. Whereas the right SEO can net you thousands of visitors and increased attention, the wrong moves can hide or bury your site deep in the search results where visibility is minimal.


4. The facets of content marketing. Though content marketing can be treated as a distinct strategy, I see it as a necessary element of the SEO process. Only by developing high-quality content over time will you be able to optimize for your target keywords, build your site’s authority, and curate a loyal recurring audience. You should know the basics, at the very least, before proceeding with other components of SEO.
Consumers today are driven by the experience. This shift from selling products to selling an experience requires a connection with customers on a deeper level, at every digital touch point. TheeDigital’s internet marketing professionals work to enhance the customer experience, grow your online presence, generate high-quality leads, and solve your business-level challenges through innovative, creative, and tactful internet marketing.
I have not at all seen the results I would expect in terms of page rank throughout my site. I have almost everything pointing at my home page, with a variety of anchor text, but my rank is 1. There is a page on my site with 3, though, and a couple with 2, so it certainly is not all about links; I do try to have somewhat unique and interesting content, but some of my strong pages are default page content. I will explore the help forum. (I guess these comments are nofollow :P) I would not mind a piece of this page rank …

Also hadn’t thought about decreasing the rank value based on the spammyness of sites a page is linking into. My guess on how to do it would be determining the spammyness of individual pages based on multiple page and site factors, then some type of reverse pagerank calcuation starting with the those bad scores, then overlaying that on top of the “good” pagerank calculation as a penalty. This is another thing which would be interesting to play around with in the Nutch algorithm.


Most online marketers mistakenly attribute 100% of a sale or lead to the Last Clicked source. The main reason for this is that analytic solutions only provide last click analysis. 93% to 95% of marketing touch points are ignored when you only attribute success to the last click. That is why multi-attribution is required to properly source sales or leads.
If you’re Matt Cutts and a billion people link to you because you’re the Spam guy at Google, writing great content is enough. For the rest of us in hypercompetitive markets, good content alone is not enough. There was nothing wrong with sculpting page rank to pages on your site that make you money as a means of boosting traffic to those pages. It’s not manipulating Google, there’s more than enough of that going on in the first page of results for most competitive keywords. Geez Matt, give the little guy a break!
If the algorithm really works as Matt suggests, no one should use nofollow links internally. I’ll use the example that Matt gave. Suppose you have a home page with ten PR “points.” You have links to five “searchable” pages that people would like to find (and you’d like to get found!), and links to five dull pages with disclaimers, warranty info, log-in information, etc. But, typically, all of the pages will have links in headers and footers back to the home page and other “searchable” pages. So, by using “nofollow” you lose some of the reflected PR points that you’d get if you didn’t use “nofollow.” I understand that there’s a decay factor, but it still seems that you could be leaking points internally by using “nofollow.”

This will give you an indication of how many times a search is performed in a month (low numbers are not very useful unless there is a very clear buying signal in the keyphrase – working hard for five hits a month is not recommended in most cases) and how much the phrase is “worth” per click to advertisers (e.g., how much someone will pay to use that keyphrase). The more it’s worth, the more likely it is that the phrase is delivering business results for someone.

If Google was to allow webmasters full control of their own fate it would be like giving up the farm rather than giving up to the forces of human creativity. If you feel today were in a crowded market place even with a Google’s superiority complex, wait until the web is completely machine readable and aggregated on pure laws of information. I don’t think most can comprehend the future of data management as we have yet to see readily available parsing mechanisms that evolve purely based on the principles of information theory and not merely economies of scale. Remember not too long ago when Facebook tried to change their TOS to own your links and profiles? We can see that the tragedy of the commons still shapes the decision of production with that of opting in.


The world is mobile today. Most people are searching on Google using a mobile device. The desktop version of a site might be difficult to view and use on a mobile device. As a result, having a mobile ready site is critical to your online presence. In fact, starting in late 2016, Google has begun experiments to primarily use the mobile version of a site's content41 for ranking, parsing structured data, and generating snippets.

Some search engines have also reached out to the SEO industry, and are frequent sponsors and guests at SEO conferences, webchats, and seminars. Major search engines provide information and guidelines to help with website optimization.[18][19] Google has a Sitemaps program to help webmasters learn if Google is having any problems indexing their website and also provides data on Google traffic to the website.[20] Bing Webmaster Tools provides a way for webmasters to submit a sitemap and web feeds, allows users to determine the "crawl rate", and track the web pages index status.
I think it is important you distinguish your advice about no-following INTERNAL links and no-following EXTERNAL links for user-generated content. Most popular UGC-heavy sites have no-followed links as they can’t possibly police them editorially & want to give some indication to the search engines that the links haven’t been editorially approved, but still might provide some user benefit.
Digital marketing is probably the fastest-changing marketing field out there: New tools are being built, more platforms emerge and more channels need to be included into your marketing plan. How not to get overwhelmed while staying on top of the latest marketing trends? Here are a few tools that help you scale and automate some parts of your marketing routine making you a more productive and empowered marketer: Tools to Semi-Automate Marketing Tasks 1.
I first discovered Sharpe years ago online. His story was one of the most sincere and intriguing tales that any one individual could convey. It was real. It was heartfelt. It was passionate. And it was a story of rockbottom failure. It encompassed a journey that mentally, emotionally and spiritually crippled him in the early years of his life. As someone who left home at the age of 14, had a child at 16, became addicted to heroin at 20 and clean four long years later, the cards were definitely stacked up against him.
Here’s my take on the whole pagerank sculpting situation. As I understand it, the basic idea is that you can increase your rankings in Google by channeling the page rank of your pages to the pages you want ranked. This used be done with the use of the ‘no folow’ tag. That said, things have changed, and Google has come out and said that the way ‘no follow’ use to work has changed. In short, using ‘no follow’ to channel that page rank juice is no longer as effective as it once was.
While ordinary users were not that interested in pages' scores, SEOs of a different caliber felt that this was a great opportunity to make a difference for their customers. This obsession of SEOs with PageRank made everyone feel that this ranking signal is more or less the only important one. In spite of the fact that pages with a lower PR score can beat those with a higher score! What did we receive then, as a result?

Page Rank

×