Google can’t like this. Although its great for them to have spammers out of the Wikipedias, they’re also losing a lot of very authorative input for their PR algorithm. Think about it – if every site in the world put nofollow on every link Google’s algorithm would be worthless overnight. There has been ongoing speculation as to whether or not Google ignores nofollows from certain sites like Wikipedia, something Mr Cutts has outrightly denied (but also admitted that it would be very useful to have more granular control over nofollow so that it was not an all-or-nothing situation.)
If you are serious about improving web traffic to your website, we recommend you read Google Webmasters and Webmaster Guidelines. These contain the best practices to help Google (and other search engines) find, crawl, and index your website. After you have read them, you MUST try our Search Engine Optimization Tools to help you with Keyword Research, Link Building, Technical Optimization, Usability, Social Media Strategy and more.
Black hat SEO attempts to improve rankings in ways that are disapproved of by the search engines, or involve deception. One black hat technique uses text that is hidden, either as text colored similar to the background, in an invisible div, or positioned off screen. Another method gives a different page depending on whether the page is being requested by a human visitor or a search engine, a technique known as cloaking. Another category sometimes used is grey hat SEO. This is in between black hat and white hat approaches, where the methods employed avoid the site being penalized, but do not act in producing the best content for users. Grey hat SEO is entirely focused on improving search engine rankings.
I won’t blame MC. Google, knows what they does. These are things that webmasters need not worry about. Well, it won’t make much difference as far as I think. I don’t use no follow tags specifically – I use WP for blogging purposes and it does rest of the things for me other than writing content which I do. I think it is the content and the external links that sites point to – which should be considered. I mean, if a computer blog owner posts a really fantastic computer article about something related to computer, and also puts some links to external pages (which are really useful for the readers), then that post, should be ranked high in gooogle – And I think google does this well – So, webmasters, just concentrate on yur website/blogs etc and leave rest of the things to Big G.
There are ten essential types of marketing that can be done online. Some of these can be broken down into organic marketing and others can be categorized as paid marketing. Organic, of course, is the allure of marketing professionals from around the planet. It's free and its unencumbered traffic that simply keeps coming. Paid marketing, on the other hand, is still a very attractive proposition as long as the marketing pays for itself by having the right type of offer that converts.
I think I did it by nature 🙂 I am just starting to learn all SEO and SERP stuffs. PageRank only entered my life when I quit the job last month. I don’t rush and I believe in a more natural way to get traffics, such as trying hard to create good contents and releasing some GPL works. You give something, and traffic flows in return. It’s Newton III Law. No sculpting, nor illegal way is necessary.
Backlinks can be time-consuming to earn. New sites or those expanding their keyword footprint may find it difficult to know where to start when it comes to link building. That's where competitive backlink research comes in: By examining the backlink profile (the collection of pages and domains linking to a website) to a competitor that's already ranking well for your target keywords, you can gain insight about the link building that may have helped them. A tool like Link Explorer can help uncover these links so you can and target those domains in your own link building campaigns.
And my vital question about Amazon affiliate links. I think many people also wonder about it as well. I have several blogs where I solely write unique content reviews about several Amazon products, nothing more. As you know, all these links are full of tags, affiliate IDs whatsoever (bad in SEO terms). Should I nofollow them all or leave as they are?
I first discovered Sharpe years ago online. His story was one of the most sincere and intriguing tales that any one individual could convey. It was real. It was heartfelt. It was passionate. And it was a story of rockbottom failure. It encompassed a journey that mentally, emotionally and spiritually crippled him in the early years of his life. As someone who left home at the age of 14, had a child at 16, became addicted to heroin at 20 and clean four long years later, the cards were definitely stacked up against him.
For most parts the sophistication in this system is simplified here. I still have trouble understanding the difference between letting link flow withing my pages without thinking about a loop. For example, page A, B and C link to each other from all angles therefore the link points should be shared. But in this loop formula, page B does not link to A. It just goes to C and loops. How does this affect navigation bars? As you know they are meant to link stay on top and link to all pages. I’m lost.
Peter made a very good point in all of this, and Michael Martinez did in a backhanded way as well. Talking about a concept related PageRank sounds cool. It doesn’t actually have to be useful or practical, and it usually isn’t; but as long as the impression of something productive is given off, then that can be all that matters in the eyes of those who lack sufficient knowledge.
Simple question – Lets say I have a blog/site with lot of outgoing links (avg 10 links per page). All the outgoing links (in the editorial content and user generated ones) are nofollowed, while all the internal links are “open”. I might have manually “opened up” some links in the editorial content because I’m so sure of their authority (ex:-google faq pages).
Link text is the visible text inside a link. This text tells users and Google something about the page you're linking to. Links on your page may be internal—pointing to other pages on your site—or external—leading to content on other sites. In either of these cases, the better your anchor text is, the easier it is for users to navigate and for Google to understand what the page you're linking to is about.
An SEO technique is considered white hat if it conforms to the search engines' guidelines and involves no deception. As the search engine guidelines are not written as a series of rules or commandments, this is an important distinction to note. White hat SEO is not just about following guidelines, but is about ensuring that the content a search engine indexes and subsequently ranks is the same content a user will see. White hat advice is generally summed up as creating content for users, not for search engines, and then making that content easily accessible to the online "spider" algorithms, rather than attempting to trick the algorithm from its intended purpose. White hat SEO is in many ways similar to web development that promotes accessibility, although the two are not identical.
I compare the latest Google search results to this: Mcdonalds is the most popular and is #1 in hamburgers… they dont taste that great but people still go there. BUT I bet you know a good burger joint down the road from Google that makes awesome burgers, 10X better than Mcdonalds, but “we” can not find that place because he does not have the resources or budget to market his burgers effectively.
Cause if I do that, If I write good content, whilst my 100+ competitors link build, article market, forum comment, social bookmark, release viral videos, buy links, I’ll end up the very bottom of the pile, great content or not and really I am just as well taking my chances pulling off every sneaky trick in the book to get my site top because, everyone does it anyway and if I don’t what do have to lose?”
This is so funny. Google stifled the notion of linking to “great content” the minute they let on to how important linking was to passing pagerank. In effect, the importance of links has indeed led to pagerank hoarding and link commoditization which in turn leads to all of the things google doesn’t like such as spammy links, link farms, link selling, link buying, etc. What you end up with is a system, much like our economic system, where the rich get richer and poor get poorer. Nobody has a problem linking to CNN, as if they really needed the links. On the flip side who wants make a dofollow link to a site that’s 2 days old, great content or not when you can provide your visitors a nofollow link which is just as valuable to them. The whole notion of benefiting from a quality outbound link is a joke, the outbound linker receives 0 benefit when you factor the outflow of pagerank.
3) Some people don’t believe things have changed. In fact, if things really changed substantially a year ago, you’d think a few of the advanced SEOs out there would have noticed this and talked about it. But nada. There are lots of reasons why the change could have happened and not been spotted. Sculpting might really have been a second or third order factor, as Matt calls it — not helping things as much as some have assumed. SEOs that spotted it might have stayed quiet. Or, it didn’t change — and still hasn’t changed — and sculpting does work even better than Google thought, so it wants to put out a message that it doesn’t, in hopes of putting the genie back in the bottle. That’s probably the major conspiracy theory out there.
I am not worried by this; I do agree with Danny Sullivan (Great comment Danny, best comment I have read in a long time). I will not be changing much on my site re: linking but it is interesting too see that Google took over a year to tell us regarding the change, but was really happy to tell us about rel=”nofollow” in the first place and advised us all to use it.
One thing is certain: interlinking sites doesn't help you from a search engine standpoint. The only reason you may want to interlink your sites in the first place might be to provide your visitors with extra resources to visit. In this case, it would probably be okay to provide visitors with a link to another of your websites, but try to keep many instances of linking to the same IP address to a bare minimum. One or two links on a page here and there probably won't hurt you.
Links still matter as part of the algorithmic secret sauce. The influence of a site’s link profile is plain to see in its search engine rankings, whether for better or worse, and changes in that link profile cause noticeable movement up or down the SERP. An SEO’s emphasis today should be on attracting links to quality content naturally, not building them en masse. (For more on proper link building today, see http://bit.ly/1XIm3vf )
Place strategic search phrases on pages. Integrate selected keywords into your website source code and existing content on designated pages. Make sure to apply a suggested guideline of one to three keywords/phrases per content page and add more pages to complete the list. Ensure that related words are used as a natural inclusion of your keywords. It helps the search engines quickly determine what the page is about. A natural approach to this works best. In the past, 100 to 300 words on a page was recommended. Many tests show that pages with 800 to 2,000 words can outperform shorter ones. In the end, the users, the marketplace, content and links will determine the popularity and ranking numbers.
Great article and writing in general. My company just published a 5,000 word Keyword targeting best practices guide for PPC and SEO, and we linked to your article “10 Reasons You Should Use Google Trends for More Than Just Keyword Research”. http://vabulous.com/keyword-research-targeting-for-ppc-and-seo-guide/ I would love if you checked it out and possibly shared it if you like it.
As for the use of nofollow as a way to keep pages that shouldn’t be indexed out of Google (as with your feed example) is terrible advice. Your use of it on your feed link does nothing. If anyone links to your feed without nofollow, then it’s going to get indexed. Things that shouldn’t be indexed need to use either robots.txt or meta robots blocking. Nofollow on links to those items isn’t a solution.
Our team is made up of industry-recognized thought leaders, social media masters, corporate communications experts, vertical marketing specialists, and internet marketing strategists. Members of the TheeTeam host SEO MeetUp groups and actively participate in Triangle area marketing organizations. TheeDigital is an active sponsor of the AMA Triangle Chapter.
Steve, sometimes good information to users is a consolidation of very high quality links. We have over 3000 links to small business sites within the SBA as well as links to the Harvard and Yale library, academic journals, etc. But because we have the understanding that there should be no more than a hundred links in a website (more now from what Matt said) we have used nofollow on all of them out of fear that Google will penalize our site because of the amount of links.
I work on a site that allows users to find what they are looking for by clicking links that take them deeper and deeper into the site hierarchy. Content can be categorised in lots of different ways. After about three steps the difference between the results pages shown is of significance to a user but not to a search engine. I was about to add nofollow to links that took the browser deeper than 3 levels but after this announcement I won’t be…
Now that you know that backlinks are important, how do you acquire links to your site? Link building is still critical to the success of any SEO campaign when it comes to ranking organically. Backlinks today are much different than when they were built in 7-8 years back. Simply having thousands of backlinks or only have link from one website isn’t going to affect your rank position. There are also many ways to manage and understand your backlink profile. Majestic, Buzzstream, and Moz offer tools to help you manage and optimize your link profile. seoClarity offers an integration with Majestic, the largest link index database, that integrates link profile management into your entire SEO lifecycle.
Using influencers to market your products or services is a great way to quickly saturate yourself into the marketplace, no matter what you're peddling. However, finding the right influencer at the right price is the hard part. You don't necessarily have to go to the top-tier influencers; you can also opt for micro-influencers (those that have 10,000 to 100,000 followers or fans).
Also hadn’t thought about decreasing the rank value based on the spammyness of sites a page is linking into. My guess on how to do it would be determining the spammyness of individual pages based on multiple page and site factors, then some type of reverse pagerank calcuation starting with the those bad scores, then overlaying that on top of the “good” pagerank calculation as a penalty. This is another thing which would be interesting to play around with in the Nutch algorithm.
Journalists and writers are always on the lookout for experts to contribute quotes for their articles. Some (but not all) will include backlinks to their sources’ websites. Getting quotes in media outlets is a great way to not only get backlinks, but also build credibility within your industry. Even in instances where you don't get backlinks, this profile page for PMM's CEO Josh Rubin is a good example of how you can showcase your media appearances - something which both Google and your clients value when it comes to evaluating your authority.
There are numerous repositories to source affiliate products and services from. However, some of the biggest are sites like Clickbank, Commission Junction, LinkShare and JVZoo. You'll need to go through an application process, for the most part, to get approved to sell certain products, services or digital information products. Once approved, be prepared to hustle.
Start Value (In this case) is the number of actual links to each “node”. Most people actually set this to 1 to start, but there are two great reasons for using link counts. First, it is a better approximation to start with than giving everything the same value, so the algorithm stabilizes in less iterations and it is so useful to check my spreadsheet in a second… so node A has one link in (from page C)
That type of earth-shattering failure and pain really does a number on a person. Getting clean and overcoming those demons isn't as simple as people make it out to be. You need to have some serious deep-down reasons on why you must succeed at all costs. You have to be able to extricate yourself from the shackles of bad habits that have consumed you during your entire life. And that's precisely what Sharpe did.
It helps to improve your ranking for certain keywords. If we want this article to rank for the term ’SEO basics’ then we can begin linking to it from other posts using variations of similar anchor text. This tells Google that this post is relevant to people searching for ‘SEO basics’. Some experts recommend varying your anchor text pointing to the same page as Google may see multiple identical uses as ‘suspicious’.
What are backlinks doing for your SEO strategy? Well, Google considers over 200 SEO ranking factors when calculating where a page should rank, but we know that backlinks are one of the top three (the other two are content and RankBrain, Google’s AI). So while you should always focus on creating high-quality content, link-building is also an important factor in ranking your pages well on Google.
Ah – well the Reasonable Surfer is a different patent (and therefore a different algorithm) to PageRank. I would imagine that initially, only the first link counted – simply because there either IS or IS NOT a relationship between the two nodes. This mean it was a binary choice. However, at Majestic we certainly think about two links between page A and Page B with separate anchor texts… in this case in a binary choice, either the data on the second link would need to be dropped or, the number of backlinks can start to get bloated. I wrote about this on Moz way back in 2011!
For example, what are the quality and quantity of the links that have been created over time? Are they natural and organic links stemming from relevant and high quality content, or are they spammy links, unnatural links or coming from bad link neighborhoods? Are all the links coming from the same few websites over time or is there a healthy amount of global IP diversification in the links?
In order to be a data driven agency, we foster a culture of inspired marketing entrepreneurs that collaborate, innovate, and are constantly pushing the threshold of marketing intelligence. Our analytics team is well versed in mathematics, business analytics, multi-channel attribution modeling, creating custom analytics reporting dashboards, and performing detailed analysis and reporting for each client.
Regarding nofollow on content that you don’t want indexed, you’re absolutely right that nofollow doesn’t prevent that, e.g. if someone else links to that content. In the case of the site that excluded user forums, quite a few high-quality pages on the site happened not to have links from other sites. In the case of my feed, it doesn’t matter much either way, but I chose not to throw any extra PageRank onto my feed url. The services that want to fetch my feed url (e.g. Google Reader or Bloglines) know how to find it just fine.
There are also many keyword research tools (some free and some paid) that claim to take the effort out of this process. A popular tool for first timers is Traffic Travis, which can also analyse your competitors’ sites for their keyword optimization strategies and, as a bonus, it can deliver detailed analysis on their back-linking strategy, too. You can also use Moz.com’s incredibly useful keyword research tools – they’re the industry leader, but they come at a somewhat higher price.
Hi Matt, I have a question about PR: N/A. With the recent update I found many sites including mine went from PR: 3 to PR: N/A. I Googled for Site:mydomain.com to find it its banned, but I found its not banned, I posted this question on Google Webmaster forum and couple of other places but I didn’t get any help to fix it. I don’t know whom to ask, or how to figure this out. Could you please help me out please?
Numerous academic papers concerning PageRank have been published since Page and Brin's original paper. In practice, the PageRank concept may be vulnerable to manipulation. Research has been conducted into identifying falsely influenced PageRank rankings. The goal is to find an effective means of ignoring links from documents with falsely influenced PageRank.
Although online marketing creates many opportunities for businesses to grow their presence via the Internet and build their audiences, there are also inherent challenges with these methods of marketing. First, the marketing can become impersonal, due to the virtual nature of message and content delivery to a desired audience. Marketers must inform their strategy for online marketing with a strong understanding of their customer’s needs and preferences. Techniques like surveys, user testing, and in-person conversations can be used for this purpose.
Mathematical PageRanks for a simple network, expressed as percentages. (Google uses a logarithmic scale.) Page C has a higher PageRank than Page E, even though there are fewer links to C; the one link to C comes from an important page and hence is of high value. If web surfers who start on a random page have an 85% likelihood of choosing a random link from the page they are currently visiting, and a 15% likelihood of jumping to a page chosen at random from the entire web, they will reach Page E 8.1% of the time. (The 15% likelihood of jumping to an arbitrary page corresponds to a damping factor of 85%.) Without damping, all web surfers would eventually end up on Pages A, B, or C, and all other pages would have PageRank zero. In the presence of damping, Page A effectively links to all pages in the web, even though it has no outgoing links of its own.
PageRank gets its name from Google cofounder Larry Page. You can read the original ranking system to calculate PageRank here, if you want. Check out the original paper about how Google worked here, while you’re at it. But for dissecting how Google works today, these documents from 1998 and 2000 won’t help you much. Still, they’ve been pored over, analyzed and unfortunately sometimes spouted as the gospel of how Google operates now.
We help clients increase their organic search traffic by using the latest best practices and most ethical and fully-integrated search engine optimization (SEO) techniques. Since 1999, we've partnered with many brands and executed campaigns for over 1,000 websites, helping them dominate in even highly competitive industries, via capturing placements that maximize impressions and traffic.
Backlinks are important for a number of reasons. The quality and quantity of pages backlinking to your website are some of the criteria used by search engines like Google to determine your ranking on their search engine results pages (SERP). The higher you rank on a SERP, the better for your business as people tend to click on the first few search results Google, Bing or other search engines return for them.
When calculating PageRank, pages with no outbound links are assumed to link out to all other pages in the collection. Their PageRank scores are therefore divided evenly among all other pages. In other words, to be fair with pages that are not sinks, these random transitions are added to all nodes in the Web. This residual probability, d, is usually set to 0.85, estimated from the frequency that an average surfer uses his or her browser's bookmark feature. So, the equation is as follows:
Hey – I love this article. One thing I’ve done with a little bit of success is interview “experts” in whatever niche. In my case this is a mattress site and I sent questions to small business owners with the information I was looking for. Some were happy to help and I would send them a link to the article once it was live. I didn’t ask for a link, but in some cases they would feature the link on their own website.
SEO is also about making your search engine result relevant to the user's search query so more people click the result when it is shown in search. In this process, snippets of text and meta data are optimized to ensure your snippet of information is appealing in the context of the search query to obtain a high CTR (click through rate) from search results.
Having a ‘keyword rich’ domain name may lead to closer scrutiny from Google. According to Moz, Google has “de-prioritized sites with keyword-rich domains that aren’t otherwise high-quality. Having a keyword in your domain can still be beneficial, but it can also lead to closer scrutiny and a possible negative ranking effect from search engines—so tread carefully.”
A decent article which encourages discussion and healthy debate. Reading some of the comments I see it also highlights some of the misunderstandings some people (including some SEOs) have of Google PageRank. Toolbar PageRank is not the same thing as PageRank. The little green bar (Toolbar PageRank) was never a very accurate metric and told you very little about the value of any particular web page. It may have been officially killed off earlier this year, but the truth is its been dead for many years. Real PageRank on the other hand, is at the core of Google’s algorithm and remains very important.