Case Study: One Sites Recovery from an Ugly SEO Mess: Posted by - TopicsExpress



          

Case Study: One Sites Recovery from an Ugly SEO Mess: Posted by AlanBleiweiss This past March, I was contacted by a prospective client: My site has been up since 2004. I had good traffic growth up to 2012 (doubling each year to around a million page views a month), then suffered a 40% drop in mid Feb 2012. Ive been working on everything that I can think of since, but the traffic has never recovered. Since my primary business is performing strategic site audits, this is something I hear often. Site appears to be doing quite well, then gets slammed. Site owner struggles for years to fix it, but repeatedly comes up empty. It can be devastating when that happens. And now more than ever, site owners need real solutions to serious problems. As this chart shows, when separating out the expected roller coaster effect, Google organic traffic took a nose-dive in early February of 2012. First step: check and correlate with known updates When this happens, the first thing I do is jump to Mozs Google Algorithm Change History charts to see if I can pinpoint a known Google update that correlates to a drop. Except in this case, there was no direct same day update listed. A week before, theres an entry listed that references integrating Panda into the main index more, however discussion around that is this change happened sometime in January. So maybe its Panda, maybe its not. Expand your timeline: look for other hits At this point, I expanded the timeline view to see if I could spot other specific drops and possibly associate them to known updates. I did this because some sites that get hit once, get hit again and again. Well now we have a complete mess. At this point, if youre up for the challenge, you can take the time to carefully review all the ups and downs manually, comparing drops to Mozs change history. Personally, when I see something this ugly, I prefer to use the Panguin Tool. It allows you to see this timeline with a known update overlay for various Google updates. Saves a lot of time. So thats what I did. Well okay this is an ugly mess as well. If you think you can pin enough of the drops on specific factors, thats great. What I like about the Panguin Tool is you can turn off or hide different update types to try and look for a consistent issue type. Alternately, you can zoom in to look at individual updates and see if they align with a specific, clear drop in traffic. Looking at this chart, its pretty clear the site saw a second dropoff beginning with Panda 3.3. The next dropoff aligned with updates appears to be Panda 3.4, however it was already in a slide after Panda 3.3 so we cant be certain of this one. Multiple other updates took place after that where there may or may not have been some impact, followed by further cascading downward. Then, in the midst of THAT dropoff, we see a Penguin update that also MAY or MAY NOT have played into the problem. The ambiguous reality This is a great time to bring up the fact that one of the biggest challenges we face in dealing with SEO is the ambiguous nature of what takes place. We cant always, with true certainty, know whether a site has been hit by any single algorithm change. In between all the known updates, Google is constantly making adjustments. The other factor here is that when a given update takes place, it doesnt always roll out instantly, nor is every site reprocessed against that latest change right away. The cascading impact effect Heres where evaluating things becomes even more of a mess. When an algorithm update takes place, it may be days or even weeks before a site sees the impact of that, if at all. And once it does, whatever changes to the overall status of a site come along due to any single algorithm shift, other algorithms are sometimes going to then base formulaic decisions on that new status of a site. So if a site becomes weaker due to a given algorithm change, even if the drop is minimal or non-observant, it can still suffer further losses due to that weakened state. I refer to this as the cascading impact effect. The right solution to cascading impact losses Okay so lets say youre dealing with a site that appears to have been hit by multiple algorithm updates. Maybe some of them are Panda, maybe others arent Panda. The only correct approach in this scenario is to step back and understand that for maximum sustainable improvement, you need to consider every aspect of SEO. Heck, even if a site was ONLY hit by Panda, or Penguin, or the Above the Fold algorithm, I always approach my audits with this mindset. Its the only way to ensure that a site becomes more resilient to future updates of any type. And when you approach it this way, because youre looking at the across-the-board considerations, youre much more likely to address the actual issues that you can associate with any single algorithm. The QUART mindset It was at this point where I began to do my work. A couple years ago, I coined the acronym QUART—what I call the five super-signals of SEO: * Quality * Uniqueness * Authority * Relevance * Trust With every single factor across the full spectrum of signals in SEO, I apply the QUART* test. Any single signal needs to score high in at least three of the five super-signals. Whether its a speed issue, a crawl efficiency issue, topical focus, supporting signals on-site or off-site, whatever it is, if that signal does not score well with quality, uniqueness or relevance, it leaves that page, that section of a site, or that site as a whole vulnerable to algorithmic hits. If you get those three strong enough, that signal will, over time, earn authority and trust score value as well. If you are really strong with relevance with any single page or section of the site, but weak in quality or uniqueness, you can still do well in SEO if the overall site is over-the-top with authority and trust. *When I first came up with this acronym, I had the sequence of letters as QURTA, since quality, uniqueness, and relevance are, in my opinion, the true ideal target above all else. New sites dont have authority or trust, yet they can be perfectly good, valuable sites if they hit those three. Except Jen Lopez suggested that if I shift the letters for the acronym, it would make it a much easier concept for people to remember. Thanks Jen! A frustratingly true example Lets say you have one single page of content. Its only okay or may even be dismal in regard to quality and uniqueness. Even if you do, and if the sites overall authority and trust are strong enough, you can outrank an entire site devoted to that specific topic. This happens all the time with sites like Wikipedia, or Yahoo Answers. Dont you hate that? Yeah, I know—Yahoo Answers? Trust? Ha! Sadly, some sites have, over time, built so much visibility, brand recognition, and trust for enough of their content, that they can seemingly get away with SEO murder. Its frustrating to see. Yet the foundational concept as to WHY that happens is understandable if you apply the QUART test. Spot mobile site issues One challenge this site has is that theres also a separate mobile subdomain. Looking at the Google traffic for that shows similar problems, beginning back in February of 2012. Note that for the most part, the mobile site suffered from that same major initial hit and subsequent downslide. The one big exception was a technical issue unique to the mobile site at the end of 2012 / beginning of 2013. Identify and address priority issues Understanding the QUART concept, and having been doing this work for years, I dove head-first into the audit. Page processing and crawl efficiency NOTE: This is an educational site – so all educational page labels refer to different primary pages on the site. For my audits, I rely upon Google Analytics Page Timings data, URIValet 1.5 mbps data, and also WebPageTest.org (testing from different server locations and at different speeds including DSL, Cable and Mobile). Speed improvement goals Whenever I present audit findings to a client, I explain Heres the ideal goal for this issue, yet I dont expect you to hit the ideal goal, only that you do your best to make improvements without becoming bogged down in this one issue. For this site, since not every single page had crisis speed problems, I was looking to have the site owner at least get to a point of better, more consistent stability. So while theres still room for vast improvement, the work performed went quite far in the right direction. Speed issues addressed: domain and process calls The first issue tackled was the fact that at the template level, the various pages on the site were calling several different processes across several different domains. A great resource to use for generating lists of what third party processes individual pages use that I rely upon is a report in the WebPageTest.org results. It lists every domain called for the page tested, along with total processes called from those, and gives separate data on the total file sizes across each. Reducing the number of times a page has to call a third party domain, and the number of times an individual process needs to be run is often a way to help speed up functionality. In the case of this site, several processes were eliminated: * Google APIs * Google Themes * Google User Content * Clicktale * Gstatic * RackCDN By eliminating functionality that was dependent upon third party servers meant less DNS lookups, and less dependance upon connections to other servers somewhere else on the web. Typical service drains can often come from ad blocks (serving too many ads from too many different ad networks is a frequent speed drain culprit), social sharing widgets, third party font generation, and countless other shiny object services. Clean code Yeah, I know—you dont have to have 100% validated code for SEO. Except what Ive found through years of this work, is that the more errors you have in your markup, the more likely there will be potential for processing delays, and beyond that, the more likely search algorithms will become confused. And even if you cant prove in a given site that cleaner code is a significant speed improvement point, its still a best practice, which is what I live for. So its always included in my audit evaluation process. Improve process efficiency Next up on the list was the range of issues all too many sites have these days regarding efficiency within a sites own content. Tools to help here include Google Webmaster Tools, Google Page Speed Insights, and again WebPageTest.org among others. Issues Im talking about here include above-the-fold render-blocking JavaScript and CSS, lack of browser caching, lack of compression of static content, server response times, a host of code-bloat considerations, too-big image sizes, and the list goes on... NOTE: This is an educational site – so all educational page labels refer to different primary pages on the site. Note: Google Page Speed Insights recommendations and WebPageTest.orgs grade reports only offer partial insight. What they do offer however, can help you go a long way to making speed improvements. Also, other speed reporting tools abound, to differing degrees of value, accuracy and help. The most important factor to me is to not rely on any single resource, and do your own extensive testing. Ultimately, enough effort in research and testing needs to be performed with followup checking to ensure you address the real issues on a big enough scale to make a difference. Just glossing over things or only hitting the most obvious problems is not always going to get you real long-term sustainable results... Correct crawl inefficiency Another common problem I find is where a site evolves over time, many of the URLs change. When this happens, site owners dont properly clean up their own internal links to those pages. The end result is a weakening of crawl efficiency and then user experience quality and trust signals. Remember that Google and Bing are, in fact, users of your site. Whether you want to admit it or not. So if theyre crawling the site and run into too many internal redirects (or heaven forbid redirect loops), or dead ends, thats going to make their systems wary to want to bother continuing the crawl. And abandoning the crawl because of that is not helpful by any stretch of the imagination. It also confuses algorithms. To that end, I like to crawl a sampling of a sites total internal links using Screaming Frog. That tool gives me many different insights, only one of which happens to be internal link problems. Yet its invaluable to know. And if I find enough of a percentage of that sample crawl URLs are redirecting or dead ends, that needs to get fixed. Note: for reference sake, the total number of pages on the entire site is less than 5,000. So thats a lot of internal inefficiency for that size site... Dont ignore external links While having link redirects and dead ends pointing to outside sources isnt ideal, its less harmful most of the time than internal redirects and dead ends. Except when its not. In this case, the site had previously been under a different domain name prior to a rebranding effort. And after the migration, it resulted in some ugly redirect loops involving the old domain! Topical focus evaluation At this point, the audit moved from the truly technical issues to the truly content related issues. Of course, since its algorithms that do the work to figure it all out, even content issues are technical in nature. Yet thats a completely different rant. So lets just move on to the list of issues identified that we can associate with content evaluation. Improve H1 and H2 headline tags Yeah, I know—some of you think these are irrelevant. Theyre really not. They are one more contributing factor when search engines look to multiple signals for understanding the unique topic of a given page. Noindex large volume of thin content pages Typical scenario here: a lot of pages that have very little to no actual unique content—at least not enough crawlable content to justify their earning high rankings for their unique focus. Be aware—this doesnt just include the content within the main content area of a page. If its surrounded (as was the case on this site) by blocks of content common to other pages, or if the main navigation or footer navigation are bloated with too many links (and surrounding words in the code), and if you offer too many shiny object widgets (as this site had), that unique content evaluation is going to become strained (as it did for this site). Add robust crawlable relevant content to video pages You can have the greatest videos on the planet on your site. And yet, if youre not a CNN, or some other truly well established high authority site, you are almost always going to need to add high quality, truly relevant content to pages that have those videos. So that was done here. And Im not just talking about filler content. In this case (as it always should be) the new content was well written and supportive of the content in the videos. Eliminate shiny object generic content that was causing duplicate content / topical dilution confusion across the site On pages that were worth salvaging but where there was thin content, I never recommend throwing those out. Instead, take the time to add more value content, yes. But also, consider eliminating some of those shiny objects. For this site, the reduction of those vastly improved the uniqueness of those pages. Improve hierarchical URL content funnels reducing the flat nature of content Flat architecture is an SEO myth. Want to know how I know this? I read it on the Internet, thats how! Oh wait. That was ME who said it. Seriously, though. If all your content looks like this: domain/category domain/subcategory domain/productdetailpage Thats flat architecture. It claims every one of these pages is as important as every other page on my site. And thats a fantasy. It also severely harms your need to communicate heres all the content specific to this category, or this sub-category. And THAT harms your need to say hey, this site is robust with content about this broad topic. So please. Stop with the flat architecture. And no, this is NOT just for search engines. Users who see proper URL funnels can rapidly get a cue as to where they are on the site (or as they look at that in the search results, more confidence about trust factors). So for this site, reorganization of content was called for and implemented. Add site-wide breadcrumb navigation Yes—breadcrumbs are helpful. because they reinforce topical focus, content organization, and improvements to user experience. So these were added. Noindex,nofollowed over 1,300 orphaned pages Pop-up windows. Theyre great for sharing additional information to site visitors. Except when you allow those to become indexable by search engines. Then all of asudden, youve got countless random pages that, on their own, have no meaning, no usability, and no way to communicate this is how this page relates to all these other pages over there. Theyre an SEO signal killer. So we lopped them out of the mix with a machete. Sometimes you may want to keep them indexable. If you do, they need full site navigation and branding, and proper URL hierarchical designations. So pay attention to whether its worth it to do that or not. Remove UX confusing widget functionality One particular widget on the site was confusing from a UX perspective. This particular issue had as much to do with site trust and overall usability as anything, and less to do with pure SEO. Except it caused some speed delays, needless site-jumping, repetition of effort and a serious weakening of brand trust. And those definitely impact SEO, so it was eliminated. Noindex internal site search results pages Duplicate content. Eliminated. nuff said? Eliminate multiple category assignments for blog articles More duplicate content issues. Sometimes you can keep these, however if multiple category assignments get out of hand, it really IS a duplicate content problem. So in this case, we resolved that. Unify brand identity across pages from old branding that had been migrated Old brand, new brand—both were intermingled after the site migration I previously described. Some of it was a direct SEO issue (old brand name in many page titles, in various on-site links and content) and some was purely a UX trust factor. Unify main navigation across pages from old branded version that had been migrated Again, this was a migration issue gone wrong. Half the site had consistent top navigation based on the new design, and half had imported the old main navigation. An ugly UX, crawl and topical understanding nightmare. Add missing meta descriptions Some of the bizarre auto-generated meta descriptions Google had been presenting on various searches was downright ugly. Killed that click-block dead by adding meta descriptions to over 1,000 pages. Remove extremely over-optimized meta keywords tag Not a problem you say? Ask Duane Forrester. Hell confirm—its one of many signal points they use to seek out potential over-optimization. So why risk leaving them there? About inbound links While I found some toxic inbound links in the profile, there werent many on this site. Most of those actually disappeared on their own thanks to all the other wars that continue to rage in the penguin arena. So for this site, no major effort has yet gone into cleaning up the small number that remain. Results Okay so what did all of this do in regard to the recovery I mention in the title? You tell me. And heres just a small fraction of the top phrase ranking changes: Next steps While the above charts show quite serious improvements since the implementation was started, theres more work that remains. Google Ad Scripts continue to be a big problem. Errors at the code level and processing delays abound. Its an ongoing issue many site owners struggle with. Heck—just eliminating Googles own ad server tracking code has given some of my clients as much as one to three seconds overall page processing improvement depending on the number of ad blocks as well as intermittent problems on Googles ad server network. Except at a certain point, ads are the life-blood of site owners. So thats a pain-point we may not be able to resolve. Other third party processes come with similar problems. Sometimes third party solution providers are helpful to want to improve their offerings, however the typical answer to your widget is killing my site is blah blah blah not our fault blah blah blah when I know for a fact from countless tests, that it is. So in this case, the client is doing what they can elsewhere for now. And ultimately, if need be, will abandon at least some of those third parties entirely if they can get a quality replacement. And content improvements—theres always more to do on that issue. Bottom line This is just one site, in one niche market. The work has been and continues to be extensive. It is, however, quite typical of many sites that suffer from a range of issues, not all of which can be pinned to Panda. Yet where ignoring issues you THINK might not be Panda specific is a dangerous game, especially now in 2015, where its only going to get uglier out there... So do the right thing for the site owner / your employer / your client / your affiliate site revenue... Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you dont have time to hunt down but want to read!
Posted on: Thu, 08 Jan 2015 04:03:43 +0000

Trending Topics



Recently Viewed Topics




© 2015