Released on April 24th 2012, Penguin 2.0 has so far continued the work of the original Google Penguin algorithm in attempting to tackle web spam and poor linking practices. When Penguin was first launched last year, Google framed it as a way to punish sites that were engaged in ‘black hat’ SEO practices, from buying up links through web directories, to stuffing pages and meta tags with keywords. Penguin 2.0 aims to build on this approach, and aligns with Google Panda in making it harder for sites to inflate their organic visibility without having quality content and organic linking from high quality sites to back up their position. What, then, do we know about Penguin 2.0 so far, and what can sites do to improve their ranking if they’ve been punished by the algorithm change? What Penguin 2.0 Does It’s perhaps best to start by looking to Google’s Matt Cutts, the chief engineer behind Penguin, to explain what some of the major features of version 2.0 are. As with the original Penguin, 2.0 is all about tackling web spam, but goes into more detail over what counts as links from low quality sites. Google have also updated their Panda algorithm, which is now in a state of ‘everflux’ so gaming Google is now harder than ever. Cutts suggests that Penguin 2.0 will affect about 2.3 per cent of English and US search queries, and will also be rolled out to include other languages. The algorithm will work to promote good SEO practices from webmasters, which means generating relevant, well written content on sites, and building up a search profile based on links from high quality sites, rather than from pages that have been identified as participating in […]
In an attempt to reduce the number of hacked or compromised sites, Google has released an online tutorial series dubbed “Webmasters help for hacked sites”. There are 8 different steps in the recovery process, each of which is targeted towards a different level of user. If you are unfamiliar with any of the steps outlined in the guide, then Google recommends contacting your hosting provider or support team to let them handle the issue. Even the most tech savvy webmasters occasionally fall prey to vulnerabilities in popular scripts. This is particularly true of CMS platforms like WordPress or Joomla, with one of more memorable issues in recent history being the heavily exploited vulnerability within TimThumb. If you’ve recently seen instances within Google’s SERPS of any of the following messages for your website: This site may be compromised This site may harm your computer Reported attack page! Phishing (web forgery) suspected Notice of suspected hacking Then your website may be compromised. Head on over to http://www.google.com/webmasters/hacked/ to learn more about the removal process or start by watching the video below.
For those that feel that they may have been hit by a penalty, Matt Cutts has released a video explaining the various responses that you can expect to receive. In the wake of the numerous Panda/Penguin updates, Google also released the disavow links tool (16th October 2012). As a result, the response time for a recon-request may have increased while webmasters purge their low-quality backlinks and disavow links they were unable to move. (more…)
Over the past few weeks Google have released an onslaught of new updates dedicated to reducing spam and increasing the quality of their organic results. As if releasing the EMD and Panda updates back to back wasn’t bad enough, Google decided to release their latest data refresh of Penguin yesterday (5th October 2012). Today I’m going to be covering what these new updates mean for you and discussing the close proximity of all three updates dubbed “The Trilogy of Search Terror” by Search Engine Watch. The Timeline Google Panda “20″ – 27th September 2012 The EMD Update – 28th September 2012 Google Penguin 3 – 5th October 2012 (more…)