3 Lessons for Search Engine Optimization After SMX Advanced 2012

There were 3 big lessons coming out of SMX Advanced this week.

1. SEO is hard.
2. Spamming/shortcuts have a short life expectancy.
3. Focus on marketing ━ presenting your biz in the best light to your audience.

You can think about how much of a “lesson” those really are since you probably already knew all that. Yet the conference gave a fresh face to these issues, which search marketers have dealt with since the inception of the industry.

Discussions tackled challenges posed by search personalization and code markups for enhanced search feature. And of course there was lots of talk about Google’s invigorated efforts to cut out cheap SEO tricks link paid links… lots of talk about Penguin.

This is the story of SMX Advanced 2012 and the state of the search engine optimization industry.

Helpful resources that go into detail on topics covered at SMX:

pen

SEO Is Hard

A feeling of relief washed over me when Matt Cutts, head of the Google web spam team, tells Search Engine Land editor Danny Sullivan that SEO is hard. (Internal reaction: I’m not alone in thinking so! I’m not crazy!) Unfortunately a second later I had to swallow down the hard reality that it’s just getting harder. Matt’s comment was along the lines of: Ask any SEO if it’s harder than 5-7 years ago, they’ll say yes. You can expect that metabolism to continue.

Breaking down the specific search engine marketing challenges was the purpose of the conference, and pulling from the sessions I saw, I’ll highlight four things that are shaping the nature of search and searcher behavior.

100% of Search Is Personalized
Ranking-based SEO died with the end of the 10 blue links. Personalization of search results today is increasingly complex and no two users’ results pages are the same. During the session Surviving Personalization with Google and Bing, Danny Sullivan, editor of Search Engine Land, brought home the case for optimizing for a personal search environment by explaining that 100% of search results are personalized. Even with personal search turned off in Google via the World toggle, results are still personalized by location. Tactics for personal search must be part of your SEO strategy.

Microformats Add Elements to Search
Author pics and info, star ratings, rich snippets ━ microformats and Schema.org are helping search engines deliver more info within results pages. Rich snippets improve a user experience, increasing the likelihood a searcher will find your content relevant to their search. Implementing authorship markup adds authority and personality to a result, as well as an image which draws user eyeballs. Ratings establish brand trust and authority. Search marketers are marketers after all. Take every advantage to market in SERPs. Implement available, relevant elements to make your listing stand out to your target audience.

Providing a Good Mobile Experience

Searchers use mobile devices a lot. As Carlos del Rio shared at the session iConvert: Landing Pages and Conversion in a Smartphone and Tablet World, wireless Internet connections outnumber wired connections 2:1. The ubiquity of connecting via mobile devices means many things for businesses aiming to provide a good experience to customers online. Growing mobile traffic has implications for how users use a site and the location and type of calls to actions used.

Conscious strategy must be dedicated to mobile site design and content delivery. Google this week recommended responsive site design, a way of delivering a smartphone and tablet experience when a mobile device is detected. And mobile browsing means there’s a significant demand for SPEED. Site abandonment skyrockets after six second load times.

Google Is a Website’s Friend
From time to time it can feel like SEO is in opposition to Google. It’s not uncommon to read about Google’s “war against SEO” or efforts to “kill SEO”, even if only as headline bait. But we got a view of Google’s perspective of SEO from Maile Ohye, Google Developer Programs Tech Lead during her presentation on pagination and canonicalization issues.

She gave credit to SMX Advanced attendees for helpful feedback that has resulted in support for new protocols and tools. In 2011 Maile heard of one company’s application of rel=canonical to consolidate paginated sets even though content was not duplicated. Maile went back to her team with that info and 5 months later Google offered support for the rel=prev/next protocol. Google also improved the URL parameters tool in Webmaster Tools after feedback from the SEO community.

My takeaway: Google understands that the commercial Web environment can pose issues for crawling and indexing; Google is willing to work with site owners to better understand the problems so the search engine can go about giving users helpful, relevant info they’re looking for.

Linking: The Rant

One of the most memorable events of the conference came during the last session of the final day in the always packed Ask the SEOs session. The now famous “rant” by Danny Sullivan provides valuable historical context and a reminder of the purpose of links. It comes at a time when it’s easy to get lost in the barking and moaning about Google Penguin and who’s to blame for the rise of link manipulation.

Danny took us back to the time when links were used to point to more information about a topic. By contrast, link building methods used today often look at links as the end goal rather than a means to an end. Before Google was the primary means of online content discovery, there were three directories: Yahoo!, the Open Directory and Look Smart. Danny recalled a time when directories were used to find things. Likewise, when Google came along, they used these directories to locate sites to crawl and index on the Web.

When search adoption grew, website owners asked Google how to get their sites indexed. Google opened a Pandora’s box when they suggested sites get listed in directories. Now, 20,000 directories later, link spam has evolved.

In the aftermath of Penguin, the possibility of a link disavow option has been raised to help sites clean up a tarnished link profile. During the Q&A keynote, Matt Cutts mentioned that such a protocol may be on the way, but that if it is, it’s months away. He advises that webmasters and SEOs today make their best efforts to remove low quality links and to document the case for Google. He explained that if they sample the low quality backlink profile and find that, for instance, 90 percent of links are gone, the good faith effort will be accepted.

In the same keynote, Danny asked Matt if Google was in a war against SEO. To that Matt said that, if anything, he’s engaged in a battle against spam. In his own words, he summed up the mission of Google in its relationship with the SEO community this year as being about improved transparency and about policing its own quality guidelines. Matt described spammers as those who are essentially asking, “How can I fake being awesome?” If you’ve ever found yourself asking that question, refer to number 1, 2 and 3 above.

Virginia Nussey is the director of content marketing at MobileMonkey. Prior to joining this startup in 2018, Virginia was the operations and content manager at Bruce Clay Inc., having joined the company in 2008 as a writer and blogger.

See Virginia's author page for links to connect on social media.

Comments (8)
Filed under: SEO — Tags:
Still on the hunt for actionable tips and insights? Each of these recent SEO posts is better than the last!

8 Replies to “3 Lessons for Search Engine Optimization After SMX Advanced 2012”

Thanks for the update Virginia… I’m really curious as to how the larger sites will work on getting the 10,000 bad links out of the system in one shot. Definitely think that Google should disregard the existing link juice and focus on the clickthroughs the links get from high quality sites… What I would’ve loved to hear more on how much more google+ and other social media sites can be gamed as linking used to be in the past…

I do hate what I am doing (SEO) because it will never finish. And I hate Google sometimes because we have no way to catch up.

But, once you realize it is a human world. Put heavy focus on human is the key thing. Don’t optimize for machine, but for human. It does reduce a lot of unnecessary works.

Google is continually tweaking the algorithm and SEO is always evolving. If you aren’t keeping up with trends and continue to use outdated tactics (like expecting links from hundreds of low quality directories to mean anything) you won’t succeed. Focus on local/personalized search and creating content that target audience members will actually find beneficial.

I’d like to give my two cents about SEO being hard and spamming/shortcuts having a short life expectancy…

1) SEO is hard. My opinion? This can also work FOR YOU. You see, the harder it gets the steeper the learning curve becomes and, when you’re not in the SEO industry itself, that means that competitors will have a harder time learning what you already know. It’ll be a barrier for market entry. Knowledge is and always will be power. I like the fact that SEO is hard, because even in the SEO industry I’m in I see new guys struggling with Panda, Penguin, etc. while I know how not to struggle with it.

2) Spamming/shortcuts have a short life expectancy: things never changed, because this was always the case. You can’t succeed online (or offline) by taking shortcuts. There is NO such thing as a magic pill that grants success overnight. Success requires sacrifice and usually the less sacrifice it requires the shorter lived it’ll be. Taking shortcuts leads to more cuts as in… seeing your website traffic and income cut in half, or worse. With all due respect, but you can’t win the online game by running around and throwing money at stuff anymore. You need to have a real strategy. As we always say: get rich quick doesn’t exist… but get richer slowly does!

Regards,

Dennis Miedema

It’s interesting to see that Matt Cutts is still advising the manual removal of low quality links. Surely Google realise the difficulty involved in manually removing links?

Identifying low quality links is one thing however removing them is a completely different story. Even if you manage to identify whom to contact to get any low quality links removed there’s still no guarantee thet they will actually be removed. Unless you’ve participated in some hardcore link schemes (such as Build My Rank) whereby links can be revoked then manually removing links is not an easy task!

Not to mention the issues of malicious negative SEO! If your website has been (or still is!) a target for negative SEO it could be a continue battle to remove low-quality links.

I understand Google’s fight against low-quality links is on-going, but putting all of the burden on webmasters to manually remove the links doesn’t quite fit with me.

I could not agree more with the points in this post. SEO is a long, tedious process to do correctly. The more people take shortcuts, the more they can expect their rankings to head south. Great recap, thanks for sharing Virginia!

Thanks for letting us inside and the helpful links. I don’t think of SEO as being hard as much as to do it right, demands patience. Thanks again Virginia.

Thanks for the link to the Danny Sullivan rant – it’s really worth listening!
I think that right now the SEO market is changing – even in Austria where SEO-trends are picked up like months later.

LEAVE A REPLY

Your email address will not be published. Required fields are marked *

Serving North America based in the Los Angeles Metropolitan Area
Bruce Clay, Inc. | PO Box 1338 | Moorpark CA, 93020
Voice: 1-805-517-1900 | Toll Free: 1-866-517-1900 | Fax: 1-805-517-1919