Matt Cutts Just Announced A Google Algorithm Change

Labels: , , , 1 comments

says it will reduce low-quality “exact-match” domains in search results. It sounds like an extension of the last change he tweeted about, which was aimed at improving domain diversity. Here’s the new tweet:

Update: Cutts tweeted a follow-up:

Probably good of him to clear that up right away.
Google is about due to publish its big list of algorithm changes for the months of August and September. When that happens, it will be interesting to see how many entries are related to domains. It seems like there are typically visible themes in the lists. For example, in the June list, there were a lot of changes related to improving how Google deals with natural language.
Have you seen any effects from this update? Let us know with your comments.

Google News Gets A New Ranking Signal, And It’s A Keywords Meta Tag

Labels: , , , 0 comments

Google announced the news_keywords metatag for publishers in Google News to help Google better identify and understand content that is related to things that are in the news.
Do you think this is a good direction for Google News? Let us know what you think.
Here’s what it looks like:
<meta name=”news_keywords” content=”World Cup, Brazil 2014, Spain vs Netherlands, soccer, football”>
If you use it, use commas to separate phrases. You can add up to ten phrases per article, and each keyword is given equal value.
The company says it’s a way to empower writers to express stories freely, while helping Google News propertly understand and classify content. In a blog post, Google News product manager Rudy Galfi explains the thought process behind the feature:
The day after the historic 1929 stock market crash, Variety bannered their front page with these words: “WALL ST. LAYS AN EGG.” It’s a great headline: pithy, catchy, and expressive of the substance of the story as well as the scale of its consequences. It’s also worth noting that Variety’s editors had a full day to write the headline—millions of readers weren’t trying to search for the story within seconds of hearing about it.
The Web has transformed both how news organizations report information and the way users find it. Imagine if “WALL ST. LAYS AN EGG” were used as a headline today by an online news site. Since the headline is a sequence of text that’s only readily understandable by a human, most machine algorithms would probably attach some sort of biological association to it. In turn, this would make it difficult for millions of curious users who are using Google.com or Google News to find the best article about the stock market crash they just heard about.
With the news_keywords metatag, publishers can specify specific keywords that apply to news articles, basically like the classic keywords metatag.
The whole thing is pretty interesting, considering that Google has downplayed the regular keywords metatag. In fact, earlier this year, in a Webmaster Help video, Matt Cutts said, “You shouldn’t spend time on the meta keywords tag. We don’t use it. I’m not aware of any major search engine that uses it these days.”
Of course, this is a different tag, and it’s specifically news-related, though news results often appear in regular Google results. Cutts did say in a tweet:

Google is careful to note that the tag will be only “one signal among many” that its algorithms use to determine ranking.
“The news_keywords metatag is intended as a tool — but high-quality reporting and interesting news content remain the strongest ways to put your newsroom’s work in front of Google News users,” says Galfi.
Keep in mind, Google still frowns upon keyword stuffing (unless that’s going away in an upcoming version of its Webmaster Guidelines, which is highly doubtful).
In case you need a refresher, here’s Google’s quality guidelines for News:
News content. Sites included in Google News should offer timely reporting on matters that are important or interesting to our audience. We generally do not include how-to articles, advice columns, job postings, or strictly informational content such as weather forecasts and stock data.
We mean it — stick to the news! Google News is not a marketing service. We don’t want to send users to sites created primarily for promoting a product or organization.
Unique articles. Original reporting and honest attribution are longstanding journalistic values. (If your site publishes aggregated content, you will need to separate it from your original work, or restrict our access to those aggregated articles via your robots.txt file.)
Authority. Write what you know! The best news sites exhibit clear authority and expertise.
Accountability. Users tell us they value news sites with author biographies and clearly accessible contact information, such as physical and email addresses, and phone numbers.
User-friendly. Sites should load quickly and use URL redirects rarely. Clearly written articles with correct spelling and grammar also make for a much better user experience. Keep in mind that we can only include sites that follow the Webmaster Guidelines.

Google Launched An Update This Week To Improve Domain Diversity

Labels: , , , , 1 comments

Google launched an algorithm update that affects the diversity of search results. Google’s head of webspam and Distinguished Engineer, tweeted:

Matt Cutts
@mattcutts

Just fyi, we rolled out a small algo change this week that improves the diversity of search results in terms of different domains returned.

Reply · Retweet · Favorite
20 minutes ago via web · powered by @socialditto

There have been complaints in recent weeks about Google showing search results pages with a lot of results from the same domain for a lot of queries. Presumably that will be better now, and users will get a more diverse set of results in more cases. Or maybe it’s just about spreading the love among more domains in general (and not just per page).

That’s as much as we know about the update for now, but it’ll be interesting to see if the change is noticeable on a day to day basis.

There has been talk from webmasters that there may have been a new Panda update this week. We’ve not heard from Google on that front, and it’s unclear at this point whether this could have been the change people were noticing.

Google’s big list of algorithm changes for the month of August is due out any time now, and when it’s released, we’ll get more insight into the direction Google is going on, and its core areas of focus in recent weeks. Stay tuned.

Fear Of Google Ironically Has People Considering Making Natural Links Unnatural

Labels: , 0 comments

We recently published an article called “Links Are The Web’s Building Blocks, And Fear Of Google Has Them Crumbling“. This was about the panic Google has caused among webmasters with its messages about links. It’s a panic that has led to many webmasters requesting to have links removed from sites that they would otherwise find valuable, if not for fear that Google will not like them and hurt their rankings.

Is all of this fear over Google an overreaction, or is it justified? Let us know what you think.

I noticed a post in WebmasterWorld that expresses this point to perfectly. The title of the post says it all: “New link to my site worries me — but it’s a good link!” Senior member crobb305 writes:

Got an unsolicited citation from a media source but they used anchor text that I have been penalized on. FUD! Should I ask them to change it? By doing that I make a natural link unnatural, and Googlebot will detect that change (obvious tinkering). Nevertheless, I do have an OOP and received the infamous link warnings about 5 months ago.

I hate it that we have to live with this type of fear.

This person has been a member of WebmasterWorld since 2002, so they’ve clearly been in this world for quite a while. Yet here they are concerned that a completely natural link might draw negative attention from Google. The person is even wondering if they should go out of their way to make the link unnatural to please Google. How’s that for irony? Sadly, it’s highly likely that plenty of other webmasters are thinking similar thoughts.

As shared in the article mentioned at the beginning, there is plenty of overreaction from webmasters out there, and I would say that Google would rather see the link occur in its natural form, but this is the kind of fear people are dealing with to please Google and maintain some form of visibility in search results (which is getting harder and harder for other reasons entirely). Should people have to be this worried about links (the building blocks of the web)?

It probably doesn’t help that Google has reportedly indicated that forthcoming algorithm updates will be more “jarring.”

Another forum senior member later responded, “But seriously, some of the sites of mine that went down the Google drain were clean, ‘link building’ was not done, just attracted some real nice ones and yet the project died due to ‘penalties’. I went out of answers to this somewhere in the middle of 2011 and focus on cool stuff, HTML5, content (I think some tools can be considered good content) and ultimately ranking solid on Bing. Google does whatever Google wants to do.”

Likewise, Chris Lang from Gadget MVP tells me on Google+, “I never have worried about Google. I just do what seems natural. Never been slapped once…. At least not by Google.”

WebmasterWorld moderator goodroi tells the user, “One link from a quality, relevant website is not the problem. The hundreds of links with identical anchor text coming from blog spam, directory submission schemes and other short cuts are the problem.’

“I tend to focus more of my efforts on improving backlink profiles by adding quality links instead of focusing on deleting bad links,” goodroi adds. “Even if you delete every single bad link (and somehow are lucky not to accidentally delete a good link) you still need to build legitimate links. So if you start working on legitimate links you may end up getting enough good links that it naturally defuses the bad link issues.”

Unfortunately, many are seemingly still eager to kill significantly more links than they may really need to. On the flipside, even some publishers are growing leery of including guest content on their sites. This fear, apparently is coming from the Penguin update.

Barry Schwartz at Search Engine Roundtable points to a post from Cre8asite Forums, where user EGOL writes:

Since Penguin, I am getting a flood of article offers. Most of this content is crap. Some of it is “average” quality (which I don’t publish). Some can be excellent, unique, highly desirable. So now I am deciding if I want to accept some of this content, knowing that I could be publishing links to sites that could have past, present or future manipulation.

I have a potential article that I really like and that would be very popular with my visitors. The author’s site ranks #1 in a difficult niche and they don’t have enough content on their site to hold that position from editorial links (IMO).

I have not seen any articles or discussion about the cautions that a publisher should be following in these days of post-penguin linking.

So, not only are people afraid to have links out there that they would find valuable, if not for fear of Google, but some are also afraid to publish quality content, for fear that it might somehow be connected to something Google will not like. Ironically, quality content is what Google wants from sites above all else.

Are webmasters worrying about Google too much, or are these simply rational concerns, with Google being such a dominant force on the web? Comments please...

 
Internet Marketing Expert, SEO Latest Google Updates - Naveen Kumar © 2012 | Designed by Meingames and Bubble shooter