QUOTE: “(Google Panda) measures the quality of a site pretty much by looking at the vast majority of the pages at least. But essentially allows us to take quality of the whole site into account when ranking pages from that particular site and adjust the ranking accordingly for the pages. So essentially, if you want a blunt answer, it will not devalue, it will actually demote. Basically, we figured that site is trying to game our systems, and unfortunately, successfully. So we will adjust the rank. We will push the site back just to make sure that it’s not working anymore.”  Gary Illyes – Search Engine Land

QUOTE: “I think there is probably a misunderstanding that there’s this one site-wide number that Google keeps for all websites and that’s not the case.  We look at lots of different factors and there’s not just this one site-wide quality score that we look at. So we try to look at a variety of different signals that come together, some of them are per page, some of them are more per site, but it’s not the case where there’s one number and it comes from these five pages on your website.” John Mueller, Google
Where the free Google tools can provide complementary value is in fact-checking. If you're checking out more than one of these SEO tools, you'll quickly realize this isn't an exact science. If you were to look at the PA, DA, and keyword difficulty scores across KWFinder.com, Moz, SpyFu, SEMrush, Ahrefs, AWR Cloud, and Searchmetrics for the same set of keywords, you might get different numbers across each metric separated by anywhere from a few points to dozens. If your business is unsure about an optimization campaign on a particular keyword, you can cross-check with data straight from a free AdWords account and Search Console. Another trick: Enable Incognito mode in your browser along with an extension like the free Moz Toolbar and you can run case-by-case searches on specific keywords to get an organic look at your target search results page.
When Google trusts you it’s because you’ve earned its trust to help it satisfy its users in the quickest and most profitable way possible. You’ve helped Google achieve its goals. It trusts you and it will reward you with higher rankings. Google will list “friends” it trusts the most (who it knows to be reputable in a particular topic) at the top of SERPs.
Think about how Google can algorithmically and manually determine the commercial intent of your website – think about the signals that differentiate a real small business website from a website created JUST to send visitors to another website with affiliate links, on every page, for instance; or adverts on your site, above the fold, etc, can be a clear indicator of a webmaster’s particular commercial intent – hence why Google has a Top Heavy Algorithm.
Naturally, business owners want to rank for lots of keywords in organic listings with their website. The challenge for webmasters and SEO is that Google doesn’t want business owners to rank for lots of keywords using autogenerated content especially when that produces A LOT of pages on a website using (for instance) a list of keyword variations page-to-page.

When you write a page title, you have a chance right at the beginning of the page to tell Google (and other search engines) if this is a spam site or a quality site – such as – have you repeated the keyword four times or only once? I think title tags, like everything else, should probably be as simple as possible, with the keyword once and perhaps a related term if possible.
An SEO expert could probably use a combination of AdWords for the initial data, Google Search Console for website monitoring, and Google Analytics for internal website data. Then the SEO expert can transform and analyze the data using a BI tool. The problem for most business users is that's simply not an effective use of time and resources. These tools exist to take the manual data gathering and granular, piecemeal detective work out of SEO. It's about making a process that's core to modern business success more easily accessible to someone who isn't an SEO consultant or expert.
QUOTE: “Content which is copied, but changed slightly from the original. This type of copying makes it difficult to find the exact matching original source. Sometimes just a few words are changed, or whole sentences are changed, or a “find and replace” modification is made, where one word is replaced with another throughout the text. These types of changes are deliberately done to make it difficult to find the original source of the content. We call this kind of content “copied with minimal alteration.” Google Search Quality Evaluator Guidelines March 2017

QUOTE: “7.4.3 Automatically ­Generated Main Content Entire websites may be created by designing a basic template from which hundreds or thousands of pages are created, sometimes using content from freely available sources (such as an RSS feed or API). These pages are created with no or very little time, effort, or expertise, and also have no editing or manual curation. Pages and websites made up of auto­generated content with no editing or manual curation, and no original content or value added for users, should be rated Lowest.” Google Search Quality Evaluator Guidelines 2017

Sometimes, Google turns up the dial on demands on ‘quality’, and if your site falls short, a website traffic crunch is assured. Some sites invite problems ignoring Google’s ‘rules’ and some sites inadvertently introduce technical problems to their site after the date of a major algorithm update and are then impacted negatively by later refreshes of the algorithm.
The emphasis on tools, meaning plural, is important because there's no one magical way to plop your website atop every single search results page, at least not organically, though there are best practices to do so. If you want to buy a paid search ad spot, then Google AdWords will happily take your money. This will certainly put your website at the top of Google's search results but always with an indicator that yours is a paid position. To win the more valuable and customer-trusted organic search spots (meaning those spots that start below all of those marked with an "Ad" icon), you must have a balanced and comprehensive SEO strategy in place.
John Lincoln is CEO of Ignite Visibility, one of the top digital marketing agencies in the nation and a 2017, 2018 and 2019 Inc. 5000 company. Lincoln is consistently named one of the top marketing experts in the industry. He has been recipient of the Search Engine Land "Search Marketer of the Year" award, named the #1 SEO consultant in the USA by Clutch.co, most admired CEO and 40 under 40. Lincoln has written two books (The Forecaster Method and Digital Influencer) and made two movies (SEO: The Movie and Social Media Marketing: The Movie) on digital marketing. He is a digital marketing strategy adviser to some of the biggest names in business.
Many search engine marketers think who you link out to (and who links to you) helps determine a topical community of sites in any field or a hub of authority. Quite simply, you want to be in that hub, at the centre if possible (however unlikely), but at least in it. I like to think of this one as a good thing to remember in the future as search engines get even better at determining topical relevancy of pages, but I have never actually seen any granular ranking benefit (for the page in question) from linking out.
However, we do expect websites of large companies and organizations to put a great deal of effort into creating a good user experience on their website, including having helpful SC. For large websites, SC may be one of the primary ways that users explore the website and find MC, and a lack of helpful SC on large websites with a lot of content may be a reason for a Low rating.
Google states, “News articles, Wikipedia articles, blog posts, magazine articles, forum discussions, and ratings from independent organizations can all be sources of reputation information” but they also state specifically boasts about a lot of internet traffic, for example, should not influence the quality rating of a web page. What should influence the reputation of a page is WHO has shared it on social media etc. rather than just raw numbers of shares. CONSIDER CREATING A PAGE with nofollow links to good reviews on other websites as proof of excellence.
Taboola (like many websites) would benefit from iterative SEO, meaning its team should make distinct revisions to the SEO page titles, content headers, and other text to see whether rankings improve without hurting their top positions. But they shouldn’t make too many changes at one time or they won’t know what worked and could disrupt rankings. They could monitor rankings after testing keyword-placement scenarios. (Most website pages are indexed within days or a couple weeks.)
Forget whether or not to put your keyword in it, make it relevant to a searcher and write it for humans, not search engines. If you want to have this 20-word meta description which accurately describes the page you have optimised (for one or two keyword phrases) when people use Google to search, make sure the keyword is in there. There is more likelihood it will be used by Google as your snippet description in SERPs.
KISS does not mean boring web pages. You can create stunning sites with smashing graphics – but you should build these sites using simple techniques – HTML & CSS, for instance. If you are new to web design, avoid things like Flash and JavaScript, especially for elements like scrolling news tickers, etc. These elements work fine for TV – but only cause problems for website visitors.
Mike Murray has shaped online marketing strategies for hundreds of businesses since 1997, including Fortune 500 companies. A former journalist, he has led SEO studies and spoken at regional and national Internet conferences. Founder of Online Marketing Coach, Mike is passionate about helping clients identify their best opportunities for online marketing success based on their strengths, his advice and industry trends. You can find him at his blog, Online Marketing Matters or on Twitter @mikeonlinecoach.
One common scam is the creation of "shadow" domains that funnel users to a site by using deceptive redirects. These shadow domains often will be owned by the SEO who claims to be working on a client's behalf. However, if the relationship sours, the SEO may point the domain to a different site, or even to a competitor's domain. If that happens, the client has paid to develop a competing site owned entirely by the SEO.
Advertising with Google won't have any effect on your site's presence in our search results. Google never accepts money to include or rank sites in our search results, and it costs nothing to appear in our organic search results. Free resources such as Search Console, the official Webmaster Central blog, and our discussion forum can provide you with a great deal of information about how to optimize your site for organic search.
Try and get links within page text pointing to your site with relevant, or at least, natural looking, keywords in the text link – not, for instance, in blogrolls or site-wide links. Try to ensure the links are not obviously “machine generated” e.g. site-wide links on forums or directories. Get links from pages, that in turn, have a lot of links to them, and you will soon see benefits.

Consider the length of your typical customer buying cycle. If your products and services have a short customer buying cycle, meaning your customers know what they want, search for it, and buy it, you may benefit from SEM ads that put your product right where customers will see it. Longer buying cycles, where customers research and compare for weeks or months, may not perform as well with SEM, as there isn’t an immediate buy after seeing one ad.
Experience can educate you when a page is high-quality and yet receives no traffic. If the page is thin, but is not manipulative, is indeed ‘unique’ and delivers on a purpose with little obvious detectable reason to mark it down, then you can say it is a high-quality page – just with very little search demand for it. Ignored content is not the same as ‘toxic’ content.
Rob Marvin served as PCMag's Associate Features Editor from December 2017 to December 2019. He wrote features, news, and trend stories on all manner of emerging technologies. Beats included: big tech coverage, startups, business and venture capital, blockchain and cryptocurrencies, artificial intelligence, augmented and virtual reality, IoT and aut... See Full Bio
When using the Keyword Explorer, Ahrefs will also produce the "parent topic" of the keyword you looked up, as you can see in the screenshot above, underneath the Keyword Difficulty meter. A keyword's parent topic is a broader keyword with higher search volume than your intended keyword, but likely has the same audience and ranking potential -- giving you more a valuable SEO opportunity when optimizing a particular blog post or webpage.
Google will select the best title it wants for your search snippet – and it will take that information from multiple sources, NOT just your page title element. A small title is often appended with more information about the domain. Sometimes, if Google is confident in the BRAND name, it will replace it with that (often adding it to the beginning of your title with a colon, or sometimes appending the end of your snippet title with the actual domain address the page belongs to).
AWR Cloud, our third Editors' Choice, is rated slightly lower than Moz Pro and SpyFu as an all-in-one SEO platform. However, AWR Cloud leads the pack in ongoing position monitoring and proactive search rank tracking on top of solid overall functionality. On the ad hoc keyword research front, the KWFinder.com tool excels. DeepCrawl's laser focus on comprehensive domain scanning is unmatched for site crawling, while Ahrefs and Majetic can duke it out for the best internet-wide crawling index. When it comes to backlinks tracking, LinkResearchTools and Majestic are the top choices. SEMrush and Searchmetrics do a bit of everything.
So you have a new site. You fill your home page meta tags with the 20 keywords you want to rank for – hey, that’s what optimisation is all about, isn’t it? You’ve just told Google by the third line of text what to filter you for. The meta name=”Keywords” was actually originally for words that weren’t actually on the page that would help classify the document.
In particular, the Google web spam team is currently waging a PR war on sites that rely on unnatural links and other ‘manipulative’ tactics (and handing out severe penalties if it detects them). And that’s on top of many algorithms already designed to look for other manipulative tactics (like keyword stuffing or boilerplate spun text across pages).
Structured data21 is code that you can add to your sites' pages to describe your content to search engines, so they can better understand what's on your pages. Search engines can use this understanding to display your content in useful (and eye-catching!) ways in search results. That, in turn, can help you attract just the right kind of customers for your business.

Provide full functionality on all devices. Mobile users expect the same functionality - such as commenting and check-out - and content on mobile as well as on all other devices that your website supports. In addition to textual content, make sure that all important images and videos are embedded and accessible on mobile devices. For search engines, provide all structured data and other metadata - such as titles, descriptions, link-elements, and other meta-tags - on all versions of the pages.


At first glance, the Ads or SC appear to be MC. Some users may interact with Ads or SC, believing that the Ads or SC is the MC.Ads appear to be SC (links) where the user would expect that clicking the link will take them to another page within the same website, but actually take them to a different website. Some users may feel surprised or confused when clicking SC or links that go to a page on a completely different website.

If you are just starting out, don’t think you can fool Google about everything all the time. Google has VERY probably seen your tactics before. So, it’s best to keep your plan simple. GET RELEVANT. GET REPUTABLE. Aim for a healthy, satisfying visitor experience. If you are just starting out – you may as well learn how to do it within Google’s Webmaster Guidelines first. Make a decision, early, if you are going to follow Google’s guidelines, or not, and stick to it. Don’t be caught in the middle with an important project. Do not always follow the herd.
When I think ‘Google-friendly’ these days – I think a website Google will rank top, if popular and accessible enough, and won’t drop like a f*&^ing stone for no apparent reason one day, even though I followed the Google SEO starter guide to the letter….. just because Google has found something it doesn’t like – or has classified my site as undesirable one day.
QUOTE: “alt attribute should be used to describe the image. So if you have an image of a big blue pineapple chair you should use the alt tag that best describes it, which is alt=”big blue pineapple chair.” title attribute should be used when the image is a hyperlink to a specific page. The title attribute should contain information about what will happen when you click on the image. For example, if the image will get larger, it should read something like, title=”View a larger version of the big blue pineapple chair image.” John Mueller, Google
Many websites rely on other traffic generation methods such as traffic from social media, email, referrals, and direct traffic sources over search engines. For sites like these, SEO errors aren’t as important because search engines aren’t their #1 traffic source. For a smaller website, a couple of errors can have a much bigger negative effect than those same errors on a larger website.
×