Google’s Site Reputation Abuse Policy: A Temporary Fix for a Systemic Problem
Google’s recently updated site reputation abuse policy aims to combat the growing issue of large, authoritative websites leveraging their domain...
2 min read
Writing Team : Dec 16, 2024 11:37:40 AM
A newly discovered exploit by Mark Williams-Cook has provided valuable insights into over 2,000 factors that Google uses to classify queries and websites. The findings shed light on processes like consensus scoring, query classifications, and site quality scores, offering a deeper understanding of how Google ranks content in its search results.
Watch him talk here -
The revelations from this exploit build upon previous insights into Google’s ranking systems, such as the Content API Warehouse leak earlier this year. These discoveries enhance our knowledge of scoring mechanisms, query classifications, and how site quality influences rankings.
One of the key revelations is Google's use of consensus scoring. This involves assessing how many passages within content align with, contradict, or remain neutral to a “general consensus.” This scoring system appears to influence rankings, especially for queries aimed at debunking misinformation (e.g., [Is the earth flat?]).
Google has categorized nearly all search queries into eight specific semantic classes:
These classifications determine how Google tailors its algorithm for each query type. For instance, since 2019, it has been understood that YMYL queries are weighted differently to ensure reliability and accuracy.
Google ranks results based on site quality scores, as detailed by Williams-Cook. These scores, calculated on a subdomain level, are influenced by factors such as:
Sites that fail to meet a minimum threshold (e.g., scoring below 0.4 on a 0-1 scale) may be excluded from search features like featured snippets or the "People Also Ask" section.
While Google does not directly use click-through rate (CTR) in its ranking algorithm, it does incorporate a “click probability” metric. This predicts how likely a user is to click on a specific result. Adjustments to elements like page titles can impact this probability. Williams-Cook noted that tools like the Google Ads Planner offer hints by estimating click-through rates.
Williams-Cook and his team analyzed 2 terabytes of data from over 90 million queries as part of their research. For uncovering this exploit, Google compensated the team with $13,337, acknowledging the significance of their findings.
These insights highlight the sophistication of Google’s ranking systems and underscore the importance of understanding factors like query classifications, consensus scoring, and site quality scores. These revelations provide actionable knowledge for content creators and SEO professionals to optimize content better and adapt to Google’s ever-evolving algorithms.
Google’s recently updated site reputation abuse policy aims to combat the growing issue of large, authoritative websites leveraging their domain...
1 min read
In a recent interview, Google's Gary Illyes provided insights into why the company trusts less in signals directly controlled by site owners and...
The Google August 2024 Core Update has officially completed its rollout, marking an impactful shift in the search engine landscape. The update, which...