Google processes 3.5 billion searches every single day. And somewhere in that flood of queries, your carefully written blog post is getting ignored - not because the content is bad, but because the title is speaking a language Google no longer prioritises.
I know this intimately. A few years back, I spent three weeks building out a content cluster for a SaaS client, keyword-optimised titles and all. Solid search volume, reasonable competition.
The results? Buried on page two, collecting dust.
The titles were technically correct. They just weren't semantically meaningful to Google's increasingly sophisticated understanding of what a piece of content is actually about.
That failure sent me down a rabbit hole into Google's Natural Language Processing (NLP) API - a free tool that lets you see, with uncomfortable clarity, how Google reads and categorises your content. What I found changed how I approach every title I write. Titles are consistently ranked among the top three on-page ranking factors, yet the way most content teams write them hasn't caught up with how Google now evaluates them.
This article walks through the full process: why traditional title formulas are losing ground, how NLP entities and salience scores reveal what Google actually values, and how to use that data to draft and refine titles that match both search intent and semantic context. You will also get a practical look at iterative testing with the NLP API, plus a framework for keeping titles competitive as the SERP landscape shifts.
No coding required. No jargon overload. Just a methodical, data-backed approach to one of the highest-leverage tweaks in SEO - writing a title that Google understands, and readers want to click.
Why Your Current Titles Aren't Landing Traffic
Only 15% of daily searches on Google are queries the engine has never seen before - and yet Google handles billions of searches a day. That single statistic should make you uncomfortable about any title strategy built purely around keyword lists.
Keyword matching was the dominant logic for years. You found a phrase with search volume, put it in your title, and waited. That approach worked reasonably well until 2019, when Google's BERT update fundamentally changed how the engine reads text - not word by word, but by understanding the relationship between words in context.
BERT wasn't a minor tweak. It was Google admitting that matching strings of characters to a query was a crude proxy for understanding what a searcher actually wanted. Then MUM arrived in 2021, pushing that capability further - handling nuance, multilingual context, and complex intent that a keyword-stuffed title has no hope of satisfying.
So if your titles still read like a keyword tool threw up on them - "Best CRM Software Best CRM for Small Business 2024" - you're not optimising for Google. You're optimising for a version of Google that no longer exists.
If your title doesn't accurately reflect your page content, Google will replace it in the SERP with text it pulls from your page - usually your H1 or a sentence it deems more representative.
That replacement behaviour is telling. Google isn't just reading your title tag; it's evaluating it against the full semantic context of your page. Tools like the Google Natural Language API can surface exactly what entities and categories Google extracts from a piece of text - which is precisely why some SEOs are already using that kind of analysis to audit titles before publishing.
Vague titles cause the same problem, just from the opposite direction. A title like "Content Tips for Marketers" signals almost nothing about intent. Is this a beginner's guide?
A tactical breakdown? A tool comparison?
Google doesn't know. Your reader doesn't know. And ambiguity at the title level costs you the click before the content ever gets a chance.
The uncomfortable truth is that most title strategies fail at intent matching, not keyword selection. You can rank a page for a term and still bleed clicks because the title misrepresents what the content delivers - or because it's so generic that a competitor's more specific title wins the tap. Research consistently shows that shorter titles for SEO tend to outperform bloated ones, partly because clarity suffers the moment you try to pack three keyword variants into a single phrase.
Stuffing, vagueness, and misalignment with content aren't just stylistic problems. They're signals Google now has the architecture to detect and penalise - quietly, without a manual action, just by choosing to display something else instead of what you wrote.
Decoding Google's Brain with NLP Entities
You can spend hours debating title formulas, or you can look directly at the data Google already uses to evaluate your content. The second approach wins. Every time.
The Google Natural Language API Demo at cloud.google.com/natural-language is free, requires no coding, and does something most SEO tools can't: it shows you how Google's own systems classify and weight the content on any page you feed it. That includes your competitors' top-ranking pages.
What the API Actually Tells You
Paste a competitor's full page content into the demo and hit "Analyze." Two outputs matter most here.
The Entities section lists every person, organisation, location, event, price, and concept Google detected - each with a salience score between 0 and 1. That score is Google's signal for how central each entity is to the overall content. A salience score of 0.4+ on your target topic means it's dominating the semantic fingerprint of that page. Below 0.1, and Google barely registers it as relevant.
The Categories section shows content classification with a confidence score. Aim for a confidence score above 0.85 in your target category - below that, Google is essentially guessing what your page is about, which is a ranking liability you don't want.
Run the same analysis on pages ranking on page two of results - the entity patterns that are missing from those pages are just as instructive as what the top-ranking pages include.
Running the Competitor Analysis
Budget 15–30 minutes per competitor page. That's the realistic time to do this properly - not a quick skim, but a structured read of the entity list.
- Pull the top 1–3 organic results for your target keyword and copy each page's full text into the demo tool.
- Record high-salience entities - anything scoring above 0.2 is worth noting. These are the concepts Google associates most strongly with the topic.
- Note the category confidence scores - if three competing pages all classify at 0.90+ in the same subcategory, that's your target benchmark.
- Identify LSI patterns - Latent Semantic Indexing (LSI) keywords are semantically related terms that appear consistently across top-ranking pages. The entity list surfaces these automatically. They belong in your title considerations, not just your body copy.
- Repeat for page-two results - compare entity profiles against page-one winners to spot the semantic gaps holding those pages back.
After reviewing dozens of these analyses, the pattern is consistent: pages that rank well don't just match the primary keyword - they carry a dense, coherent cluster of related entities that reinforce the topic from multiple angles. This is exactly what Google AI search summaries now reward at the top of results.
The uncomfortable question this raises: if your title contains the right keyword but none of the high-salience entities your competitors carry, what exactly are you signalling to Google's classifier?
Crafting High-Salience Titles from Entity Data
Titles built on gut instinct and a rough keyword idea tend to get quietly rewritten by Google in the SERPs - which is a polite way of saying Google doesn't trust them to describe the page accurately. The entity data you've already pulled from competitor analysis is the fix for that. The question is how to turn a spreadsheet of salience scores into a title that actually works.
Start with your highest-salience entities - the ones scoring closest to 1.0 in the NLP API output. Those are the concepts Google considers central to the topic, not peripheral. Your title needs at least one or two of them front-loaded, because Google's parsing of a title isn't uniform; prominence matters.
But this isn't just a keyword-insertion exercise. Salience without intent match is wasted signal. A title can contain every high-scoring entity from your competitor analysis and still fail if it doesn't answer what the user actually came to find. "Best CRM Software" hits the entity. "Best CRM Software for Small Teams Under 10 People" hits the entity and the intent.
Question-based formats deserve more credit than they get. The NLP API specifically looks for interrogative structures to match against user queries - which is exactly why question titles show up in "People Also Ask" boxes. "What Is the Fastest CRM for Small Teams?" is a dead simple format, but it maps cleanly to both entity recognition and featured snippet eligibility.
Run your drafted titles back through the NLP API demo before committing - check whether your intended entities actually surface with high salience, or whether the API reads your title as being about something adjacent.
For initial drafts at volume, tools like TextBuilder AutoWriter are worth knowing about here. Its LLM aggregation approach - routing different parts of content generation to the model best suited for each task - produces titles with more semantic variety than single-model outputs, which helps when you need 20 candidate titles rather than three.
Avoid two specific traps. First, keyword stuffing: "Best CRM Software CRM Tools Small Business CRM" tanks readability and Google's NLP reads the repetition as low-quality signal. Second, vague language: titles like "Everything You Need to Know About CRM" contain no salient entities and give Google nothing concrete to classify.
- Pull your top 3-5 entities by salience score from the competitor analysis
- Identify the primary user intent behind the search (informational, comparative, transactional)
- Draft 5-8 title variations - mix statement formats with question formats
- Integrate LSI keywords (semantically related terms) naturally, not forced
- Flag any title with repeated root keywords for immediate revision
After running this process on roughly 50 title sets, the pattern is clear: the titles that hold their position in SERPs are the ones where the entity structure in the title matches the entity structure of the body content. A mismatch there - even a subtle one - and Google substitutes its own version.
Iterative Refinement with the NLP API
Your first draft title almost never survives contact with the API. That's not a failure - that's the process working exactly as it should.
After you've built your initial drafts from entity data, paste each title directly into the Google Natural Language API demo and run the analysis. What you're looking for is specific: do the entities the API detects match your intended topic, and does the category confidence score clear the 0.85 threshold? Below that number, Google is essentially uncertain about what your content covers. That uncertainty costs you rankings.
The Refinement Loop
- Check Entity Salience First - If your primary topic entity returns a low salience score, the title is burying the lead. Restructure so the most important entity appears earlier and with tighter surrounding language. Cut anything that dilutes focus.
- Fix Wrong Entity Detection - Ambiguous terms get misclassified constantly. "Apple," "Sage," "Mercury" - the API reads context, not intent. Add a qualifying word to disambiguate. If the API still misreads the entity, structured data via Schema markup gives Google the explicit signal the title can't.
- Push the Category Score Above 0.85 - A low confidence score means your title is too vague or too broad. Make the subject explicit. Concrete language outperforms clever language every time in categorization.
- Audit for Uniqueness and Length - Google rewrites titles it considers too long, inaccurate, or boilerplate. If your title duplicates phrasing used across other pages on your site, or runs past the display width Google allocates in SERPs, expect it to be replaced with something you didn't write. Neither outcome is acceptable.
- Retest After Every Change - A single word swap can shift salience scores meaningfully. This isn't theoretical - I've seen a title's primary entity drop from 0.91 salience to 0.43 after adding a subtitle that introduced a competing concept. Run the API again every time you edit.
Budget the time honestly. Initial drafting and the first round of API testing runs 1–2 hours. Revisions - depending on how many titles you're refining and how far off the first pass lands - add another 1–2 hours on top. That's a real time commitment, and it's worth it, but don't treat this as a 20-minute task.
One thing worth flagging: the refinement work you do now has a shelf life. Search categories shift, competitor content changes, and what scores 0.91 today may read differently to the API six months from now as Google's understanding of your topic space evolves. The teams that treat this as a quarterly audit rather than a one-time fix consistently outperform those that don't.
A title that passes every NLP check today still faces the question of whether it will hold that position as the SERP around it changes.
Sustaining Title Performance in a Changing SERP
Titles you optimised last quarter are already competing against a Google that has learned something new since then. BERT and MUM don't sit still, and neither does your competition - which means a title that earned position three in January can quietly slip to page two by March without a single change on your end.
Title decay is rarely dramatic. It creeps. You stop watching the data for a few weeks, and by the time you notice the traffic drop, a competitor has already claimed the semantic ground you abandoned.
Set aside 1–2 hours weekly for monitoring keyword trends and competitor movement. That's not a suggestion - it's the minimum viable effort to catch a ranking shift before it compounds. When you do spot a title losing ground, budget 30–45 minutes per piece for content optimisation, including the title itself.
If Google is replacing your title in the SERP with its own rewrite, that's a direct signal: your title no longer matches the content or the dominant user intent Google has identified for that query. Rewrite immediately.
Google rewriting your title in search results is not a cosmetic issue. It means the algorithm has decided your title misrepresents what the page actually delivers. Fix the mismatch - either update the title to match the content, or update the content to match the intent your title implies.
Beyond titles, the broader page signals matter more than most people account for. NLP-driven title optimisation is one layer of a larger system. Page speed, internal linking structure, external links to authoritative sources, and mobile experience all feed into how Google evaluates the page as a whole. A semantically perfect title sitting on a slow, poorly linked page still loses.
The Pitfall of Set-and-Forget AI
AI-generated titles are a starting point. Full stop. Tools that handle programmatic SEO - semantic richness, automated external linking, structured content formatting - handle the scaffolding well. But human refinement is what separates a ranking title from a generic one that Google ignores or rewrites.
I've watched well-structured AI content plateau at position eight because the titles were technically correct but tonally flat - no brand voice, no specificity, nothing that matched how the target audience actually searches. The NLP signals were there. The human layer wasn't.
- Check Google Search Console weekly for title rewrites and CTR drops
- Re-run your titles through the Natural Language API when you update underlying content
- Audit competitor titles for new entity patterns every 30 days
- Verify that your title's primary entity still carries the highest salience score after any content edits
- Never let an AI-generated title publish without a human read-through for brand alignment
Google's understanding of your content evolves whether you update the page or not. The index is a moving target, and your title optimisation strategy has to move with it.
Conclusion
Google stopped rewarding keyword density years ago. What it rewards now is semantic clarity - titles that signal the right entities, at the right salience, in the right category context. That shift is the whole argument of this article.
The NLP API demo isn't a gimmick. It's a window into how Google actually reads your content. Use it like one.
- Salience scores matter more than keyword frequency. A title that surfaces your core entity with a high salience score tells Google what the page is about - not just what words it contains.
- Category confidence below 0.85 is a warning sign. If Google can't confidently classify your content, your title likely needs more topical focus, not more keywords.
- Competitor analysis takes roughly 15–30 minutes per page in the NLP demo. That's a small investment for understanding exactly which entity patterns are earning top rankings in your niche.
- Titles getting rewritten in the SERPs is a symptom, not a random penalty. It means your title and your content are telling Google different stories.
- NLP optimisation isn't a one-time task. Algorithms like BERT and MUM continue to evolve, and titles that performed well six months ago may need refinement today.
Two things you can do right now: open the Google Natural Language API demo at cloud.google.com/natural-language and run your top three competitor URLs through the entity analyser - note every high-salience term. Then paste your current title into the same tool and compare what's missing.
The gap between those two analyses is your optimisation roadmap.
