Stephanie asks, “Should I be creating content for search engines or people?”
With the advent of modern, AI-driven SEO, this is largely the same thing. Transformer-based natural language processing makes writing for people and writing for machines much closer, thanks to the way machines understand and predict. Models like BERT now understand sentences and paragraphs, which is a major change from keywords. They also specialize in tasks like question answering and prediction. Create content for these model applications.
Can’t see anything? Watch it on YouTube here.
Listen to the audio here:
- Got a question for You Ask, I’ll Answer? Submit it here!
- Subscribe to my weekly newsletter for more useful marketing tips.
- Find older episodes of You Ask, I Answer on my YouTube channel.
- Need help with your company’s data and analytics? Let me know!
- Join my free Slack group for marketers interested in analytics!
What follows is an AI-generated transcript. The transcript may contain errors and is not a substitute for watching the video.
In today’s episode, Stephanie asks, should I be creating content for search engines or for people? Kind of the same thing.
In the old days of SEO, we would focus very heavily on things like individual words or keyword phrases that we needed to create content around, that we needed to create, you know, text for, that took advantage of the way that search engines used to work in terms of how they identified what a page was about how they identified what what were the most important words on a page with the advent of the more sophisticated language models, in search engines that those tactics have largely gone by the wayside, not because we stopped using those words or phrases, but because of the way the engines themselves can understand language.
A few years ago, 2015, we started to see the advent of these new models called transformers, not anything related to the the toy line that can process text in very different ways.
And probably one of the more important and impactful frameworks that came out of this was one called bidirectional encoding representations for transformers, which we now know as Google’s Bert model.
And what these do differently, then, the weight search engines used to process text in the old days is they understand at the sentence and paragraph level, almost at the document level, as opposed to the word the phrase level.
A lot of the techniques that SEO folks who haven’t gotten up to speed, still focus on are very much still at the word level.
And the word level is not as important anymore.
It is about sentences, paragraphs and context.
With these new models, they’re able to read a sentence and predict what the next sentence logically should be.
They’re able to predict what a sequence of sentences are, and then generate a an understanding of what the next paragraph logically could be about.
Obviously, this technology has a great deal of power when it comes to things like question answering.
So if you were to feed a question to one of these models, they can come up with what is likely to be a solid predicted answer.
Now, what does this mean? This means that these models that are trained on an awful lot of text and tuned, using things like Google Search Quality Rating guidelines, can look at a page and say, how closely does this page, this page has content match? What is predicted, like what should logically be here, if this is a page that starts off with a sentence about coffee shops, and coffee, it should be looking for a discussion of you know, these are the hours that we’re open, these are the drinks that we serve, etc.
And the further your content deviates from what the model understands, the less likely it is that the model is going to think that your page is relevant.
Right? So when we talked about creating content for search engines, people that really is the same thing, because you would not take a page that you were putting on a website and say, a coffee shop and start putting like cooked ham recipes on there.
Right? You would not take a paragraph that was about when your coffee shop was open and insert, you know, random facts about you on there, you shouldn’t, because as a human, you would read that go that makes absolutely no sense.
What, who put this page together, this is terrible.
And so understanding how the natural language processing models work, helps us understand how we should be writing content, we should be creating content for people.
What are the logical things that you as a person would want to know? What are the logical next things that would occur any piece of text? Think about it if you were writing a page about your coffee shop, you’ve got a page of frequently asked questions.
What are the logical questions that should be on there? What time he opened? How much do things cost? How do I dry? How do I get there? You know, how do you treat things for allergens, and by building a page that would satisfy a human who wanted to know all the things that logically you’d want to know about a coffee shop, you’re also satisfying These modern search engines that have the ability to predict text.
And guess essentially, how, how aligned is your page with the things that it is predicting should be on there.
Now, is it perfect? No.
And are there constant updates? Yes.
But the way these algorithms work, there’s not as much drift as long as they have that.
That solid training data set, which again, comes from two things.
It comes from Google Search Quality Rating guidelines where quality raters go out and rate existing pages as to how how well they display things like expertise, authority, and trustworthiness, how, what kind of quality is paid as a high quality page with a lots of original unique content? Is it does it demonstrate expertise, does it demonstrate topical relevance, those are all things that are in the quality weighting guidelines, which I strongly encourage you to read.
Google publishes them.
The other thing that Google has is a group called Pygmalion, which has, I believe, two or three dozen linguistic experts on PhDs in linguistics, who are constantly helping understand the mechanics of language, and how the search queries that are being processed are being essentially are, how correctly they’re processing language.
And ways to deconstruct language, because one of the challenges of content on the web is goes a whole bunch, but one is technically a hot mess.
But something that is authoritative.
Authority varies wildly between different industries and disciplines.
You know, what should be authoritative text, say in surgery is going to be very different than authoritative text in plumbing, right? You’re not going to expect a plumber to write at a graduate school level on their website, but they may have great content that makes them the popular choice for answers about two plumbing questions.
So a lot of the these models on the back end have to contend with these different topics.
So what do you do with this? Well, number one, you create content for people.
And the best way to create content for people that’s relevant is to actually ask people, what questions do you have? Right? Go into your customer service inbox, and just fish out every single question.
And for any question that occurs more than like, two or three times, guess what, you should have content on the website about that? Because it’s clearly a question people have, that they want to know more about.
Go into your, your meeting notes with all your customers and all your clients, and pull out the questions that they keep asking that people keep asking over and over again, again, that’s content that should be on your website.
And when you create stuff that is aligned with people, and reflects reality, you’re going to do better in search engines, because that’s what Google is, is bringing in from all these other websites that are out there.
And that these models will come to essentially a consensus like this is authoritative, based on the training guidelines, based on the expert models, and based on the user behavior signals that Google can see, when people search on search engines, you know, if this site reads like it should be authoritative, but everybody who clicks with clicks back to Google in two seconds, like, yeah, something’s wrong there.
As opposed to, you know, people click through to a website, and they stay there for 30 minutes.
Okay, they, they must have found what they were looking for.
And those data points go in as well.
So that’s how to create content in the modern era focus on what people want, what they need, what questions they have, anticipate the questions they’re going to have in advance and put that content up there.
Do a show like this? I mean, that this this whole show that I’ve been doing for three years now is all about answering questions that real people have, because it’s how search engines function.
If you got follow up questions, leave them in the comments box below.
Subscribe to the YouTube channel on the newsletter, I’ll talk to you soon take care.
One help solving your company’s data analytics and digital marketing problems.
This is Trust insights.ai today and let us know how we can help you
You might also enjoy:
- Almost Timely News, 17 October 2021: Content Creation Hacks, Vanity Metrics, NFTs
- Is Social Listening Useful?
- B2B Email Marketers: Stop Blocking Personal Emails
- Retiring Old Email Marketing Strategies
- How I Think About NFTs
Want to read more like this from Christopher Penn? Get updates here:
Get your copy of AI For Marketers