Determine the semantic core. Semantic core! Engine in blog promotion! Google Search Query Statistics


The semantic core (SC) is a set keywords, queries, search phrases for which you need to promote your site so that targeted visitors from search engines come to it. Compiling a semantic core is the second stage after setting up your site on hosting. It is on a well-written NL that determines how much quality traffic will be on your website.

The need to compile a semantic core lies in several points.

Firstly, it makes it possible to develop a more effective search engine promotion strategy, since the webmaster, who will create the semantic core for his project, will have a clear idea of ​​what search engine promotion methods he will need to apply to his site and decide on the cost of search promotion, which greatly depends on the level of competition key phrases in search results.

Secondly, the semantic core makes it possible to fill the resource with higher quality content, that is, content that will be well optimized for key queries. For example, if you want to make a website about suspended ceilings, then the semantic core of queries should be selected based on this keyword.

In addition, compiling a semantic core involves determining the frequency of use of keywords by search engine users, which makes it possible to determine the search system in which it is necessary to pay more attention to the promotion of a particular query.

It is also necessary to note some rules for compiling a semantic core that must be followed in order to be able to compose a high-quality and effective SL.

Thus, the semantic core must include all possible key queries for which the site can be promoted, except for those queries that cannot bring at least a small amount of traffic to the resource. Therefore, the semantic core should include high-frequency keywords (HF), medium-frequency (MF) and low-frequency (LF) queries.

Conventionally, these requests can be broken down as follows: LF - up to 1,000 requests per month, MF - from 1,000 to 10,000 requests per month, HF - more than 10,000 requests per month. It is also necessary to take into account the fact that in different topics these figures can vary significantly (in some particularly narrow topics the maximum frequency of requests does not exceed 100 requests per month).

Low-frequency queries simply need to be included in the structure of the semantic core, because any young resource has the opportunity to advance in search engines, first of all, using them (and without any external promotion site - you wrote an article for a low-frequency query, the search engine indexed it, and in a few weeks your site can be in the TOP).

It is also necessary to follow the rule of structuring the semantic core, which is that all keywords of the site must be combined into groups not only by their frequency, but also by the degree of similarity. This will make it possible to optimize content more efficiently, for example, optimize some texts for several queries.

It is also advisable to include such types of queries as synonymous queries, slang queries, abbreviated words, and so on into the structure of the semantic core of many sites.

Here clear example how you can increase site traffic just by writing optimized articles that include mid-range and a large number of low-frequency queries:

1. Men's-themed site. Although it is updated very rarely (for all time, a little more than 25 articles have been published), thanks to well-chosen queries it is gaining more and more more traffic. No links were purchased to the site at all!

On this moment the growth in traffic stalled, as only 3 articles were published.

2. Women's themed website. At first, unoptimized articles were published on it. In the summer, a semantic core was compiled, which consisted of queries with very low competition (I will tell you how to collect such queries below). In response to these requests, relevant articles were written and several links were purchased on thematic resources. You can see the results of such promotion below:

3. Medical website. A medical niche with more or less adequate competition was chosen and more than 200 articles were published for tasty queries with good bids (on the Internet, the term “bid” refers to the cost of one user click on an advertisement link). I started buying links in February, when the site was 1.5 months old. So far, most traffic comes from Google; Yandex has not yet taken into account purchased links.

As you can see, the selection of the semantic core plays a vital role in website promotion. Even with a small budget or no budget at all, you can create a traffic website that will bring you profit.

Selection of search queries

You can select queries for the semantic core in the following ways:

Using free search engine statistics services Google queries, Yandex or Rambler

Using special software or online service(free or paid).

Analyzing competitors' websites.

Selection of search queries using search query statistics services Google, Yandex or Rambler

I want to warn you right away that these services do not show entirely accurate information about the number of search queries per month.

For example, you requested “ plastic windows» view 4 pages of search results, then the statistics service will show you that this query was searched not 1, but 4 times, since not only the first page of the search results is considered as an impression, but also all subsequent ones that the user viewed. Therefore, in practice, the real numbers will be slightly lower than those shown by various services.

To determine the exact number of transitions, it is best, of course, to look at the traffic statistics of competitors who are in the TOP 10 (liveinternet or mail.ru, if it is open). This way, you can understand how much traffic the request you are interested in will bring.

You can also roughly calculate how many visitors a given request will bring you, depending on the position the site will occupy in search results. Here's how the CTR (click-through rate) of your site will change at different positions in the TOP:

Let’s take, for example, the query “!repair!apartment”, region “Moscow and region”:

In this case, if your site occupies 5th position in the search results, this request will bring you about 75 visitors per month (6%), 4th place (8%) - 100 visitors per month, 3rd place (11%) - 135 villages/month, 2nd place (18%) - 223 villages/month. and 1st position (28%) - 350 visitors per month.

You can also influence CTR using a bright snippet, thereby increasing traffic for this request. You can read how to improve a snippet and what it is.

Google Search Query Statistics

Previously, I used Google search query statistics more often, since I primarily promoted this search engine. Then it was enough to optimize the article, purchase as many different PR links for it as possible (it is very important that these links were natural and that visitors followed the links) and voila - you are in the TOP!

Now the situation in Google is such that it is not much easier to advance in it than in Yandex. Therefore, you have to pay much more attention to both writing and designing (!) articles and buying links for the site.

I would also like to draw your attention to the following fact:

In Russia, Yandex is the most popular (Yandex - 57.7%, Google - 30.6%, Mail.ru - 8.9%, Rambler -1.5%, Nigma/Ask.com - 0.4%), so if you are promoting in this country, you should first It's worth focusing on Yandex.

In Ukraine, the situation looks different: Google - 70-80%, Yandex - 20-25%. Therefore, Ukrainian webmasters should focus on promotion in Google.

To use Google query statistics, go to

Let's look at an example of selecting keywords for a culinary site.

First of all, you need to enter a basic query, on the basis of which keyword options for the future semantic core of the site will be selected. I entered the query “how to cook”.

The next step is to select the match type. There are 3 types of matches: broad, phrase and exact. I recommend selecting accurate as this option will show the most accurate information for the request. And now I will explain why.

Broad match means that impression statistics will be shown for all words that are in this query. For example, for the query “plastic windows” it will be shown for all words that include the word “plastic” and the word “windows” (plastic, window, buy windows, buy blinds for windows, PVC windows prices). In a word, there will be a lot of “garbage”.

Phrase matching means that statistics will be shown for words in the exact order in which they are listed. Along with the specified combination of words, other words may also be present in the request. For the request “plastic windows” the words “inexpensive plastic windows”, “plastic windows Moscow”, “how much do plastic windows cost”, etc. will be taken into account.

We are most interested in the indicator “Number of requests per month (target regions)” and “Approximate cost per click” (if we are going to place Adsense advertising on the site pages).

Yandex search query statistics

I use search query statistics almost every day, as they are presented in a more convenient form than their counterpart in Google.

The only disadvantage of Wordstat is that you won’t find match types in it, you won’t save selected queries to your computer, you won’t be able to find out the cost of a click, etc.

To obtain more accurate results, you need to use special operators with which you can clarify the queries that interest us. You can find a list of operators here.

If you simply enter the query “how to cook” in Wordstat, we get the following statistics:

This is the same as if we selected “broad match” in Adwords.

If you enter the same query, but in quotes, we will get more accurate statistics (analogous to phrase matching in Adwords):

Well, to get statistics only for a given request, you need to use the “!” operator: “!how!to prepare”

To get even more accurate results, you need to specify the region for which the site is being promoted:

Also in the top panel of Wordstat there are tools with which you can view the statistics of a given query by region, by month and by week. With the help of the latter, by the way, it is very convenient to analyze the statistics of seasonal requests.

For example, after analyzing the request “how to cook”, you can find out that it is most popular in the December months (this is not surprising - everyone is preparing for the New Year):

Rambler search query statistics

I would like to warn you right away that the statistics of requests from Rambler are losing their relevance every year more and more (primarily, this is due to the low popularity of this search engine). Therefore, you most likely won't even have to work with it.

There is no need to enter any operators into Adstat - it immediately displays the frequency of the request in the case in which it was entered. It also has separate query frequency statistics for the first page of search results and for all search results pages, including the first one.

Selection of search queries using special software or online services

Rookee can not only promote your queries, but can also help in creating the semantic core of the site.

With Rookee, you can easily select a semantic core for your website and be able to approximately predict the number of visits for selected queries and the cost of promoting them to the top.

Selection of queries using the free Slovoeb program

If you are going to compose the semantic core (semantic core) at a more professional level, or you need query statistics for Google.Adwords, Rambler.Adstat, social network VKontakte, various link aggregators, etc., I advise you to immediately purchase Key Collector.

If you want to create a large semantic core, but do not want to spend money on purchasing paid programs, the best option in this case would be the Slovoeb program (read information about the program here). It is the “younger brother” of Kay Collector and allows you to collect EA based on query statistics on Yandex.Wordstat.

Installing the SlovoEB program.

Download the archive with the program from .

Make sure the archive is unlocked. To do this, in the file properties (in context menu select “Properties”) click the “Unblock” / “Unblock” button, if present.

Unzip the contents of the archive.

Run executable file Slovoeb.exe

We create new project:

Select the desired region (Yandex.Wordstat Regions button):

Save the changes.

Click on the button “Left column of Yandex.Wordstat”

If necessary, we set “Stop words” (words that should not be included in our semantic core). Stop words can be the following words: “free” (if you sell something on your website), “forum”, “Wikipedia” (if you have your own information site that does not have a forum), “porn”, “sex” (well, everything is clear here), etc.

Now you need to set the initial list of words on the basis of which the SL will be compiled. Let's form the core for a company that installs suspended ceilings (in Moscow).

When selecting any semantic core, the first step is to make a classification of the analyzed topic.

In our case, suspended ceilings can be classified according to the following criteria (I make similar convenient mind maps in the MindJet MindManager program):

Helpful advice: for some reason, many webmasters forget to include the names of small localities in the semantic core.

In our case, the names of areas of interest to us in Moscow and cities in the Moscow region could be included in the NL. Even if there are very few requests per month for these keywords (“Stretch ceilings Golitsyno”, “Stretch ceilings Aprelevka”, etc.), you still need to write at least one short article for them, the title of which would include the required key. You don’t even have to promote such articles, because most often there will be very little competition for these requests.

10-20 such articles, and your site will consistently have several additional orders from these cities.

Click the “Left Column of Yandex.Wordstat” button and enter the necessary queries:

Click on the “Parse” button. As a result, we get the following list of requests:

We filter out all unnecessary requests that do not fit the theme of the site (for example, “do-it-yourself suspended ceilings” - although this request will bring some traffic, it will not bring us clients who will order the installation of ceilings). We select these queries and delete them so as not to waste time analyzing them in the future.

Now you need to clarify the frequency for each of the keys. Click on the “Collect frequencies “!” button:

Now we have the following information: the request, its general and exact frequency.

Now, based on the received frequencies, you need to review all requests again and delete unnecessary ones.

Unnecessary queries are queries that:

The exact frequency (“!”) is very low (in the topic I have chosen, in principle, you need to value every visitor, so I will filter out requests that have a monthly frequency of less than 10). If the topic was not construction, but, say, some general thematic one, then you can safely filter out requests whose frequency is below 50-100 per month.

The ratio of total and exact frequencies exceeds very high. For example, the request “will buy suspended ceilings” (1335/5) can be immediately deleted, because it is a “dummy request”.

Requests with very high competition should also be deleted; it will be difficult to advance on them (especially if you have a young site and a small budget for promotion). Such a request, in our case, is “stretch ceilings”. In addition, most often, those queries that consist of 3,4 and more words- they bring more targeted visitors.

In addition to Slovoeb, there is another excellent program for convenient automatic collection, analysis and processing of statistics on impressions of Yandex.Direct keywords -.

News search queries

In this case, the FL is created differently than for a regular content project. For a news site, first of all, you need to select the categories under which the news will be published:

After this, you need to select queries for each section. News requests can be expected or unexpected.

For the first type of requests, the informational occasion is predictable. Most often, you can even accurately determine the date when there will be a surge in the popularity of a particular request. It could be some holiday ( New Year, May 9, March 8, February 23, Independence Day, Constitution Day, church holidays), events (musical events, concerts, film premieres, sports competitions, presidential elections).

You can prepare such requests in advance and determine the approximate volume of traffic in this case using

Also, do not forget to look at the traffic statistics of your site (if you have already reviewed some event) and competitors’ sites (if their statistics are open).

The second type of request is less predictable. These include breaking news: disasters, catastrophes, any events in families famous people(birth/wedding/death), release of an unannounced software product, etc. In this case, you just need to be one of the first to publish this news.

To do this, it is not enough to simply monitor news in Google and Yandex - in this case, your site will simply be one of those that simply reprinted this event. A more effective method that allows you to hit the big jackpot is to monitor foreign sites. By publishing this news as one of the first on the RuNet, you, in addition to tons of traffic that will put the servers on your hosting, will receive a huge amount backlinks to your website.

Movie-related queries can also be classified as the first type (expected queries). Indeed, in this case, the date of the premiere of the film is known in advance, the script of the film and its actors are approximately known. Therefore, you need to prepare in advance the page on which the film will appear, and temporarily add its trailer there. You can also publish news about the film and its actors on the site. This tactic will help you take TOP positions in search engines in advance and will bring visitors to the site even before its premiere.

Do you want to find out what queries are currently trending or predict the relevance of your topic in the future? Then use services that provide information about trends. It allows you to perform many analysis operations: comparing search trends for several queries, analyzing the geographic regions of a query, viewing the hottest trends at the moment, viewing relevant current queries, exporting results to CSV format, the ability to subscribe to an RSS feed for hot trends, etc. .

How to speed up the collection of the semantic core?

I think that everyone who came across the collection of a semantic core had the thought: “How long and tedious the parsing is, I’m tired of sorting through and grouping thousands of these queries!” This is fine. This happens to me too sometimes. Especially when you have to parse and sort through a syntax that consists of several tens of thousands of requests.

Before you start parsing, I strongly advise you to divide all requests into groups. For example, if your website theme is “Building a house,” break it down into the foundation, walls, windows, doors, roof, wiring, heating, etc. Then it will simply be much easier for you to sort through and group requests when they are located in a small group and are related to each other by a specific narrow topic. If you just parse everything in one pile, you will end up with an unrealistically huge list that will take more than one day to process. And so, by processing the entire list of keywords in small steps, you will not only process all requests more efficiently, but you will also be able to simultaneously order articles from copywriters for already collected keys.

The process of collecting a semantic core for me almost always begins with automatic parsing of queries (for this I use Key Collector). You can also collect it manually, but if we work with a large number of requests, I don’t see the point in wasting my precious time on this routine work.

If you work with other programs, then they will most likely have a function for working with proxy servers. This allows you to speed up the parsing process and protect your IP from being banned by search engines. To be honest, it’s not very pleasant when you urgently need to complete an order, and your IP is banned for a day due to frequent access to the Google/Yandex statistics service. This is where paid proxies come to the rescue.

Personally, I don’t use them at the moment for one simple reason - they are constantly banned, finding high-quality working proxies is not so easy, and I don’t want to pay money for them again. So I found alternative way collection of nuclear weapons, which accelerated this process several times.

Otherwise, you will have to look for other sites for analysis.

In most cases, sites hide these statistics, so you can use entry point statistics.

As a result, we will have statistics of the most visited pages of the site. We go to them, write out the main queries for which it was optimized, and based on them we collect the SYNOPSIS.

By the way, if the site has open statistics “By search phrases”, you can make your work easier and collect queries using Key Collector (you just need to enter the site address and click on the “Get data” button):

The second way to analyze a competitor’s website is by analyzing their website.

Some resources have a “Most Popular Articles” widget (or something like that). Sometimes the most popular articles are chosen based on the number of comments, sometimes based on the number of views.

In any case, having before your eyes a list of the most popular articles, you can figure out what requests this or that article was written for.

The third way is to use tools. It was actually created to analyze the trustworthiness of sites, but, to be honest, it considers trustworthiness very bad. But what he can do well is analyze requests from competitors’ websites.

Enter the address of any site (even with closed statistics!) and click the trust check button. At the very bottom of the page, site visibility statistics by keywords will be displayed:

Website visibility in search engines (by keywords)

The only negative is that these queries cannot be exported; everything must be copied manually.

The fourth way is to use services and .

With its help, you can determine all queries for which the site ranks in the TOP 100 in both search engines. It is possible to export positions and queries to xls format, but I was never able to open this file on any computer.

Well, the last way to find out your competitors’ keywords is with the help of

Let’s analyze the query “house made of timber” as an example. In the “Competitor Analysis” tab, enter this request (if you want to analyze a specific site, then you just need to enter its url in this field).

As a result, we get information about the frequency of the request, the number of advertisers in each of the search engines (and if there are advertisers, then there is money in this niche) and the average cost of a click in Google and Yandex:

You can also see all ads in Yandex.Direct and Google Adwords:

And this is what the TOP of each PS looks like. You can click on any of the domains and see all the queries for which it is in the TOP and all its contextual ads:

There is another way to analyze competitors’ requests, which few people know about - using the Ukrainian service

I would call it the “Ukrainian version of Spywords”, since they are very similar in functionality. The only difference is that the database contains key phrases that Ukrainian users are looking for. So if you work in UA-net, this service will be very useful for you!

Request competition analysis

So, the requests have been collected. Now you need to analyze the competition for each of them in order to understand how difficult it will be to promote this or that keyword.

Personally, when I create a new website, I first of all try to write articles for queries that have low competition. This will allow for short term and with minimal investment bring the site to pretty good traffic. At least in such areas as construction and medicine, it is possible to reach 500-1000 visitors in 2.5 months. I’m generally silent about women’s issues.

Let's see how you can analyze competition manually and automatically.

Manual method of competition analysis

We enter the desired keyword in the search and look at the T0P-10 (and if necessary, T0P-20) of sites that are in the search results.

The most basic parameters you need to look at are:

The number of main pages in the TOP (if you are promoting the internal page of the site, and your competitors mainly have main pages in the TOP, then most likely you will not be able to overtake them);

The number of direct occurrences of the keyword in the title of the page.

For queries such as “website promotion”, “how to lose weight”, “how to build a house” there is unrealistically high competition (there are many main pages in the TOP with a direct inclusion of the keyword in the Title), so it is not worth promoting on them. But if you search for the query “how to build a house from foam blocks with your own hands with a basement,” then you will have a better chance of getting into the TOP. Therefore, I once again focus my attention on the fact that you need to promote queries that consist of 3 or more words.

Also, if you analyze search results in Yandex, you can pay attention to the TCI of sites (the higher it is, the harder it will be to overtake them, because a high TCI most often indicates a greater reference mass site), whether they are in the Yandex Catalog (sites with the Yandex Catalog have more trust), its age (search engines like older sites more).

If you analyze the TOP sites in Google, pay attention to the same parameters that I wrote about above, only in this case, instead of the Yandex Catalog there will be DMOZ catalog, and instead of the TCI indicator there will be a PR indicator (if the TOP site pages have PR from 3 to 10, it will not be easy to overtake them).

I recommend analyzing sites using the plugin. It shows all the information about the competitor's website:

An automatic way to analyze request competition

If there are a lot of requests, then you can use programs that will do all the work for you hundreds of times faster. In this case, you can use Sloboeb or Key Collector.

Previously, when analyzing competitors, I used the “KEI” indicator (competition indicator). This function is available in Key Collector and Slovoyobe.

In Slovoeb, the KEI indicator simply shows the total number of sites in the search results for a particular request.

In this regard, Key Collector has an advantage, since it has the ability to independently set a formula to calculate the KEI parameter. To do this, go to Program Settings – KEI:

In the “Formula for calculating KEI 1” insert:

(KEI_YandexMainPagesCount * KEI_YandexMainPagesCount * KEI_YandexMainPagesCount) + (KEI_YandexTitlesCount * KEI_YandexTitlesCount * KEI_YandexTitlesCount)

In the “Formula for calculating KEI 2” insert:

(KEI_GoogleMainPagesCount * KEI_GoogleMainPagesCount * KEI_GoogleMainPagesCount) + (KEI_GoogleTitlesCount * KEI_GoogleTitlesCount * KEI_GoogleTitlesCount)

This formula takes into account the number of main pages in the search results for a given keyword and the number of pages in the TOP 10 in which this key phrase is included in the title of the page. Thus, you can obtain more or less objective competition data for each request.

In this case, the smaller the request KEI, the better. The best keywords will be with KEI=0 (if they have at least some traffic, of course).

Click on the data collection buttons for Yandex and Google:

Then click on this button:

In the KEI 1 and KEI 2 column, data on KEI queries for Yandex and Google will appear
respectively. Let's sort the queries in ascending order of the KEI 1 column:

As you can see, among the selected queries there are some for which promotion should not pose any special problems. In less competitive topics, to bring a similar request to the TOP 10, you just need to write a good optimized article. And you won't need to buy external links to promote it!

As I said above, I have used the KEI indicator before. Now, to assess the competition, I just need to get the number of main pages in the TOP and the number of occurrences of the keyword in the Title of the page. Key Collector has a similar function:

After that, I sort the requests by the column “Number of main pages in Yandex PS” and make sure that according to this parameter there are no more than 2 main pages in the TOP and as few occurrences in the headings as possible. After the NA has been compiled for all these requests, I reduce the filter parameters. Thus, articles under NK requests will be published first on the site, and articles under SC and VK requests will be published last.

After all the most interesting queries have been collected and grouped (I’ll talk about grouping below), click on the “Export data” button and save them to a file. In the export file I usually include following parameters: frequency according to Yandex with “!” in a given region, the number of main pages of sites and the number of occurrences of the keyword in the headings.

Tip: Kay Collector sometimes does not quite correctly display the number of occurrences of requests in the headers. Therefore, it is advisable to additionally enter these queries in Yandex and look at the results manually.

You can also evaluate the competitiveness of search queries using a free

Grouping requests

After all the requests have been selected, the time comes for the rather boring and monotonous work of grouping requests. You need to select similar queries that can be combined into one small group and promoted within one page.

For example, if your list contains the following queries: “how to learn to do push-ups”, “how to learn to do push-ups at home”, “how to teach a girl to do push-ups”. Similar requests can be combined into one group and one large optimized article can be written for it.

To speed up the grouping process, parse keywords in parts (for example, if you have a site about fitness, then during parsing, break keywords into groups that will include queries related to the neck, arms, back, chest, abs, legs and etc.). This will make your work much easier!

If the received groups contain a small number of requests, then you can stop there. And when you end up with a list of several dozen, or even hundreds of queries, you can try the following methods.

Working with stop words

Key Collector has the ability to specify stop words that can be used to mark unwanted keywords in the resulting query table. Such queries are usually removed from the semantic core.

In addition to removing unwanted queries, this function can also be used to search for all the necessary word forms by a given key.

Specify the required key:

In the table with queries, all word forms of the specified key will be highlighted:

We transfer them to new tab and there we manually work with all requests.

Filtering for the "Phrase" field

You can find all word forms of a given keyword using the filtering settings for the “Phrase” field.

Let's try to find all queries that include the word “bars”:

As a result we get:

Group Analysis Tool

This is the most handy tool for grouping phrases and further manual filtering.

Go to the “Data” - “Group Analysis” tab:

And this is what is displayed if you open any of these groups:

By marking any group of phrases in this window, you simultaneously mark or unmark the phrase in the main table with all queries.

In any case, without self made you can't get by. But, thanks to Kay Collector, part of the work (and not a small one!) has already been done, and this, at least a little, simplifies the process of compiling the website’s SYNOPSIS.

Once you've manually processed all the requests, you should end up with something like this:

How to find a low-competition, profitable niche?

First of all, you must decide for yourself how you will make money on your website. Many novice webmasters make the same stupid mistake - they first create websites on a topic that they like (for example, after buying an iPhone or iPad, everyone immediately runs to make another “Apple-themed” website), and then they begin to understand that The competition in this niche is very high and it will be almost impossible for their shit site to get to the top. Or they make a culinary website because they like to cook, and then realize with horror that they don’t know how to monetize such traffic.

Before creating any website, immediately decide how you will monetize the project. If you have an entertainment theme, then teaser advertising is most likely suitable for you. For commercial niches (in which something is sold), contextual and banner advertising are perfect.

I would like to immediately dissuade you from creating websites on general topics. Sites that write “everything about everything” are now not so easy to promote, because the TOPs have long been occupied by well-known trust portals. It will be more cost-effective to create a narrowly themed website that will outstrip all competitors in a short time and gain a permanent position in the TOP. I'm telling you this from personal experience.

My narrowly themed medical website already 4 months after its creation ranks first in Google and Yandex, having overtaken the largest medical general portals.

Promoting narrowly themed sites is EASIER and FASTER!

Let's move on to choosing a commercial niche for the site on which we will make money contextual advertising. At this stage, you need to adhere to 3 criteria:

When assessing competition in a certain niche, I use a search in the Moscow region, where competition is usually higher. To do this, specify “Moscow” in the region settings in your Yandex account:

The exception is cases when the site is made for a specific region - then you need to look at competitors specifically for this region.

The main signs of low competition in search results are as follows:

Lack of complete answers to the query (irrelevant results).

There are no more than 2-3 main pages in the TOP (they are also called “faces”). A larger number of “faces” means that the entire site is purposefully promoted for this request. In this case, it will be much more difficult to promote the internal page of your site.

Inaccurate phrases in the snippet indicate that the required information is simply not available on competitors’ websites or that their pages are not optimized for this request at all.

A small number of large thematic portals. If, for many requests from your future network, there are many large portals and highly specialized sites in the TOP, you need to understand that the niche has already been formed, and the competition there is very high.

More than 10,000 requests per month based on base frequency. This criterion means that you should not “narrow” the subject of the site too much. The niche must have a sufficient amount of traffic from which you can make money in context. Therefore, the main request in the selected topic must have at least 10,000 requests per month according to Wordstat (without quotes and taking into account the region!). Otherwise, you will need to expand the topic a little.

You can also roughly estimate the amount of traffic in a niche using statistics on traffic to sites that occupy first positions in the TOP. If these statistics are closed for them, then use

Most seasoned MFA professionals look for niches for their sites not in this way, but using the Micro Niche Finder service or program.

The latter, by the way, shows the SOC parameter (0-100 - the request will reach the TOP only on internal optimization, 100-2000 - average competition). You can safely select queries with SOC less than 300.

In the screenshot below you can see not only the frequency of queries, but also the average cost per click and site position for queries:

There is also a useful feature called “Potential Ad Buyers”:

You can also simply enter the query you are interested in and analyze it and similar queries:

As you can see, the whole picture is in front of you, all that remains is to check the competition for each of the requests and select the most interesting of them.

We estimate the cost of bids for YAN sites

Now let's look at how to analyze the cost of bids for YAN sites.

Let's enter the queries that interest us:

As a result we get:

Attention! Let's look at the rate of guaranteed impressions! The actual approximate cost per click will be 4 times less than that shown by the service.

It turns out that if you display a page about the treatment of sinusitis in the TOP and visitors click on the advertisement, then the cost of 1 click on the YAN advertisement will be 12 rubles (1.55/4 = $0.39).

If you select the “Moscow and region” region, the bids will be even higher.

That is why it is so important to take first places in the TOP in this particular region.

Please note that when analyzing queries, you do not need to take into account commercial queries (for example, “buy a table”). For content projects it is necessary to analyze and promote information requests (“how to make a table with your own hands”, “how to choose a table for the office”, “where to buy a table”, etc.).

Where to look for queries to select a niche?

1. Start brainstorming and write down all your interests.

2. Write what inspires you in life (it could be some kind of dream, or relationships, lifestyle, etc.).

3. Look around and write down all the things that surround you (literally write everything in a notebook: a pen, a light bulb, wallpaper, a sofa, an armchair, paintings).

4. What is your occupation? Maybe you work in a factory behind a machine or in a hospital as an ENT doctor? Write down all the things you use at work in a notebook.

5. What are you good at? Write this down too.

6. Look through advertisements in newspapers, magazines, advertisements on websites, and spam emails. Perhaps they offer something interesting that you could make a website about.

Find sites there with 1000-3000 visitors per day. Look at the topics of these sites and write down the most frequent queries that bring them traffic. Perhaps there is something interesting there.

I will give a selection of one narrowly thematic niche.

So, I became interested in such a disease as “heel spur”. This is a very narrow micro-niche, for which you can make a small website of 30 pages. Ideally, I advise you to look for broader niches for which you could make a website of 100 or more pages.

Open and check the frequency of this request (we won’t specify anything in the region settings!):

Great! The frequency of the main request is more than 10,000. There is traffic in this niche.

Now I want to check the competition for queries for which the site will be promoted (“heel spur treatment”, “heel spur treatment”, “heel spur how to treat”)

I switch to Yandex, specify the “Moscow” region in the settings and this is what I get:

In the TOP there are many internal pages of general thematic medical sites, which will not be difficult for us to overtake (although this will happen no earlier than in 3-4 months) and there is also one site that is tailored to the topic of heel spur treatment.

The wnopa website is a 10-page shit website (which, however, has PR=3) with a traffic of 500 visitors. A similar site can also be overtaken if you make a better site and promote the main page for this request.

Having analyzed the TOP for each of the requests in this way, we can come to the conclusion that this microniche has not yet been occupied.

Now there is the last stage - checking the bids.

We go to and enter our requests there (there is no need to select a region, since our site will be visited by Russian-speaking people from all over the world)

Now I find the arithmetic average of all guaranteed impressions indicators. It turns out to be $0.812. We divide this indicator by 4, which gives an average of $0.2 (6 rubles) per click. For medical topics, this is, in principle, a normal indicator, but if you wish, you can find topics with better bids.

Is it worth choosing a similar niche for a website? It's up to you to decide. I would look for broader niches that have more traffic and better cost per click.

If you really want to find a good theme, I highly recommend you follow the next step!

Make a table with the following columns (I indicated the data approximately):

In this way, you must write out at least 25 requests per day! Spend 2 to 5 days searching for a niche and you will eventually have plenty to choose from!

Choosing a niche is one of the most important steps in creating a website and if you treat it negligently, you can assume that in the future you will simply throw away your time and money.

How to write the right optimized article

When I write an article optimized for any request, I first of all think about making it useful and interesting to site visitors. Plus, I try to format the text so that it is pleasant to read. In my understanding, an ideally formatted text should contain:

1. A bright, eye-catching headline. It is the title that first attracts the user when he views the search engine results page. Please note that the title must be completely relevant to what is written in the text!

2. Lack of “water” in the text. I think that few people like it when something that could be written in 2 words is presented on several sheets of text.

3. Simplify paragraphs and sentences. To make the text easier to read, be sure to break large sentences into short ones that fit on one line. Simplify complex turns. It’s not for nothing that in all newspapers and magazines the text is published in small sentences in columns. Thus, the reader will perceive the text better, and his eyes will not get tired so quickly.

Paragraphs in the text should also not be large. A text that alternates between small and medium-sized paragraphs will be best perceived. An ideal paragraph would be 4-5 lines.

4. Use subheadings (h2, h3, h4) in the text. It takes a person 2 seconds to “scan” open page and decide whether to read it or not. During this time, he runs his eyes across the page and identifies for himself the most important elements that help him understand what the text is about. Subheadings are precisely the tool with which you can grab the visitor’s attention.

To make the text structured and easier to understand, use one subheading for every 2-3 paragraphs.

Please note that these tags must be used in order (h2 to h4). Eg:

You don't need to do this:

<И3>Subtitle<И3>

<И3>Subtitle<И3>

<И2>Subtitle<И2>

Here's how you can do it:

<И2>Subtitle<И2>

<И3>Subtitle<И3>

<И4>Subtitle<И4>

You can also do this as follows:

<И2>How to pump up your abs<И2>

<И3>How to pump up your abs at home<И3> <И3>How to pump up your abs on the horizontal bar<И3> <И2>How to pump up your biceps<И2> <И3>How to pump up biceps without exercise equipment<И3>

Keywords should be used to a minimum in the texts (no more than 2-3 times, depending on the size of the article). The text should be as natural as possible!

Forget about putting keywords in bold and italics. This is a clear sign of over-optimization and Yandex really doesn’t like it. In this way, highlight only important sentences, thoughts, ideas (at the same time, try not to include keywords in them).

5. Use bulleted and numbered lists, quotes in all your texts (do not confuse them with quotes and aphorisms of famous people! We only include important thoughts, ideas, tips), pictures (with the alt, title and signature attributes filled in). Insert interesting thematic videos into articles, this has a very good effect on the length of time visitors stay on the site.

6. If the article is very large, then it makes sense to navigate through the text using content and anchor links.

After you have made all the necessary changes to the text, run your eyes over it again. Isn't it hard to read? Maybe some proposals need to be made even smaller? Or do you need to change the font to a more readable one?

Only after you are satisfied with all the changes in the text, you can fill out the title, description and keywords and send it for publication.

Title requirements

Title is the title in the snippet that is displayed in search results. The title in the title and the title in the article (the one that is highlighted with the tag

) must be different from each other and at the same time, they must contain the main keyword (or several keywords).

Description requirements

The description should be vivid and interesting. It needs to briefly convey the meaning of the article and do not forget to mention the necessary keywords.

Keywords requirements

Enter the promoted keywords here separated by a space. No more than 10 pieces.

So, this is how I write optimized articles.

First of all, I select the queries that I will use in the article. For example, let’s take the main request “how to learn to do push-ups.” Enter it into wordstat.yandex.ru


Wordstat displayed 14 queries. I will try, if possible, to use most of them in the article (it is enough to mention each keyword once). Keywords can and should be declined, rearranged, replaced with synonyms, and made as varied as possible. Search engines that understand Russian (especially Yandex) perfectly recognize synonyms and take them into account when ranking.

I try to determine the size of an article by the average size of articles from the TOP.

For example, if in the TOP 10 there are 7 sites in the search results, the text size of which is mainly 4,000-7,000 characters, and 3 sites in which this indicator is very low (or high), I simply do not take into account the last 3 attention.

Here is the approximate html structure of an article optimized for the query “how to learn to do push-ups”:

How can a girl learn to do push-ups and parallel bars from scratch?


……………………………

How can a girl learn to do 100 or more push-ups in 2 months?


……………………………

What should proper nutrition be like?


……………………………

Daily regime


……………………………
……………………………

Training program

Some optimizers advise that you must include the main key in the first paragraph. In principle, this is an effective method, but if you have 100 articles on your site, each of which contains a promoted query in the first paragraph, this can have unpleasant consequences.

In any case, I advise you not to get hung up on one specific scheme, experiment, experiment and experiment again. And remember, if you want to make a site that search robots will love, make sure that real people love it!

How to increase website traffic using a simple manipulation of linking?

I think that each of you has seen many different linking schemes on sites, and I am more than sure that you did not attach much importance to these schemes. But in vain. Let's look at the first linking scheme, which is found on most sites. True, it still lacks links from internal pages to the main page, but this is not so important. The important thing is that in this case, on the site, the pages of categories (which most often are not promoted and are closed from indexing in robots.txt!) and the main page of the site will have the greatest weight. Pages with posts in this case have the least weight:

Now let's think logically: optimizers who promote content projects promote which pages first? That's right - pages with posts (in our case, these are 3rd level pages). So why, promoting the site external factors(links), we so often forget about equally important internal factors (correct linking)?

Now let's look at a diagram in which inner pages will have more weight:

This can be achieved simply by properly setting the weight flow from the pages top level, to the pages of the lower, 3rd level. Also, do not forget that internal pages will be additionally boosted by external links to other sites.

How to set up this kind of weight flow? Very simple - using tags nofollow noindex.

Let's look at a site where you need to upgrade internal pages:

On the main page (1st level): all internal links are open for indexing, external links are closed via nofollow noindex, the link to the site map is open.

On pages with categories (2nd level): links to pages with posts (3rd level) and to the site map are open for indexing; all other links (to the main page, categories, tags, external) are closed via nofollow noindex.

On pages with posts (3rd level): links to pages of the same 3rd level and to the site map are open, all other links (to the main page, categories, tags, external ones) are closed using nofollow noindex.

After such linking was done, the site map received the greatest weight, which evenly distributed it across all internal pages. Internal pages, in turn, gained significant weight. The main page and the category page received the least weight, which is exactly what I wanted to achieve.

I will check the weight of the pages using the program. It helps to visualize the distribution static weight on the site, focus it on promoted pages, and thereby raise your site in search engine results, significantly saving your budget on external links.

Well, now, let's move on to the most interesting thing - the results of such an experiment.

Women's themed website. Changes to the site's weight distribution were made in January of this year. A similar experiment brought +1000 to traffic:

The site is under a Google filter, so the traffic increased only in Yandex:

Do not pay attention to the decline in traffic in May, this is seasonality (May
holidays).

2. The second site dedicated to women, changes were also made in January:

This site, however, is constantly updated and links to it are purchased.

Terms of reference for copywriters for writing articles

Article title: How can greens be useful?

Uniqueness: not lower than 95% by Etxt

Article size: 4,000-6,000 characters without spaces

Keywords:

what are the benefits of greens - the main key

healthy greens

beneficial properties of greens

benefits of greens

Requirements for the article:

The main keyword (listed first) does not have to be in the first paragraph, it just needs to appear in the first half of the text!!!

PROHIBITED: grouping keywords in parts of the text; it is necessary to distribute keywords throughout the text.

Keywords should be highlighted in bold (this is to make it easier for you to check your order)

Keywords need to be entered harmoniously (therefore, you need to use not only direct occurrences of keys, but you can also inflect them, swap words, use synonyms, etc.)

Use subheadings in your text. In subheadings, try not to use direct occurrences of keywords. Make subheadings for each article as varied as possible!

Use no more than 2-3 direct occurrences of the main query in the text!

If in an article you need to use the keywords “how greens are useful” and “how greens are useful for humans,” then by writing the second keyword, you will automatically use the first one. That is, there is no particular need to write a key twice if one already contains an entry of another key.

Before the beginning SEO promotion you need to create a semantic core of the site - a list of search queries that are used when searching for the goods or services we offer potential clients. All further work - internal optimization and work with external factors (purchase of links) are carried out in accordance with the list of requests defined at this stage.

The final cost of promotion and even the expected level of conversion (number of calls to the company) also depend on the correct collection of the core.

The more companies promote using the chosen word, the higher the competition and, accordingly, the cost of promotion.

Also, when choosing a list of queries, you should not only rely on your ideas about what words your potential clients use, but also trust the opinion of professionals, because not everyone is expensive and popular queries have high conversion and promoting some of the words directly related to your business may simply be unprofitable, even if it is possible to achieve the ideal result in the form of TOP-1.

A correctly formed semantic core, all other things being equal, ensures that the site is confidently positioned in the top positions of search results for a wide range of queries.

Principles for compiling semantics

Search queries are formed by people - potential site visitors, based on their goals. Keep up with mathematical methods statistical analysis, embedded in the algorithm of work of robots - search engines, it is difficult, especially since they are continuously refined, improved, and therefore change.

Most effective way cover the maximum number of people when forming the semantic core of the site possible requests- this is to look at it as if from the position of a person making a search request.

The search engine was created to help a person quickly find the most suitable source of information for a search query. The search engine is focused primarily on quick way narrow down to several dozen most suitable answer options for the key phrase (word) of the request.

When forming a list of these keywords, which will be the basis of the site’s semantics, the circle of its potential visitors is actually determined.

Stages of collecting the semantic core:

  • First, a list of the main key phrases and words found in the information field of the site and characterizing its target orientation is compiled. In this case, you can use the latest statistical information about the frequency of requests in the direction in question from the search engine. In addition to the main variants of words and phrases, it is also necessary to write down their synonyms and variants of other names: washing powder - detergent. The Yandex Wordstat service is perfect for this work.

  • You can also write down the components of the name of any product or subject of the request. Very often, queries include words with typos, misspellings, or simply spelled incorrectly due to the lack of literacy of a large part of Internet users. Taking this feature into account can also attract additional resources from site visitors, especially if any new names appear.
  • The most common queries, also called high-frequency queries, rarely lead a person to the site they are looking for. Low-frequency queries, that is, queries with clarification, work better. For example, the request Ring will return one top, and piston ring will provide more specific information. When collecting, it is better to focus on such requests. This will attract target visitors, that is, for example, potential buyers, if this is a commercial site.
  • When compiling a list of keywords, it is also advisable to take into account widespread slang, the so-called folk, which have become generally accepted and stable names for some objects, concepts, services, etc., for example, cell phone - mobile phone– mobile phone – mobile phone. Taking into account such neologisms in some cases can provide a significant increase in the target audience.
  • In general, when compiling a list of keys, it is better to initially focus specifically on target audience, that is, those website visitors for whom the product or service is intended. The core should not contain a little-known name of an item (product, service) as the main option, even if it needs to be promoted. Such words will be found extremely rarely in queries. It is better to use them with clarifications or use more popular similar names or analogues.
  • When the semantics are ready, it should be passed through a series of filters to remove clogging keywords, which means that they bring the wrong target audience to the site.

Taking into account the semantics of associated queries

  • To the initial list of the semantic core, compiled from the main keys, a number of auxiliary low-frequency keys should be added, which may include important but not taken into account words that did not come to mind when compiling it. The search engine itself will help you with this. When you repeatedly type key phrases from the list on a topic, the search engine itself offers for consideration options for frequently occurring phrases in this area.
  • For example, if the phrase “computer repair” is entered, and then the second query is a matrix, then search system will perceive them as associated, that is, interconnected in meaning, and will provide various frequently encountered queries in this direction to help. With such key phrases you can expand the original semantics.
  • Knowing a few main words from the core, with the help of a search engine it can be significantly expanded with associated phrases. In the event that a search engine does not produce an insufficient number of such additional keys, you can obtain them using the methods of a thesaurus - a set of concepts (terms) for a specific subject from the same conceptual area. Dictionaries and reference books can help here.

Logical scheme for selecting semantics for a site

Formation of a list of requests and their final editing

  • The key phrases that make up the semantics generated in the first two steps require filtering. Among such phrases there may be useless ones, which will only make the core heavier, without bringing any tangible benefit in attracting the target audience of site visitors. Phrases obtained by analyzing the target orientation of the site and expanded using associated keys are called masks. This is an important list that allows you to make the site visible, that is, when a search engine operates, in response to a request, this site will also be displayed in the list of suggested ones.
  • Now you need to create lists of search queries for each mask. To do this, you will need to use the search engine that this site is oriented to, for example, Yandex, Rambler, Google, or others. The created list for each mask is subject to further editing and cleaning. This work is carried out based on clarification of the information posted on the site, as well as the actual search engine ratings.
  • Cleaning consists of removing unnecessary, uninformative and harmful requests. For example, if the list of a building materials website includes phrases with the words “course work,” then it should be removed, since it is unlikely to expand the target audience. After cleaning and final editing, you will get a version of actually working key queries, the content for which will promote the site into the zone of so-called visibility for search engines. In this case, the search engine internal links site will be able to show the desired page from the semantic core.

Summarizing all of the above, we can briefly say that the semantics of a site is determined by the total number of search engine query formulations used and their total frequency in the statistics of hits for a specific query.

All work on the formation and editing of semantics can be reduced to the following:

  1. analysis of the information posted on the site, the goals pursued by the creation of this site;
  2. compiling a general list of possible phrases based on site analysis;
  3. generating an extended version of keywords using associated queries (masks);
  4. generating a list of query options for each mask;
  5. editing (cleaning) the list to exclude unimportant phrases.

From this article you learned what the semantic core of a website is and how it should be compiled.

Good afternoon friends.

Surely you have already forgotten the taste of my articles. The previous material was quite a long time ago, although I promised to publish articles more often than usual.

Recently the amount of work has increased. I created a new project (an information site), worked on its layout and design, collected a semantic core and began publishing material.

Today there will be very voluminous and important material for those who have been running their website for more than 6-7 months (in some topics for more than 1 year), have a large number of articles (on average 100) and have not yet reached the bar of at least 500-1000 visits per day . The numbers are taken to a minimum.

The importance of the semantic core

In some cases, poor website growth is caused by improper technical optimization of the website. More cases where the content is of poor quality. But there are even more cases when texts are not written according to requests at all - no one needs the materials. But there is also a very huge part of people who create a website, optimize everything correctly, write high-quality texts, but after 5-6 months the site only begins to gain the first 20-30 visitors from search. At a slow pace, after a year there are already 100-200 visitors and the income figure is zero.

And although everything was done correctly, there are no errors, and the texts are sometimes even many times higher quality than competitors, but somehow it doesn’t work, for the life of me. We begin to attribute this problem to the lack of links. Of course, links give a boost in development, but this is not the most important thing. And without them, you can have 1000 visits to the site in 3-4 months.

Many will say that this is all idle chatter and you won’t get such numbers so quickly. But if we look, such numbers are not achieved precisely on blogs. Information sites (not blogs), created for quick earnings and return on investment, after about 3-4 months it is quite possible to reach a daily traffic of 1000 people, and after a year - 5000-10000. The numbers, of course, depend on the competitiveness of the niche, its volume and the volume of the site itself for the specified period. But, if you take a niche with fairly little competition and a volume of 300-500 materials, then such figures within the specified time frame are quite achievable.

Why exactly do blogs not achieve such quick results? The main reason is the lack of a semantic core. Because of this, articles are written for just one specific request and almost always for a very competitive one, which prevents the page from reaching the TOP in a short time.

On blogs, as a rule, articles are written in the likeness of competitors. We have 2- read blog. We see that they have decent traffic, we begin to analyze their sitemap and publish texts for the same requests, which have already been rewritten hundreds of times and are very competitive. As a result, we get very high-quality content on the site, but it performs poorly in searches, because... requires a lot of age. We are scratching our heads, why is my content the best, but doesn’t make it to the TOP?

That's why I decided to write detailed material about the semantic core of the site, so that you can collect a list of queries and, what is very important, write texts for such groups of keywords that, without buying links and in just 2-3 months, reach the TOP (of course, if the content is high-quality).

The material will be difficult for most if you have never encountered this issue in its correct way. But the main thing here is to start. As soon as you start acting, everything immediately becomes clear.

Let me make a very important remark. It concerns those who are not ready to invest in quality with their hard-earned coins and always try to find free loopholes. You can’t collect high-quality semantics for free, and that’s known fact. Therefore, in this article I describe the process of collecting semantics of maximum quality. None free ways or there will be no loopholes in this post! There will definitely be a new post where I’ll tell you about free and other tools with which you can collect semantics, but not in full and without the proper quality. Therefore, if you are not ready to invest in the basics of your website, then this material is of no use to you!

Despite the fact that almost every blogger writes an article about this. kernel, I can say with confidence that there are no normal free tutorials on the Internet on this topic. And if there is, then there is no one that would give a complete picture of what should be the output.

Most often, the situation ends with some newbie writing material and talking about collecting the semantic core, as well as using a service for collecting search query statistics from Yandex (wordstat.yandex.ru). Ultimately, you need to go to this site, enter queries on your topic, the service will display a list of phrases included in your entered key - this is the whole technique.

But in fact, this is not how the semantic core is assembled. In the case described above, you simply will not have a semantic core. You will receive some disconnected requests and they will all be about the same thing. For example, let’s take my niche “website building”.

What are the main queries that can be named without hesitation? Here are just a few:

  • How to create a website;
  • Website promotion;
  • Website creation;
  • Website promotion, etc.

The requests are about the same thing. Their meaning comes down to only two concepts: creation and promotion of a website.

After such a check, the wordstat service will display a lot of search queries that are included in the main queries and they will also be about the same thing. Their only difference will be in the changed word forms (adding some words and changing the arrangement of words in the query with changing endings).

Of course, it will be possible to write a certain number of texts, since requests can be different even in this option. For example:

  • How to create a wordpress website;
  • How to create a joomla website;
  • How to create a website on free hosting;
  • How to promote a website for free;
  • How to promote a service website, etc.

Obviously, separate material can be allocated for each request. But such compilation of the semantic core of the site will not be successful, because there will be no complete disclosure of information on the site in the selected niche. All content will be about only 2 topics.

Sometimes newbie bloggers describe the process of compiling a semantic core as analyzing individual queries in the Yandex Wordstat query analysis service. We enter some separate query that does not relate to the topic as a whole, but only to a specific article (for example, how to optimize an article), we get the frequency for it, and here it is - the semantic core is assembled. It turns out that in this way we must mentally identify all possible topics of articles and analyze them.

Both of the above options are incorrect, because... do not provide a complete semantic core and force you to constantly return to its compilation (second option). In addition, you will not have the site development vector in your hands and will not be able to publish materials in the first place, which should be published among the first.

Regarding the first option, when I once bought courses on website promotion, I constantly saw exactly this explanation for collecting the core queries for a website (enter the main key and copy all queries from the service to Text Document). As a result, I was constantly tormented by the question “What to do with such requests?” The following came to mind:

  • Write many articles about the same thing, but using different keywords;
  • Enter all these keys into the description field for each material;
  • Enter all the keys from the family. kernels in the general description field for the entire resource.

None of these assumptions were correct, nor, in general, was the semantic core of the site itself.

In the final version of collecting the semantic core, we should receive not just a list of queries in the amount of 10,000, for example, but have on hand a list of groups of queries, each of which is used for a separate article.

A group can contain from 1 to 20-30 requests (sometimes 50 or even more). In this case, we use all these queries in the text and the page in the future will bring traffic to all queries every day if it gets to 1-3 positions in the search. In addition, each group must have its own competition in order to know whether it makes sense to publish text on it now or not. If there is a lot of competition, then we can expect the effect of the page only after 1-1.5 years and with regular work to promote it (links, linking, etc.). Therefore, it is better to focus on such texts as a last resort, even if they have the most traffic.

Answers to possible questions

Question No. 1. It is clear that the output is a group of queries for writing text, and not just one key. Wouldn't the keys be similar to each other in this case and why not write for each request separate text?

Question No. 2. It is known that each page should be tailored to only one keyword, but here we get a whole group and, in some cases, with a fairly large content of queries. How, in this case, does the optimization of the text itself occur? After all, if there are, for example, 20 keys, then the use of each at least once in the text (even big size) already looks like text for a search engine, not for people.

Answer. If we take the example of requests from the previous question, then the first thing to sharpen the material will be precisely for the most frequent (1st) request, since we are most interested in its reaching the top positions. We consider this keyword to be the main one in this group.

Optimization for the main key phrase occurs in the same way as it would be done when writing text for only one key (the key in the title heading, using the required number of characters in the text and the required number of times the key itself, if required).

Regarding other keys, we also enter them, but not blindly, but based on an analysis of competitors, which can show the average number of these keys in texts from the TOP. It may happen that for most keywords you will receive zero values, which means that they do not require use in the exact occurrence.

Thus, the text is written using only main request in the text directly. Of course, other queries can also be used if the analysis of competitors shows their presence in the texts from the TOP. But this is not 20 keywords in the text in their exact occurrence.

Most recently I published material for a group of 11 keys. It seems like there are a lot of queries, but in the screenshot below you can see that only the main most frequent key has an exact occurrence - 6 times. The remaining key phrases do not have exact occurrences, but also diluted ones (not visible in the screenshot, but this is shown during the analysis of competitors). Those. they are not used at all.

(1st column – frequency, 2nd – competitiveness, 3rd – number of impressions)

In most cases, there will be a similar situation, when only a couple of keys need to be used in an exact occurrence, and all the rest will either be greatly diluted or not used at all, even in a diluted occurrence. The article turns out to be readable and there are no hints of focusing only on search.

Question No. 3. Follows from the answer to Question No. 2. If the remaining keys in the group do not need to be used at all, then how will they receive traffic?

Answer. The fact is that by the presence of certain words in the text, a search engine can determine what the text is about. Since keywords contain certain individual words that relate only to this key, they must be used in the text a certain number of times based on the same competitor analysis.

Thus, the key will not be used in the exact entry, but the words from the key will be individually present in the text and will also take part in the ranking. As a result, for these queries the text will also be found in the search. But in this case, the number of individual words should ideally be observed. Competitors will help.

I answered the main questions that could drive you into a stupor. I will write more about how to optimize text for groups of requests in one of the following materials. There will be information about analyzing competitors and about writing the text itself for a group of keys.

Well, if you still have questions, then ask your comments. I will answer everything.

Now let's start compiling the semantic core of the site.

Very important. I can’t really describe the whole process as it is in this article (I’ll have to do a whole webinar for 2-3 hours or a mini-course), so I’ll try to be brief, but at the same time informative and touch on maximum amount moments. For example, I will not describe in detail the configuration of the KeyCollector software. Everything will be cut down, but as clear as possible.

Now let's go through each point. Let's begin. First, preparation.

What is needed to collect a semantic core?


Before creating the semantic core of the site, we will set up a key collector to correctly collect statistics and parse queries.

Setting up KeyCollector

You can enter the settings by clicking on the icon in the top menu of the software.

First, what concerns parsing.

I would like to note the “Number of streams” and “Use primary IP address” settings. The number of threads to assemble one small core does not require a large number. 2-4 threads are enough. The more threads, the more proxy servers are needed. Ideally, 1 proxy per 1 thread. But you can also have 1 proxy for 2 streams (that’s how I parsed it).

Regarding the second setting, in case of parsing in 1-2 threads, you can use your main IP address. But only if it is dynamic, because... If a static individual entrepreneur is banned, you will lose access to the Yandex search engine. But still, priority is always given to using a proxy server, as it is better to protect yourself.

On the Yandex.Direct parsing settings tab, it is important to add your Yandex accounts. The more there are, the better. You can register them yourself or buy them, as I wrote earlier. I bought them because it’s easier for me to spend 100 rubles for 30 accounts.

You can add it from the buffer by copying the list of accounts in the required format in advance, or load it from a file.

Accounts must be specified in the “login:password” format, without specifying the host itself in the login (without @yandex.ru). For example, “artem_konovalov:jk8ergvgkhtf”.

Also, if we use several proxy servers, it is better to assign them to specific accounts. It would be suspicious if at first the request comes from one server and from one account, and the next time a request is made from the same Yandex account, the proxy server is different.

Next to the accounts there is a column “IP proxy”. Next to each account we enter a specific proxy. If there are 20 accounts and 2 proxy servers, then there will be 10 accounts with one proxy and 10 with another. If there are 30 accounts, then 15 with one server and 15 with another. I think you understand the logic.

If we use only one proxy, then there is no point in adding it to each account.

I talked a little earlier about the number of threads and the use of the main IP address.

The next tab is “Network”, where you need to enter proxy servers that will be used for parsing.

You need to enter a proxy at the very bottom of the tab. You can load them from the buffer in the desired format. I added it in a simple way. In each column of the line I entered information about the server that is given to you when you purchase it.

Next, we configure the export parameters. Since we need to receive all requests with their frequencies in a file on the computer, we need to set some export parameters so that there is nothing superfluous in the table.

At the very bottom of the tab (highlighted in red), you need to select the data that you want to export to the table:

  • Phrase;
  • Source;
  • Frequency "!" ;
  • The best form of the phrase (you don’t have to put it).

All that remains is to configure the anti-captcha solution. The tab is called “Antikapcha”. Select the service you are using and enter special key, which is located in the service account.

A special key for working with the service is provided in a letter after registration, but it can also be taken from the account itself in the “Settings - account settings” item.

This completes the KeyCollector settings. After making the changes, do not forget to save the settings by clicking on the large button at the bottom “Save changes”.

When everything is done and we are ready for parsing, we can begin to consider the stages of collecting the semantic core, and then go through each stage in order.

Stages of collecting the semantic core

It is impossible to obtain a high-quality and complete semantic core using only basic queries. You also need to analyze competitors’ requests and materials. Therefore, the entire process of compiling the core consists of several stages, which in turn are further divided into substages.

  1. The basis;
  2. Competitor analysis;
  3. Expansion of the finished list from stages 1-2;
  4. Collection of the best word forms for queries from stages 1-3.

Stage 1 – foundation

When collecting the kernel at this stage you need to:

  • Generate a main list of requests in the niche;
  • Expansion of these queries;
  • Cleaning.

Stage 2 – competitors

In principle, stage 1 already provides a certain volume of the core, but not fully, because we may be missing something. And competitors in our niche will help us find missing holes. Here are the steps to follow:

  • Collecting competitors based on requests from stage 1;
  • Parsing competitors' queries (site map analysis, open liveinternet statistics, analysis of domains and competitors in SpyWords);
  • Cleaning.

Stage 3 – expansion

Many people stop already at the first stage. Someone gets to the 2nd, but there are a number of additional queries that can also complement the semantic core.

  • We combine requests from stages 1-2;
  • We leave 10% of the most frequent words from the entire list, which contain at least 2 words. It is important that these 10% are no more than 100 phrases, because a large number will force you to dig deep into the process of collecting, cleaning and grouping. We need to assemble the kernel in a speed/quality ratio (minimal quality loss at maximum speed);
  • We expand these queries using the Rookee service (everything is in KeyCollector);
  • Cleaning.

Stage 4 – collecting the best word forms

The Rookee service can determine the best (correct) word form for most queries. This should also be used. The goal is not to determine the word that is more correct, but to find some more queries and their forms. In this way, you can pull up another pool of queries and use them when writing texts.

  • Combining requests from the first 3 stages;
  • Collection of the best word forms based on them;
  • Adding the best word forms to the list for all queries combined from stages 1-3;
  • Cleaning;
  • Export the finished list to a file.

As you can see, everything is not so fast, and especially not so simple. I outlined only a normal plan for compiling a kernel in order to get a high-quality list of keys at the output and not lose anything or lose as little as possible.

Now I propose to go through each point separately and study everything from A to Z. There is a lot of information, but it’s worth it if you need a really high-quality semantic core of the site.

Stage 1 – foundation

First, we create a list of basic niche queries. Typically, these are 1-3 word phrases that describe a specific niche issue. As an example, I propose to take the “Medicine” niche, and more specifically, the sub-niche of heart diseases.

What main requests can we identify? Of course, I won’t write everything, but I will give a couple.

  • Heart diseases
  • Heart attack;
  • Cardiac ischemia;
  • Arrhythmia;
  • Hypertension;
  • Heart disease;
  • Angina, etc.

In simple words, these are common names for diseases. There can be quite a lot of such requests. The more you can do, the better. But you shouldn’t enter it for show. It makes no sense to write out more specific phrases from the general ones, for example:

  • Arrhythmia;
  • Arrhythmia causes;
  • Treatment of arrhythmia;
  • Arrhythmia symptoms.

The main thing is only the first phrase. There is no point in indicating the rest, because... they will appear in the list during expansion using parsing from the left column of Yandex Wordstat.

To search for common phrases, you can use both competitor sites (site map, section names...) and the experience of a specialist in this niche.


Parsing will take some time, depending on the number of requests in the niche. All requests are placed in by default new group with the name “New Group 1”, if memory serves. I usually rename groups to understand which one is responsible for what. The group management menu is located to the right of the request list.

The rename function is in the context menu when you right-click. This menu will also be needed to create other groups in the second, third and fourth stages.

Therefore, you can immediately add 3 more groups by clicking on the first “+” icon so that the group is created in the list immediately after the previous one. There is no need to add anything to them yet. Let them just be.

I named the groups like this:

  • Competitors - it is clear that this group contains a list of requests that I collected from competitors;
  • 1-2 is a combined list of queries from the 1st (main list of queries) and 2nd (competitors’ queries) stages, in order to leave only 10% of the queries consisting of at least 2 words and collect extensions from them;
  • 1-3 – combined list of requests from the first, second and third (extensions) stages. We also collect the best word forms in this group, although it would be smarter to collect them in a new group (for example, the best word forms), and then, after cleaning them, move them to group 1-3.

After completing parsing from yandex.wordstat, you receive a large list of key phrases, which, as a rule (if the niche is small) will be within several thousand. Much of this is garbage and dummy requests and will have to be cleaned up. Some things will be sifted out automatically using the KeyCollector functionality, while others will have to be shoveled by hand and sitting for a while.

When all the requests are collected, you need to collect their exact frequencies. The overall frequency is collected during parsing, but the exact frequency must be collected separately.

To collect statistics on the number of impressions, you can use two functions:

  1. With the help of Yandex Direct - quickly, statistics are collected in batches, but there are limitations (for example, phrases of more than 7 words will not work, even with symbols);
  2. Using analysis in Yandex Wordstat - very slowly, phrases are analyzed one by one, but there are no restrictions.

First, collect statistics using Direct, so that this is as fast as possible, and for those phrases for which it was not possible to determine statistics using Direct, we use Wordstat. As a rule, there will be few such phrases left and they will be collected quickly.

Impression statistics are collected using Yandex.Direct by clicking on the appropriate button and assigning the necessary parameters.

After clicking on the “Get data” button, there may be a warning that Yandex Direct is not enabled in the settings. You will need to agree to activate parsing using Direct in order to begin determining frequency statistics.

You will immediately see how packs in the “Frequency” column! » accurate impression indicators for each phrase will begin to be recorded.

The process of completing a task can be seen on the “Statistics” tab at the very bottom of the program. When the task is completed, you will see a completion notification on the Event Log tab, and the progress bar on the Statistics tab will disappear.

After collecting the number of impressions using Yandex Direct, we check whether there are phrases for which the frequency was not collected. To do this, sort the “Frequency!” column. (click on it) so that the smallest or largest values ​​appear at the top.

If all zeros are at the top, then all frequencies are collected. If they do empty cells, then impressions were undecided for these phrases. The same applies when sorting by kill, only then you will have to look at the result at the very bottom of the list of queries.

You can also start collecting using Yandex Wordstat by clicking on the icon and selecting the required frequency parameter.


After selecting the frequency type, you will see how the empty cells gradually begin to fill.

Important: do not be alarmed if empty cells remain after the end of the procedure. The fact is that they will be empty if their exact frequency is less than 30. We set this in the parsing settings in Keycollector. These phrases can be safely selected (just like regular files in Windows Explorer) and deleted. The phrases will be highlighted in blue, right-click and select “Delete selected lines.”

When all the statistics have been collected, you can start cleaning, which is very important to do at each stage of collecting the semantic core.

The main task is to remove phrases that are not related to the topic, remove phrases with stop words and get rid of too many low frequency queries.

The latter ideally does not need to be done, but if we are in the process of compiling this. cores to use phrases even with a minimum frequency of 5 impressions per month, this will increase the core by 50-60 percent (or maybe even 80%) and force us to dig deep. We need to get maximum speed with minimal losses.

If we want to get the most complete semantic core of the site, but at the same time collect it for about a month (we have no experience at all), then take frequencies from 4-5 monthly impressions. But it’s better (if you’re a beginner) to leave requests that have at least 30 impressions per month. Yes, we will lose a little, but this is the price for maximum speed. And as the project grows, it will be possible to receive these requests again and use them to write new materials. And this is only on the condition that all this. the core has already been written out and there are no topics for articles.

The same key collector allows you to filter requests by the number of impressions and other parameters fully automatically. Initially, I recommend doing just this, and not deleting garbage phrases and phrases with stop words, because... it will be much easier to do this when the total core volume at this stage becomes minimal. What's easier, shoveling 10,000 phrases or 2,000?

Filters are accessed from the Data tab by clicking on the Edit Filters button.

I recommend first displaying all queries with a frequency of less than 30 and moving them to a new group so as not to delete them, as they may be useful in the future. If we simply apply a filter to display phrases with a frequency of more than 30, then after the next launch of KeyCollector we will have to reapply the same filter, since everything is reset. Of course, you can save the filter, but you will still have to apply it, constantly returning to the “Data” tab.

To save ourselves from these actions, we add a condition in the filter editor so that only phrases with a frequency of less than 30 are displayed.




In the future, you can select a filter by clicking on the arrow next to the floppy disk icon.

So, after applying the filter, only phrases with a frequency of less than 30 will remain in the list of queries, i.e. 29 and below. Also, the filtered column will be highlighted in color. In the example below you will only see a frequency of 30, because... I show all this using the example of a kernel that is already ready and everything is cleaned. Don't pay any attention to this. Everything should be as I describe in the text.

To transfer, you need to select all phrases in the list. Click on the first phrase, scroll to the very bottom of the list, hold down the “Shift” key and click once on the last phrase. This way, all phrases are highlighted and marked with a blue background.

A small window will appear where you need to select the movement.

Now you can remove the filter from the frequency column so that only queries with a frequency of 30 and above remain.


We have completed a certain stage of automatic cleaning. Next you will have to tinker with deleting garbage phrases.

First, I suggest specifying stop words to remove all phrases containing them. Ideally, of course, this is done immediately at the parsing stage so that they do not end up on the list, but this is not critical, since clearing stop words occurs automatically using the KeyCollector.

The main difficulty is compiling a list of stop words, because... They are different for each topic. Therefore, whether we cleaned up stop words at the beginning or now is not so important, since we need to find all the stop words, and this is a laborious task and not so fast.

On the Internet you can find general thematic lists, which include the most common words like “abstract, free, download, p...rn, online, picture, etc.”

First, I suggest using a general topic list to further reduce the number of phrases. On the “Data Collection” tab, click on the “Stop words” button and add them in a list.

In the same window, click on the “Mark phrases in the table” button to mark all phrases containing the entered stop words. But it is necessary that the entire list of phrases in the group is unchecked, so that after clicking the button, only phrases with stop words remain marked. It's very easy to unmark all phrases.


When only marked phrases with stop words remain, we either delete them or move them to a new group. I deleted it for the first time, but it’s still a priority to create a group “With stop words” and move all unnecessary phrases to it.

After cleaning, there were even fewer phrases. But that’s not all, because... we still missed something. Both the stop words themselves and phrases that are not related to the focus of our site. These may be commercial requests for which texts cannot be written, or texts can be written, but they will not meet the user’s expectations.

Examples of such queries may be related to the word “buy”. Surely, when a user searches for something with this word, he already wants to get to the site where they sell it. We will write text for such a phrase, but the visitor will not need it. Therefore, we do not need such requests. We look for them manually.

We slowly and carefully scroll through the remaining list of queries to the very end, looking for such phrases and discovering new stop words. If you find a word that is used many times, then simply add it to the already existing list stop words and click on the “Mark phrases in the table” button. At the end of the list, when we have marked all unnecessary queries during manual checking, we delete the marked phrases and the first stage of compiling the semantic core is completed.

We have obtained a certain semantic core. It is not quite complete yet, but it will already allow you to write the maximum possible part of the texts.

All that remains is to add to it a small part of the requests that we might have missed. The following steps will help with this.

Stage 2 competitors

At the very beginning, we compiled a list of common phrases related to the niche. In our case these were:

  • Heart diseases
  • Heart attack;
  • Cardiac ischemia;
  • Arrhythmia;
  • Hypertension;
  • Heart disease;
  • Angina, etc.

All of them belong specifically to the “Heart Disease” niche. Using these phrases, you need to search to find competitor sites on this topic.

We enter each of the phrases and look for competitors. It is important that these are not general thematic ones (in our case, medical sites with a general focus, i.e. about all diseases). What is needed is niche projects. In our case - only about the heart. Well, maybe also about the vessels, because... the heart is connected to the vascular system. I think you get the idea.

If our niche is “Recipes for salads with meat,” then in the future these are the only sites we should look for. If they are not there, then try to find sites only about recipes, and not in general about cooking, where everything is about everything.

If there is a general thematic site (general medical, women's, about all types of construction and repair, cooking, sports), then you will have to suffer a lot, both in terms of compiling the semantic core itself, because you will have to work long and tediously - collect the main list of requests, wait a long time for the parsing process, clean and group.

If on the 1st, and sometimes even on the 2nd page, you cannot find narrow thematic sites of competitors, then try using not the main queries that we generated before the parsing itself at the 1st stage, but queries from the entire list after parsing. For example:

  • How to treat arrhythmia with folk remedies;
  • Symptoms of arrhythmia in women and so on.

The fact is that such queries (arrhythmia, heart disease, heart disease...) are highly competitive and it is almost impossible to get to the TOP for them. Therefore, in the first positions, and maybe even on the pages, you will quite realistically find only general thematic portals about everything due to their enormous authority in the eyes of search engines, age and reference mass.

So it makes sense to use lower-frequency phrases consisting of more words to find competitors.

We need to parse their requests. You can use the SpyWords service, but its query analysis function is available on a paid plan, which is quite expensive. Therefore, for one core there is no point in upgrading the tariff on this service. If you need to collect several cores over the course of a month, for example 5-10, then you can buy an account. But again - only if you have a budget for the PRO tariff.

You can also use Liveinternet statistics if they are open for viewing. Very often, owners make it open to advertisers, but close the “search phrases” section, which is exactly what we need. But there are still sites where this section is open to everyone. Very rare, but available.

The most in a simple way is a banal view of sections and site maps. Sometimes we may miss not only some well-known niche phrases, but also specific requests. There may not be a lot of materials for them and you can’t create something for them separate section, but they can add a couple of dozen articles.

When we have found another list of new phrases to collect, we launch the same collection of search phrases from the left column of Yandex Wordstat, as in the first stage. We just launch it already, being in the second group “Competitors”, so that requests are added specifically to it.

  • After parsing, we collect the exact frequencies of search phrases;
  • We set a filter and move (delete) queries with a frequency of less than 30 to a separate group;
  • We clean out garbage (stop words and queries that are not related to the niche).

So, we received another small list of queries and the semantic core became more complete.

Stage 3 – expansion

We already have a group called “1-2”. We copy phrases from the “Main list of requests” and “Competitors” groups into it. It is important to copy and not move, so that all phrases remain in the previous groups, just in case. It will be safer this way. To do this, in the phrase transfer window you need to select the “copy” option.

We received all requests from stages 1-2 in one group. Now you need to leave in this group only 10% of the most frequent queries of the total number and which contain at least 2 words. Also, there should be no more than 100 pieces. We reduce it so as not to get buried in the process of collecting the core for a month.

First, we apply a filter in which we set the condition so that at least 2-word phrases are shown.


We mark all the phrases in the remaining list. By clicking on the “Frequency!” column, we sort the phrases in descending order of number of impressions, so that the most frequent ones are at the top. Next, select the first 10% of the remaining number of queries, uncheck them (right mouse button - uncheck selected lines) and delete the marked phrases so that only these 10% remain. Don’t forget that if your 10% is more than 100 words, then we stop at line 100, no more is needed.

Now we carry out the expansion using the keycollector function. The Rookee service will help with this.


We indicate the collection parameters as in the screenshot below.

The service will collect all extensions. Among the new phrases there may be very long keys, as well as symbols, so it will not be possible to collect frequency through Yandex Direct for everyone. Then you will have to collect statistics using the button from Wordstat.

After receiving statistics, we remove requests with monthly impressions of less than 30 and carry out cleaning (stop words, garbage, keywords that are not suitable for the niche).

The stage is over. We received another list of requests.

Stage 4 – collecting the best word forms

As I said earlier, the goal is not to determine the form of the phrase that will be more correct.

In most cases (based on my experience), collecting the best word forms will point to the same phrase that is in the search list. But, without a doubt, there will also be queries for which a new word form will be indicated, which is not yet in the semantic core. These are additional key queries. At this stage we achieve the core to maximum completeness.

When assembling its core, this stage yielded another 107 additional requests.

First, we copy the keys into the “1-3” group from the “Main Queries”, “Competitors” and “1-2” groups. The sum of all requests from all previously completed stages should be obtained. Next, we use the Rookee service using the same button as the extension. Just choose another function.


The collection will begin. Phrases will be added to new column"The best form of a phrase."

The best form will not be determined for all phrases, since the Rookee service simply does not know all the best forms. But for the majority the result will be positive.

When the process is completed, you need to add these phrases to the entire list so that they are in the same “Phrase” column. To do this, select all phrases in the “Best form of phrase” column, copy (right mouse button - copy), then click on the big one green button“Add phrases” and enter them.

It is very easy to make sure that phrases appear in the general list. Since adding phrases to the table like this happens at the very bottom, we scroll to the very bottom of the list and in the “Source” column we should see the add button icon.

Phrases added using extensions will be marked with a hand icon.

Since the frequency for the best word forms has not been determined, this needs to be done. Similar to the previous stages, we collect the number of impressions. Don’t be afraid that we collect in the same group where the other requests are located. The collection will simply continue for those phrases that have empty cells.

If it is more convenient for you, then initially you can add the best word forms not to the same group where they were found, but to a new one, so that only they are there. And already in it, collect statistics, clear garbage, and so on. And only then add the remaining normal phrases to the entire list.

That's all. The semantic core of the site has been assembled. But there is still a lot of work left. Let's continue.

Before the next steps, you need to download all requests with the necessary data into an Excel file on your computer. We set the export settings earlier, so you can do it right away. The export icon in the KeyCollector main menu is responsible for this.

When you open the file, you should get 4 columns:

  1. Phrase;
  2. Source;
  3. Frequency!;
  4. The best form of the phrase.

This is our final semantic core, containing the maximum pure and necessary list of queries for writing future texts. In my case (narrow niche) there were 1848 requests, which equals approximately 250-300 materials. I can’t say for sure - I haven’t completely ungrouped all the requests yet.

For immediate use, this is still a raw option, because... requests are in a chaotic order. We also need to scatter them into groups so that each contains the keys to one article. This is the ultimate goal.

Ungrouping the semantic core

This stage is completed quite quickly, although with some difficulties. The service will help us http://kg.ppc-panel.ru/. There are other options, but we will use this one in view of the fact that with it we will do everything in the quality/speed ratio. What is needed here is not speed, but first and foremost quality.

A very useful thing about the service is that it remembers all actions in your browser using cookies. Even if you close this page or the browser as a whole, everything will be saved. This way there is no need to do everything at once and be afraid that everything may be lost in one moment. You can continue at any time. The main thing is not to clear your browser cookies.

I will show you how to use the service using the example of several fictitious queries.

We go to the service and add the entire semantic core (all queries from the excel file) exported earlier. Just copy all the keys and paste them into the window as shown in the image below.

They should appear in the left column “Keywords”.

Ignore the presence of shaded groups on the right side. These are groups from my previous core.

Let's look at left column. There is an added list of queries and a “Search/Filter” line. We will use the filter.

The technology is very simple. When we enter part of a word or phrase, the service in real time leaves in the list of queries only those that contain the entered word/phrase in the query itself.

See below for more clarity.


I wanted to find all queries related to arrhythmia. I enter the word “Arrhythmia” and the service automatically leaves in the list of queries only those that contain the entered word or part of it.

The phrases will move to a group, which will be called one of the key phrases of this group.

We received a group containing all the keywords for arrhythmia. To see the contents of a group, click on it. A little later we will further divide this group into smaller groups, since there are a lot of keys with arrhythmia and they are all under different articles.

Thus, at the initial stage of grouping, you need to create large groups that combine a large number of keywords from one niche question.

If we take the same topic of heart disease as an example, then first I will create a group “arrhythmia”, then “heart disease”, then “heart attack” and so on until there are groups for each disease.

As a rule, there will be almost as many such groups as the main niche phrases generated at the 1st stage of collecting the core. But in any case, there should be more of them, since there are also phrases from stages 2-4.

Some groups may contain 1-2 keys altogether. This may be due, for example, to a very rare disease and no one knows about it or no one is looking for it. That's why there are no requests.

In general, when the main groups are created, it is necessary to break them down into smaller groups, which will be used to write individual articles.

Inside the group, next to each key phrase there is a cross; by clicking on it, the phrase is removed from the group and goes back to the ungrouped list of keywords.

This is how further grouping occurs. I'll show you with an example.

In the image you can see that there are key phrases related to arrhythmia treatment. If we want to define them in a separate group for a separate article, then we remove them from the group.


They will appear in the list in the left column.

If there are still phrases in the left column, then to find deleted keys from the group, you will have to apply a filter (use the search). If the list is completely divided into groups, then only deleted requests will be there. We mark them and click on “Create group”.


Another one will appear in the “Groups” column.

Thus, we distribute all the keys by topic and, ultimately, a separate article is written for each group.

The only difficulty in this process lies in analyzing the need to ungroup some keywords. The fact is that there are keys that are inherently different, but they do not require writing separate texts, but detailed material is written on many issues.

This is clearly expressed in medical topics. If we take the example of arrhythmia, then there is no point in making the keys “arrhythmia causes” and “arrhythmia symptoms”. The keys to treating arrhythmia are still in question.

This will be found out after analyzing the search results. We go to Yandex search and enter the analyzed key. If we see that the TOP contains articles devoted only to symptoms of arrhythmia, then given key allocated to a separate group. But, if the texts in the TOP reveal all the issues (treatment, causes, symptoms, diagnosis, etc.), then ungrouping into in this case need not. We cover all these topics in one article.

If in Yandex exactly such texts are at the top of the search results, then this is a sign that ungrouping is not worth doing.

The same can be exemplified by the key phrase “causes of hair loss.” There may be “causes of hair loss in men” and “...in women.” Obviously, you can write a separate text for each key, based on logic. But what will Yandex say?

We enter each key and see what texts are there. There are separate ones for each key detailed texts, then the keys are ungrouped. If in the TOP for both queries there are general materials on the key “causes of hair loss”, within which questions regarding women and men are disclosed, then we leave the keys within one group and publish one material where topics on all keys are revealed.

This is important, since it is not for nothing that the search engine identifies texts in the TOP. If the first page contains exclusively detailed texts on a specific issue, then there is a high probability that by breaking a large topic into subtopics and writing material on each, you will not get to the TOP. And so one material has every chance of getting good positions for all requests and collecting good traffic for them.

In medical topics, a lot of attention needs to be paid to this point, which significantly complicates and slows down the ungrouping process.

At the very bottom of the “Groups” column there is an “Unload” button. Click on it and we get a new column with a text field containing all the groups, separated by a line indent.

If not all keys are in groups, then in the “Upload” field there will be no spaces between them. They only appear when ungrouping is completely completed.

Select all the words (key combination Ctrl+A), copy them and paste them into new excel file.

Under no circumstances click on the “Clear all” button, as absolutely everything you have done will be deleted.

The ungrouping phase is over. Now you can safely write the text for each group.

But, for maximum efficiency, if your budget does not allow you to write out everything. core in a couple of days, and there is only a strictly limited opportunity for regular publication of a small number of texts (10-30 per month, for example), then it is worth determining the competition of all groups. This is important because the groups with the least competition produce results in the first 2-3-4 months after writing without any links. All you need to do is write high-quality, competitive text and optimize it correctly. Then time will do everything for you.

Definition of group competition

I would like to note right away that a low-competition request or group of requests does not mean that they are very micro low-frequency. The beauty is that there is a fairly decent number of requests that have low competition, but at the same time have a high frequency, which immediately gives such an article a place in the TOP and attracts solid traffic to one document.

For example, a very realistic picture is when a group of queries has 2-5 competition, and the frequency is about 300 impressions per month. Having written only 10 such texts, after they reach the TOP, we will receive at least 100 visitors daily. And these are only 10 texts. 50 articles - 500 visits and these figures are taken as a ceiling, since this only takes into account traffic for exact queries in the group. But traffic will also be attracted from other tails of requests, and not just from those in groups.

This is why identifying the competition is so important. Sometimes you can see a situation where there are 20-30 texts on a site, and there are already 1000 visits. And the site is young and there are no links. Now you know what this is connected with.

Request competition can be determined through the same KeyCollector absolutely free of charge. This is very convenient, but ideally this option is not correct, since the formula for determining competition is constantly changing with changes in search engine algorithms.

Better identify competition using the service http://mutagen.ru/. It is paid, but is closer to real indicators as much as possible.

100 requests cost only 30 rubles. If you have a core for 2000 requests, then the entire check will cost 600 rubles. 20 free checks are given per day (only to those who top up their balance by any amount). You can evaluate 20 phrases every day until you determine the competitiveness of the entire core. But this is very long and stupid.

Therefore, I use the mutagen and have no complaints about it. Sometimes there are problems related to processing speed, but this is not so critical, since even after closing the service page, the check continues in the background.

The analysis itself is very simple. Register on the site. We top up the balance with any in a convenient way and verification is available at the main address (mutagen.ru). We enter a phrase into the field and it immediately begins to be evaluated.

We see that for the query being checked, the competitiveness turned out to be more than 25. This is a very high indicator and can be equal to any number. The service does not display it as real, since this does not make sense due to the fact that such competitive requests are almost impossible to promote.

The normal level of competition is considered to be up to 5. It is precisely such requests that are easily promoted without unnecessary gestures. Slightly higher indicators are also quite acceptable, but queries with values ​​greater than 5 (for example, 6-10) should be used after you have already written texts for minimal competition. Maximum importance quick text in the TOP.

Also during the assessment, the cost of a click in Yandex.Direct is determined. It can be used to estimate your future earnings. We take into account guaranteed impressions, the value of which we can safely divide by 3. In our case, we can say that one click on a Yandex Direct advertisement will bring us 1 ruble.

The service also determines the number of impressions, but we do not look at them, since the frequency of the “!request” type is not determined, but only the “request”. The indicator turns out to be inaccurate.

This analysis option is suitable if we want to analyze a single request. If you need a mass check of key phrases, then there is a special link on the main page at the top of the key entry field.


On the next page we create a new task and add a list of keys from the semantic core. We take them from the file where the keys are already ungrouped.

If there are enough funds on the balance for analysis, the check will begin immediately. The duration of the check depends on the number of keys and may take some time. I analyzed 2000 phrases for about 1 hour, when the second time checking only 30 keys took several hours.

Immediately after starting the scan, you will see the task in the list, where there will be a “Status” column. This will help you understand whether the task is ready or not.

In addition, after completing the task, you will immediately be able to download a file with a list of all phrases and the level of competition for each. The phrases will all be in the same order as they were ungrouped. The only thing is that the spaces between the groups will be removed, but this is not a big deal, because everything will be intuitive, because each group contains keys on a different topic.

In addition, if the task has not yet completed, you can go inside the task itself and see the competition results for already completed requests. Just click on the task name.

As an example, I will show the result of checking a group of queries for the semantic core. Actually, this is the group for which this article was written.

We see that almost all requests have a maximum competition of 25 or closer to it. This means that for these requests, I either won’t see the first positions at all or won’t see them for a very long time. I wouldn’t write such material on the new site at all.

Now I published it only to create quality content for the blog. Of course, my goal is to get to the top, but that’s later. If the manual reaches the first page, at least only for the main request, then I can already count on significant traffic only to this page.

The last step is to create the final file, which we will look at in the process of filling the site. The semantic core of the site has already been assembled, but managing it is not entirely convenient yet.

Creating a final file with all data

We will need the KeyCollector again, as well as the latest excel file that we received from the mutagen service.

We open the previously received file and see something like the following.

We only need 2 columns from the file:

  1. key;
  2. competition.

You can simply delete all other contents from this file so that only the necessary data remains, or you can create a completely new file, make a beautiful header line there with the names of the columns, highlighting it, for example, in green, and copy the corresponding data into each column.

Next, we copy the entire semantic core from this document and add it to the key collector again. Frequencies will need to be collected again. This is necessary so that the frequencies are collected in the same order as the key phrases are located. Previously, we collected them to weed out garbage, and now to create the final file. Of course, we add phrases to a new group.

When the frequencies are collected, we export the file from the entire frequency column and copy it into the final file, which will have 3 columns:

  1. key phrase;
  2. competitiveness;
  3. frequency.

Now all that’s left to do is sit a little and do the math for each group:

  • Average group frequency - it is not the indicator of each key that is important, but the average indicator of the group. It is calculated as the usual arithmetic mean (in Excel - the "AVERAGE" function);
  • Divide the total frequency of the group by 3 to bring the possible traffic to real numbers.

So that you don’t have to worry about mastering calculations in Excel, I have prepared a file for you where you just need to enter data in the required columns and everything will be calculated fully automatically. Each group must be calculated separately.

There will be a simple calculation example inside.

As you can see, everything is very simple. From the final file, simply copy all the entire phrases of the group with all the indicators and paste them into this file in place of the previous phrases with their data. And the calculation will happen automatically.

It is important that the columns in the final file and in mine are in exactly this order. First the phrase, then the frequency, and only then the competition. Otherwise, you will count nonsense.

Next, you will have to sit a little to create another file or sheet (I recommend) inside the file with all the data about each group. This sheet will not contain all the phrases, but only the main phrase of the group (we will use it to determine what kind of group it is) and the calculated values ​​of the group from my file. This is a kind of final file with data about each group with calculated values. This is what I look at when choosing texts for publication.

You will get the same 3 columns, only without any calculations.

In the first one we insert the main phrase of the group. In principle, any one is possible. It is only needed to copy it and, through a search, find the location of all the group keys that are on another sheet.

In the second we copy the calculated frequency value from my file, and in the third we copy the average value of the group’s competition. We take these numbers from the “Result” line.

The result will be the following.

This is the 1st sheet. The second sheet contains all the contents of the final file, i.e. all phrases with indicators of their frequency and competition.

Now we look at the 1st sheet. We select the most “cost-effective” group of keys. We copy the phrase from the 1st column and find it using search (Ctrl + F) in the second sheet, where the rest of the group phrases will be located next to it.

That's all. The manual has come to an end. At first glance, everything is very complicated, but in reality it is quite simple. As I already said at the very beginning of the article, you just have to start doing it. In the future, I plan to make a video manual using these instructions.

Well, on this note I end the instructions.

All friends. If you have any suggestions or questions, I'm waiting for you in the comments. See you later.

P.S. Record for volume of content. In Word it turned out to be more than 50 sheets of small print. It was possible not to publish an article, but to create a book.

Best regards, Konstantin Khmelev!


The semantic core of the site is what users on the Internet use to search for services, products and any other information that this site offers. For webmasters, this is an action plan for promoting the resource. In an ideal plan The semantic core of the site is created once, before optimization and promotion begin.


The semantic core of a website is usually compiled in several stages:

  1. All sorts of words (phrases) that are appropriate to the topic of the site are selected. At first, you can limit yourself to 100–200 search queries. In order to know which queries are suitable for you, answer yourself the question “What do I want to dedicate my site to?”
  2. Expansion of the semantic core through associative queries
  3. Inappropriate words should be eliminated. Here you filter out those phrases that you will not use to promote your site. There are usually more than half of such words.
  4. Highly competitive queries for which there is no point in promoting the site are eliminated. Typically, three words out of five or more are removed.
  5. And lastly, this is the correct distribution of the list of search queries on the resource pages. It is recommended to leave highly competitive queries on the main page of the resource; less competitive ones should be grouped according to their meaning and placed on other pages. To do this, you need to create a document in Excel and break down the keywords into pages.

Selection of search queries and checking frequency

The first thing you need to do is collect as many different queries as possible on your topic that interest users on the Internet. There are two methods for this:

  • Free, which includes: Wordstat Yandex, Slovoeb, the old-fashioned way, hints from Google (External Keyword Tool), analysis of the semantics of competitors and search tips.
  • Paid ones which include Key Collector, Semrush, Pastukhov databases and some other services.

These tools are suitable for various purposes (for example, Semrush is best used for burzhunet). Of course, all this can be entrusted to optimizers, but there is a possibility that you will receive an incomplete semantic core.

Many people use Pastukhov’s database to collect key phrases, but with Key Collector it is much more convenient to collect requests from services.

At the initial stage, it is better to collect queries in Excel; it looks like this:


If Google is more important for your resource, then focus on it, but also take into account and analyze keywords from Yandex. It is also very important to collect a long tail of low-frequency queries; they will get you traffic much faster.

Another option you can use is to find out key phrases (words) from your competitors and use them. At this stage, you simply collect as many key phrases (words) that are relevant to the topic of your resource as possible, and then move on to the next stage - filtering.

Analysis of requests, removal of dummies

This stage is already simpler; here you need to filter out dummy words and those that are not related to the theme of the site. For example, you have lunch delivery in Kyiv, but there are other cities on the list.

How to identify empty requests? Go to Yandex Wordstat and enter the keyword:


You see 881 impressions per month, but to be more precise:


Now a completely different picture emerges. This may not be the best example, but the main thing is that you get the idea. There are a large number of key phrases for which sufficient traffic is visible, although in reality there is nothing there. That's why you need to weed out such phrases.

For example, if a person before (or after) typing the request “lunch delivery” entered another phrase in the search bar (called one search session), then Yandex makes the assumption that these search phrases are somehow interconnected. If such a relationship is observed among several people, then such associative queries are shown in the right column of wordstat.


Such search queries are sorted in the wordstat window in descending order of frequency of their entry in conjunction with the main query this month (the frequency of their use in the Yandex search engine is shown). You need to use this information to expand the semantic core of your resource.

Distribution of requests across pages

After this, you need to distribute the keywords (phrases) you collected on the pages of your site. Distribution is much easier when you don’t yet have the pages themselves.

Focus primarily on keywords in search queries and their frequency. To deal with competition, you should do this: dedicate the main page to one or two highly competitive queries.

For moderately competitive or low competitive queries, optimize section and article pages accordingly.

If there is semantic similarity in the search queries, simply collect the same phrases and define them in one group. When creating keywords to promote a resource, always use not only standard tools, but also a creative approach.

Combining non-standard and classic methods You can simply and quickly create the semantic core of a site, choose the most optimal promotion strategy and achieve success much faster!

Greetings, dear reader of the web-revenue blog!

Today I decided to tell you about the basics of SEO promotion, namely compiling the semantic core of the site (SY).

Semantic Core- this is a library search words or phrases and their morphological forms that most accurately characterize the activities of the site, as well as the goods or services offered by the site. Roughly speaking, the compilation of NL is the compilation of the link structure of the site’s target queries for which it is planned to promote the site!

Why is the semantic core of a website created?

1.SYA forms the theme of the site, which will be taken into account by search engines.

2. A correctly formed syntax is the basis for the optimal structure of a web resource.

3. In order to link each page of the site with semantic information to a specific part of the language (keywords).

4. To form a limited set of keywords in order to rationally allocate resources for website promotion for specific keywords (words).

5. To estimate the cost of website promotion in search engines (search engines)

Basic Concepts

Before we begin compiling a semantic core, let’s look at a few basic concepts.

1. All queries that users enter into search engines can be divided into:

High frequency (HF)

Mid frequency (MF)

Low frequency (LF)

How to find out which group this or that request belongs to, you ask. In general, there are no strict boundaries and boundaries that separate high-frequency from mid-frequency, and mid-frequency from low-frequency requests. This largely depends on the theme of the site. If we take average values, then we will consider low-frequency queries that are received up to 450-700 times a month; medium frequency - up to 1.2 - 2 thousand times a month; high frequency – over 2 thousand times a month.

Many webmasters recommend starting to promote sites, first for low-frequency and medium-frequency queries, this is on the one hand correct, but there is one thing: some low-frequency and mid-frequency queries have such high competition that it will be no easier to promote for such queries, What would you use to promote high frequencies?

So, when compiling a website’s SYNOPSIS, you should not rely only on the frequency of words; you also need to determine how difficult it will be to compete for a given request.

Therefore, we will introduce 3 more groups of queries:

Highly competitive (VC);

Moderately competitive (SC);

Low competitive (LC);

Many people consider VC to be highly competitive, SC – moderately competitive, and NC – low competitive. However, this is not always the case. Nowadays, in many areas, low-frequency queries have become so in demand that it is better not to try to get to the TOP using them. Sometimes it’s easier to reach the top in midrange frequencies (but this is also rare). Sometimes you need to take into account words that people often misspell (for example: Volkswagen it can be spelled Volcwagen or Volswagen) or words that a person types without forgetting to change the keyboard layout - “cjplfnm uba fybvfwb” instead of “create a GIF animation”. Such mistakes in words can also be used to promote a website well!

And three more important concepts:

Primary queries are queries that characterize the resource “in general” and are the most general in the subject of the site. For example, the primary requests for my website are: website creation, website promotion, website promotion, making money on the website, etc.

The main ones are queries included in the list of the semantic core, those for which promotion is advisable. For my blog: how to create a website, how to promote a website, making money on a website, etc.

Auxiliary (associative) - queries that were also typed by people entering the main queries. They are usually similar to the main queries. For example, for the query SEMANTIC CORE, internal optimization, website promotion, SEO will be associative.

I have explained the basic theory, now we can move on to the basics of compiling a semantic core:

1. If you are composing a SL for your website, then first sit down and think about what your website is about, what queries a person can use to find it, try to come up with as many keywords (sentences) for your topic as possible and write them down in a text document. For example, if you are going to make a website about various drinks, cocktails, etc., after a little thought you can write down something like the following words: soft drinks, cocktail recipes, making cocktails, fruit drinks..., etc.

And if you are doing this for any client, then we find out from the client a list of words by which he wants to promote his site.

2. We analyze the sites of competitors from the top 10 (we look at what queries they are promoted for and receive most of the traffic)

3. We use the client’s price list (name of goods, services, etc.)

4. We try to find synonyms for keywords (hard drive - hard drive - HDD)

5. Collection of keywords that are suitable for your personal blog, Internet resource or business. Here you can use wordstat search query statistics, or it is better to use special software such as Key Collector.

6. Traffic counting for selected search queries. For this, you can also use a key collector or link aggregators: seopult or webeffector.

7. Removing dummy requests. These are search queries for which the impression values ​​are greatly inflated or even inflated. You won’t get visitors from dummy requests.

8. Removing keywords with a very high promotion budget. You can again find out the approximate budget: seopult or webeffector. You can also filter out highly competitive queries.

Then we distribute them throughout the site.

The general scheme for compiling a strategic language looks something like this:

As a result, we will receive a list of keywords for our site. That's basically the whole scheme. It is not that complicated, but it is quite labor-intensive and takes quite a lot of time. But as I wrote above, this is the basis of the site, which is worth paying close attention to.

Mistakes that are usually made when compiling SYNOPSIS:

When selecting keywords, try to avoid the following problems:

SYNOPSIS should not consist of general phrases that poorly characterize your site or, conversely, too narrow. For example, if a visitor wants to learn about creating a vertical drop-down menu in Wordpress, then he will type “creating a vertical drop-down menu in Wordpress”, and not “creating a website”, “creating a site menu”, “web site”, etc. In general, you should cover more specific queries. On the contrary, too narrow queries will not give you enough visitors. Try to find a middle ground.

If you have little text, you shouldn’t put a lot of keywords on it. For example, this article is tailored to 3 keys! But the volume of text is quite substantial - more than 6 thousand characters. Ideally, there should be 1 key per article. But you can use the following rule: one or two keywords per 2 thousand characters of an article.

When creating the site's language, misspelled words that users accidentally make when typing are not taken into account. I spoke about them above.

Well, I think the theory is enough for you, and in the next article we will talk about!







2024 gtavrl.ru.