# Site review for an Irish online paper

### Summary <a href="#h.6kh58mvcc7q" id="h.6kh58mvcc7q"></a>

#### Goals <a href="#h.rqx7bx4bhk8v" id="h.rqx7bx4bhk8v"></a>

1. Increase user traffic through empowering organic search users.
2. Improve Behavior Metrics including ‘Time of page’ and ‘Pages/ Session’ .
3. Improve the overall website’s Google AdSense performance .

#### Key research takeaways <a href="#h.quee3rn80usz" id="h.quee3rn80usz"></a>

1. User engagement level is low.
2. Content organization / grouping is ineffective.
3. Keywords number is relatively low.
4. Website has technical issues and content duplication problems.

#### Key suggestions <a href="#h.kvh7w8xepqci" id="h.kvh7w8xepqci"></a>

1. Disallow crawling/indexing of an alternative domain version (*com* vs. *ie*).
2. Stick to a single URL structure and use “rel=canonical” to signify what’s your preferred URL.
3. Map different URL paths together to cure a high rate of 404 (page not found) errors. &#x20;
4. Adopt a fine Category -> Sub-category / Tag structure to better organize your content.
5. Update the ‘physical’ URL structure to mirror the structure above and give Google a better sense of attribution / connection between the content.
6. Use Breadcrumbs to allow for better user navigation at a post level: Main -> Category / Tag -> Article.
7. Update your page’s layout to allow for more effective internal navigation and distribution of PageRank.

#### Suggestions of further actions <a href="#h.588l1s1ski7x" id="h.588l1s1ski7x"></a>

1. Fix the alternative domain issue
2. Fix the concurrent URLs issue
3. Create the robots.txt on the base domain
4. Configure the .htaccess file to map the ‘old’ URL vs. the ‘new’ ones
5. Create a map for content breakdown & attribution: category > sub-categories / tags
6. Update the URL structure to mirror the map above for newly created content
7. Produce mockups of new landing pages: category (sub-categories / tags) and post
8. Develop widgets for supplementary navigation
9. Update a basic XML sitemap and re-submit the website to Google index

### Current metrics analysis <a href="#h.1soe6u4r6y5d" id="h.1soe6u4r6y5d"></a>

Organic traffic (Google) accounts for around 27% of all users. 47% of all traffic comes from social media (Facebook & Twitter). Interestingly, Facebook mobile users account for almost as much traffic as organic ones and have a potential to outperform organic very soon.

The average website’s user is rather loyal, i.e. 45% of daily visits are attributed to the returning users who visited your pages 3 times or more already. Now, the loyal users are something you may want to dissect and compare as it gives a sense of:

1. What content is the sweetest?
2. What is the most popular behavior pattern?
3. What’s working for them the best?

Now, loyal visitors are into TV/Movie and Politics. The same concerns the new users, so it’s arguably your user segments’ characteristics that have to be addressed through a more focused content strategy.

<figure><img src="https://909672381-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2Ft8rE11rvpFLCUkZIlwPg%2Fuploads%2FbYtdOHUhSQJA9WgQr342%2Fimage2.png?alt=media&#x26;token=15768ebc-dadc-4500-9d38-d25e282475fe" alt=""><figcaption></figcaption></figure>

Loyal visitors would use Desktop as much as mobile devices. New users would prefer mobile devices more. Obviously, it has to do with the fact that you are able to drive more \[organic] traffic from Mobile than Desktop: your pages’ organic ranking position is 2 times better on Mobile than on Desktop (due to relatively effective mobile optimization) – see below.     &#x20;

<figure><img src="https://909672381-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2Ft8rE11rvpFLCUkZIlwPg%2Fuploads%2Fsw0T5Zs5qKFh1Tc61kel%2Fimage4.png?alt=media&#x26;token=ecf24921-2f43-499f-86c6-19eb856d6ba7" alt=""><figcaption></figcaption></figure>

In most cases loyal users would come to the site through Social (54%), 24% of them would use organic (Google) to type in the brand name mostly, and 16% would just come from bookmark (Direct). By contrast, only 30% of new users come from Social while 30% come from Organic. It means that currently your Social media strategy needs to be focused at attracting new users to the website vs. maintaining the community.

Notably, even the loyal users have low engagement rate with your content. It means that you cannot make them move further beyond one-page-visit-only: the ‘through traffic’ from the starting page is rather low, i.e. 4-6% depending on the post. It has to do with supplementary navigation, i.e. other pieces of reading that may interest your users. The most popular routes (stacks of pages) were – see below.

<figure><img src="https://909672381-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2Ft8rE11rvpFLCUkZIlwPg%2Fuploads%2FAtwiGMOuXMcnHx539Ugx%2Fimage3.png?alt=media&#x26;token=22538f25-3b1d-469e-bfe3-4b8479db24f5" alt=""><figcaption></figcaption></figure>

#### Metrics analysis takeaways <a href="#h.utiu9jknsp6m" id="h.utiu9jknsp6m"></a>

Organic channels for user acquisition are underdeveloped. Organic users would mostly type in your brand, podcast and anchors-related search terms.

Importantly, there has not been detected any actual search terms pattern on your website: most search terms that you were ranked for are very time-and-events-sensitive (for example, derek davis \[death]) or incidental /trivial (for example, sex video or Leicester city sex video).

Social accounts play an important role in maintaining the core of daily traffic, but they are relatively ineffective in both driving new users and developing loyal visitors’ engagement with the content.

Your page’s layout needs a quality supplementary content (i.e. related items, ‘you might be also interested in’ or ‘most popular in the category’) to be able to effectively improve engagement metrics.

### Site SEO research <a href="#h.p05ldzk5gtic" id="h.p05ldzk5gtic"></a>

The whole picture of your current SEO profile is below. It particularly shows that the website has:

1. Relatively  low number of keywords (around 850).
2. The vast majority of keywords are trivial AND ranked the second page of Google (11-20).
3. Relatively good backlinks profile.

<figure><img src="https://909672381-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2Ft8rE11rvpFLCUkZIlwPg%2Fuploads%2FaDFsugsAvdjkMD2bBmnR%2Fimage6.png?alt=media&#x26;token=a280349e-c602-4cea-940d-a3f61721350d" alt=""><figcaption></figcaption></figure>

For comparison, [http://www.irishmirror.ie/](https://www.google.com/url?q=http://www.irishmirror.ie/\&sa=D\&source=editors\&ust=1701101022373027\&usg=AOvVaw1LYeq4-NemS6-FnvP-WHvX) has 7,000 keywords and 275k organic traffic vs. 12k of yours.

The core problem lies in ineffective content organization. So, what you have to do is improve Content organization & internal linking through updating the whole page template.

Now, the job suggested has to secure:

1\. A clear attribution of each content piece with a category > subcategory (tag) through:

1. the very URL structure, i.e. ‘main/category/subcategory/humanly-readable-news-title’ for example like this [http://www.irishmirror.ie/news/world-news/schoolboy-left-fighting-life-after-5833522](https://www.google.com/url?q=http://www.irishmirror.ie/news/world-news/schoolboy-left-fighting-life-after-5833522\&sa=D\&source=editors\&ust=1701101022373993\&usg=AOvVaw2lkHecUP549Wvtn6UPmfHz)
2. the breadcrumbs navigation – see below: *News > Weird News > Religion*

<figure><img src="https://909672381-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2Ft8rE11rvpFLCUkZIlwPg%2Fuploads%2Fkqgs7f1BpkUWBYIWytiS%2Fimage8.png?alt=media&#x26;token=df487587-9096-4f17-925e-f0eb7a050294" alt=""><figcaption></figcaption></figure>

c) use of tags to group the like content

2\. Effective internal navigation through:

1. Main menu supplementary navigation to most trending topics
2. Sidebar navigation to the hottest news / most promoted content
3. ‘Related articles’ navigation that showcases similar articles in the \[same] category
4. Footer navigation that brings main content categories in an expanded format to get them accessed by crawlers more often

**It’s important because it**:

1. creates citation of potential keywords on greater number of pages
2. allows for a better distribution of PageRank, i.e. enhances chances of ranking of each single page on the website
3. creates better user experience

For example, consider ‘Derek Davis’ search term. Google normally would give precedence in ranking to websites that have more content on the search term (deemed more relevant) as well as render more quality content (expertness, supplementary content, number of backlinks, etc).

Now your website has around 230 pages featuring that term: [https://www.google.ie/webhp?hl=ru#q=site:www.newstalk.com+derek+davis](https://www.google.com/url?q=https://www.google.ie/webhp?hl%3Dru%23q%3Dsite:www.newstalk.com%2Bderek%2Bdavis\&sa=D\&source=editors\&ust=1701101022375620\&usg=AOvVaw26dV-cZA8k_IkuvWd8xp0K) whereas [http://www.irishexaminer.com/](https://www.google.com/url?q=http://www.irishexaminer.com/\&sa=D\&source=editors\&ust=1701101022375827\&usg=AOvVaw0iCysxy9HAoWWZpEXtuOL6) has 15 times more pages:  [https://www.google.ie/#q=site:www.irishexaminer.com+derek+davis](https://www.google.com/url?q=https://www.google.ie/%23q%3Dsite:www.irishexaminer.com%2Bderek%2Bdavis\&sa=D\&source=editors\&ust=1701101022376000\&usg=AOvVaw3dZAIUUNhz7bJd2-wbwP39).

What’s important to understand is that through proper content organization and navigation [http://www.irishexaminer.com/](https://www.google.com/url?q=http://www.irishexaminer.com/\&sa=D\&source=editors\&ust=1701101022376285\&usg=AOvVaw0wP5krppbXhSq5yLndjiGx) wins the battle off the lower number of original pieces of content – it would just better use the supplementary navigation machinery.

In addition to content organization and navigation (internal linking) your website has some issues with the technical side of SEO, i.e. crawlability, missing and duplicate content. It has core importance as to the overall organic ranking ability. &#x20;

#### Crawlability and indexing <a href="#h.lyvs3jqmssl6" id="h.lyvs3jqmssl6"></a>

Crawlability and indexing is the compliance to Google tech requirements regarding ability to access/crawl/index the pages as well as non-duplication.

The website is accessible under two different domains [https://newstalk.ie/](https://www.google.com/url?q=https://newstalk.ie/\&sa=D\&source=editors\&ust=1701101022377213\&usg=AOvVaw05ibYwox0vNVEwgKJl-qv5) vs. [http://www.newstalk.com/](https://www.google.com/url?q=http://www.newstalk.com/\&sa=D\&source=editors\&ust=1701101022377388\&usg=AOvVaw1bCL4y7FQOC1gTkNbpRR8b). It may cause a content duplication issues that make it harder for Google Algorithm to figure out what page to show up in search results. Normally, large scale of duplication breaks the credibility of tour pages and results in poor indexing of your content.

Suggestion: pick one domain version and disallow ALL other version from indexing by Googlebot.

Here’s what you need to do for [https://newstalk.ie](https://www.google.com/url?q=https://newstalk.ie\&sa=D\&source=editors\&ust=1701101022378419\&usg=AOvVaw371brUzSiohqWbhPRIHdlf) domain:

1. Create a robots.txt file in your document root
2. Add the following statement to it (means disallow all crawlers):
3. User-agent: \*
4. Disallow:/
5. Submit robots.txt via Google Webmaster Tools
6. Remove the domain from your webmaster console

Another issue is preferred URL structure. It means that on your ‘base domain’, i.e. [http://www.newstalk.com/](https://www.google.com/url?q=http://www.newstalk.com/\&sa=D\&source=editors\&ust=1701101022379152\&usg=AOvVaw35tm5FiUN4Ea93JF88-P01) there are 2 kinds of URL structures in place: [http://www.newstalk.com/reader/47.301/44495/0/](https://www.google.com/url?q=http://www.newstalk.com/reader/47.301/44495/0/\&sa=D\&source=editors\&ust=1701101022379321\&usg=AOvVaw0bbulYy5MstxMLwWfnWMMt) vs. [http://www.newstalk.com/Kildare-Newbridge-houses-fire-court-Millfield-Manor-man-arrested](https://www.google.com/url?q=http://www.newstalk.com/Kildare-Newbridge-houses-fire-court-Millfield-Manor-man-arrested\&sa=D\&source=editors\&ust=1701101022379538\&usg=AOvVaw2FSUTmbQ0y7m2Vph5s-L6_). You have already put a redirect for browser users; yet, Googlebot can still see both URLs as it does not use a browser to access your pages. It causes another round of duplication issues. As a result, you have more than 10,000 duplicate page titles & obviously pages content.

I recommend that you used “rel=canonical” Meta tag to show Google the proffered version of your page (the one you want to appear on search results). It can be done through installing ready-made extensions (widgets) to your CRM or through coding your files ‘manually’.

#### Server connectivity and missing page (404) errors <a href="#h.zff5c8o6s3hf" id="h.zff5c8o6s3hf"></a>

The Google Webmaster Tools shows that recently Googlebot’s experienced problems with accessibility of your web pages on the server. Obviously, your script cannot process search queries of the kind of ‘search.php?search\_term=ireland\&currentpage=1’, i.e. the search option is broken. Suggestion: fix or remove search all together.

You have a relatively high number of 404 errors, i.e. your page has been removed and is no longer accessible from the old URL address. As a result of lots of 404 errors your web pages PageRank (PR) cannot normally circulate.  The PR is an index historically used by Google to determine the relative probability of a page's ranking in organic search. It is said to be gained by counting-in the ‘quality’ of incoming links and is distributed through the internal linking across different pages. Likewise, if a page get a ‘404 error’ the PR would not flow to the relevant pages within the same category but at best would be channeled to the ‘default’ pages (Home).

404 errors originate out of two concurrent URL structures, i.e. ‘/reader/ based’ vs. ‘human readable URLs’. Apparently, these are produced by operation of your Content management system (CMS).

So, the first thing to do is stop producing alternative URL paths to the same content. Alternatively, specify a canonical URL path for ‘/reader/ based’ URL. I recommend that Google would only index ‘human readable URLs’ – those will be canonical ones.

Afterwards, you will need to map the ‘/reader/ based’ URL vs. ‘human readable URLs’. In my observation, most of 404 errors have the same pattern: inaccurate mapping of the concurrent URLs. Likewise, [http://www.newstalk.com/Passenger-removed-from-plane-over-notebook-doodles](https://www.google.com/url?q=http://www.newstalk.com/Passenger-removed-from-plane-over-notebook-doodles\&sa=D\&source=editors\&ust=1701101022381437\&usg=AOvVaw3O4VyvbSShpUIRE2s1Wl6C) gets 404 errors while [http://www.newstalk.com/Passenger-removed-from-plane-over-notebook-doodles-](https://www.google.com/url?q=http://www.newstalk.com/Passenger-removed-from-plane-over-notebook-doodles-\&sa=D\&source=editors\&ust=1701101022381673\&usg=AOvVaw3ucjf9mDIDoFBe5iHrj77X) gets a user redirected to a ‘/reader/ based’ URL (no 404 error).

**Suggestions of further actions (repeated):**

10. Fix the alternative domain issue
11. Fix the concurrent URLs issue
12. Create the robots.txt on the base domain
13. Configure the .htaccess file to map the ‘old’ URL vs. the ‘new’ ones
14. Create a map for content breakdown & attribution: category > sub-categories / tags
15. Update the URL structure to mirror the map above for newly created content
16. Produce mockups of new landing pages: category (sub-categories / tags) and post
17. Develop widgets for supplementary navigation
18. Update a basic XML sitemap and re-submit the website to Google index
