Site review for an Irish online paper
SEO review and optimisation suggestion for an Irish online magazine
Summary
Goals
Increase user traffic through empowering organic search users.
Improve Behavior Metrics including βTime of pageβ and βPages/ Sessionβ .
Improve the overall websiteβs Google AdSense performance .
Key research takeaways
User engagement level is low.
Content organization / grouping is ineffective.
Keywords number is relatively low.
Website has technical issues and content duplication problems.
Key suggestions
Disallow crawling/indexing of an alternative domain version (com vs. ie).
Stick to a single URL structure and use βrel=canonicalβ to signify whatβs your preferred URL.
Map different URL paths together to cure a high rate of 404 (page not found) errors.
Adopt a fine Category -> Sub-category / Tag structure to better organize your content.
Update the βphysicalβ URL structure to mirror the structure above and give Google a better sense of attribution / connection between the content.
Use Breadcrumbs to allow for better user navigation at a post level: Main -> Category / Tag -> Article.
Update your pageβs layout to allow for more effective internal navigation and distribution of PageRank.
Suggestions of further actions
Fix the alternative domain issue
Fix the concurrent URLs issue
Create the robots.txt on the base domain
Configure the .htaccess file to map the βoldβ URL vs. the βnewβ ones
Create a map for content breakdown & attribution: category > sub-categories / tags
Update the URL structure to mirror the map above for newly created content
Produce mockups of new landing pages: category (sub-categories / tags) and post
Develop widgets for supplementary navigation
Update a basic XML sitemap and re-submit the website to Google index
Current metrics analysis
Organic traffic (Google) accounts for around 27% of all users. 47% of all traffic comes from social media (Facebook & Twitter). Interestingly, Facebook mobile users account for almost as much traffic as organic ones and have a potential to outperform organic very soon.
The average websiteβs user is rather loyal, i.e. 45% of daily visits are attributed to the returning users who visited your pages 3 times or more already. Now, the loyal users are something you may want to dissect and compare as it gives a sense of:
What content is the sweetest?
What is the most popular behavior pattern?
Whatβs working for them the best?
Now, loyal visitors are into TV/Movie and Politics. The same concerns the new users, so itβs arguably your user segmentsβ characteristics that have to be addressed through a more focused content strategy.
Loyal visitors would use Desktop as much as mobile devices. New users would prefer mobile devices more. Obviously, it has to do with the fact that you are able to drive more [organic] traffic from Mobile than Desktop: your pagesβ organic ranking position is 2 times better on Mobile than on Desktop (due to relatively effective mobile optimization) β see below.
In most cases loyal users would come to the site through Social (54%), 24% of them would use organic (Google) to type in the brand name mostly, and 16% would just come from bookmark (Direct). By contrast, only 30% of new users come from Social while 30% come from Organic. It means that currently your Social media strategy needs to be focused at attracting new users to the website vs. maintaining the community.
Notably, even the loyal users have low engagement rate with your content. It means that you cannot make them move further beyond one-page-visit-only: the βthrough trafficβ from the starting page is rather low, i.e. 4-6% depending on the post. It has to do with supplementary navigation, i.e. other pieces of reading that may interest your users. The most popular routes (stacks of pages) were β see below.
Metrics analysis takeaways
Organic channels for user acquisition are underdeveloped. Organic users would mostly type in your brand, podcast and anchors-related search terms.
Importantly, there has not been detected any actual search terms pattern on your website: most search terms that you were ranked for are very time-and-events-sensitive (for example, derek davis [death]) or incidental /trivial (for example, sex video or Leicester city sex video).
Social accounts play an important role in maintaining the core of daily traffic, but they are relatively ineffective in both driving new users and developing loyal visitorsβ engagement with the content.
Your pageβs layout needs a quality supplementary content (i.e. related items, βyou might be also interested inβ or βmost popular in the categoryβ) to be able to effectively improve engagement metrics.
Site SEO research
The whole picture of your current SEO profile is below. It particularly shows that the website has:
Relatively low number of keywords (around 850).
The vast majority of keywords are trivial AND ranked the second page of Google (11-20).
Relatively good backlinks profile.
For comparison, http://www.irishmirror.ie/ has 7,000 keywords and 275k organic traffic vs. 12k of yours.
The core problem lies in ineffective content organization. So, what you have to do is improve Content organization & internal linking through updating the whole page template.
Now, the job suggested has to secure:
1. A clear attribution of each content piece with a category > subcategory (tag) through:
the very URL structure, i.e. βmain/category/subcategory/humanly-readable-news-titleβ for example like this http://www.irishmirror.ie/news/world-news/schoolboy-left-fighting-life-after-5833522
the breadcrumbs navigation β see below: News > Weird News > Religion
c) use of tags to group the like content
2. Effective internal navigation through:
Main menu supplementary navigation to most trending topics
Sidebar navigation to the hottest news / most promoted content
βRelated articlesβ navigation that showcases similar articles in the [same] category
Footer navigation that brings main content categories in an expanded format to get them accessed by crawlers more often
Itβs important because it:
creates citation of potential keywords on greater number of pages
allows for a better distribution of PageRank, i.e. enhances chances of ranking of each single page on the website
creates better user experience
For example, consider βDerek Davisβ search term. Google normally would give precedence in ranking to websites that have more content on the search term (deemed more relevant) as well as render more quality content (expertness, supplementary content, number of backlinks, etc).
Now your website has around 230 pages featuring that term: https://www.google.ie/webhp?hl=ru#q=site:www.newstalk.com+derek+davis whereas http://www.irishexaminer.com/ has 15 times more pages: https://www.google.ie/#q=site:www.irishexaminer.com+derek+davis.
Whatβs important to understand is that through proper content organization and navigation http://www.irishexaminer.com/ wins the battle off the lower number of original pieces of content β it would just better use the supplementary navigation machinery.
In addition to content organization and navigation (internal linking) your website has some issues with the technical side of SEO, i.e. crawlability, missing and duplicate content. It has core importance as to the overall organic ranking ability.
Crawlability and indexing
Crawlability and indexing is the compliance to Google tech requirements regarding ability to access/crawl/index the pages as well as non-duplication.
The website is accessible under two different domains https://newstalk.ie/ vs. http://www.newstalk.com/. It may cause a content duplication issues that make it harder for Google Algorithm to figure out what page to show up in search results. Normally, large scale of duplication breaks the credibility of tour pages and results in poor indexing of your content.
Suggestion: pick one domain version and disallow ALL other version from indexing by Googlebot.
Hereβs what you need to do for https://newstalk.ie domain:
Create a robots.txt file in your document root
Add the following statement to it (means disallow all crawlers):
User-agent: *
Disallow:/
Submit robots.txt via Google Webmaster Tools
Remove the domain from your webmaster console
Another issue is preferred URL structure. It means that on your βbase domainβ, i.e. http://www.newstalk.com/ there are 2 kinds of URL structures in place: http://www.newstalk.com/reader/47.301/44495/0/ vs. http://www.newstalk.com/Kildare-Newbridge-houses-fire-court-Millfield-Manor-man-arrested. You have already put a redirect for browser users; yet, Googlebot can still see both URLs as it does not use a browser to access your pages. It causes another round of duplication issues. As a result, you have more than 10,000 duplicate page titles & obviously pages content.
I recommend that you used βrel=canonicalβ Meta tag to show Google the proffered version of your page (the one you want to appear on search results). It can be done through installing ready-made extensions (widgets) to your CRM or through coding your files βmanuallyβ.
Server connectivity and missing page (404) errors
The Google Webmaster Tools shows that recently Googlebotβs experienced problems with accessibility of your web pages on the server. Obviously, your script cannot process search queries of the kind of βsearch.php?search_term=ireland¤tpage=1β, i.e. the search option is broken. Suggestion: fix or remove search all together.
You have a relatively high number of 404 errors, i.e. your page has been removed and is no longer accessible from the old URL address. As a result of lots of 404 errors your web pages PageRank (PR) cannot normally circulate. The PR is an index historically used by Google to determine the relative probability of a page's ranking in organic search. It is said to be gained by counting-in the βqualityβ of incoming links and is distributed through the internal linking across different pages. Likewise, if a page get a β404 errorβ the PR would not flow to the relevant pages within the same category but at best would be channeled to the βdefaultβ pages (Home).
404 errors originate out of two concurrent URL structures, i.e. β/reader/ basedβ vs. βhuman readable URLsβ. Apparently, these are produced by operation of your Content management system (CMS).
So, the first thing to do is stop producing alternative URL paths to the same content. Alternatively, specify a canonical URL path for β/reader/ basedβ URL. I recommend that Google would only index βhuman readable URLsβ β those will be canonical ones.
Afterwards, you will need to map the β/reader/ basedβ URL vs. βhuman readable URLsβ. In my observation, most of 404 errors have the same pattern: inaccurate mapping of the concurrent URLs. Likewise, http://www.newstalk.com/Passenger-removed-from-plane-over-notebook-doodles gets 404 errors while http://www.newstalk.com/Passenger-removed-from-plane-over-notebook-doodles- gets a user redirected to a β/reader/ basedβ URL (no 404 error).
Suggestions of further actions (repeated):
Fix the alternative domain issue
Fix the concurrent URLs issue
Create the robots.txt on the base domain
Configure the .htaccess file to map the βoldβ URL vs. the βnewβ ones
Create a map for content breakdown & attribution: category > sub-categories / tags
Update the URL structure to mirror the map above for newly created content
Produce mockups of new landing pages: category (sub-categories / tags) and post
Develop widgets for supplementary navigation
Update a basic XML sitemap and re-submit the website to Google index
Last updated