Reformulate text – Paraphrase Online https://www.paraphrase-online.com/blog Creative Writing Blog Wed, 13 Apr 2022 06:08:24 +0000 en hourly 1 https://wordpress.org/?v=5.0.16 Case study – furniture industry https://www.paraphrase-online.com/blog/word-changer/case-study-furniture-industry/ Thu, 02 Sep 2021 05:42:50 +0000 https://www.paraphrase-online.com/blog/?p=1321 Continue readingCase study – furniture industry]]> Google Ads and website positioning are two promotional activities that exist in full harmony – supporting short- and long-term sales and increasing brand recognition on the Internet. Although the cost of this specific struggle for a customer on two fronts may seem considerable, it is worth treating it as an investment that pays back quickly and with an excess, which may even exceed the contribution several times over. Let’s show it on the example of our client operating in the furniture industry.

Positioning the website

Our client’s company has been operating for over twenty years, providing high-quality furniture throughout Europe. Nevertheless, when we started SEO in December 2020, its online visibility was very low. The website did not appear on any phrase, neither in TOP3 nor in TOP10 of search results, i.e. it was never visible on the first page in Google for phrases related to the industry. What’s more – it appeared only for one phrase in the TOP50, which almost completely limited the organic traffic on the site.

Top phrases in December 2020
Therefore, it was necessary to take both on-site activities, i.e. within the website itself, and off-site activities – consisting in acquiring high-quality external links. We started out by optimizing the site technically to improve loading times and usability for the user. We also focused on adding meta elements, important for the website evaluation by search engine robots. However, these changes were only the beginning. Together with the client, we also decided to implement a blog with tips on the website, and saturate each entry with key phrases best suited to the store’s offer. In addition, we strengthened the website by acquiring links on external websites leading to the client’s website.

Thanks to these actions, after six months we can see a significant improvement in the visibility of the website – we currently have 5 phrases in TOP3 of search results and 25 phrases in TOP10, i.e. on the first page. Naturally, the TOP50 saw the greatest progress – here we have already introduced over 500 phrases and we are constantly fighting for their even better positions.

Top phrases in June 2021
It is worth remembering that indexing pages by search engine algorithms is a time-consuming process, and it is sometimes even a dozen or so months for Google robots to be “recognized”. So this is just the beginning of the SEO road, and even better results can be expected in the coming months.

In a short time, however, you can achieve a significant increase in traffic and conversion thanks to Google Ads. Let us compare the results of our client’s campaign from 3 months ago with those achieved now.

In the period from February 16 to March 16, the ads generated over 21,700 clicks. In the period May 16 – June 14, this number dropped by approximately 1,500 clicks. However, does this mean a decrease in the effectiveness of the campaign? On the contrary. Clicks themselves do not bring the customer tangible profits – they only mean that the user has entered his website. In the process of optimizing ads, we try to limit accidental, empty clicks from recipients who are not really interested in the offer and do not plan to make a purchase. Choosing the right target group may therefore slightly reduce the number of clicks, but this does not translate into a reduction in profits.

The number and cost of the conversion

During the three months of campaigning, the number of conversions doubled – there were 319 of them at the turn of February and March, and now their number is 663 per month. At the same time, however, the cost of acquiring one conversion decreased by over 50%, which means that at the price you had to pay three months ago to acquire one customer, you can now acquire two.

How it’s possible? The answer is simple. The longer the campaign runs and the more data our specialists and bid optimization systems have to analyze, the more effective your ads can be. We constantly monitor when there is the greatest chance of conversion, what devices are used by users who most often show interest in the offer, where are the people who convert and many other factors. On this basis, we can predict which actions will bring the best effect and optimize the rates per click to achieve the campaign’s goal at the lowest possible cost.

What profits did the campaign bring?

Together with the client, we agreed that we would stick to the chosen budget for the campaign – for 3 months it increased by only 2.6%. With almost the same budget, revenues increased by more than 56%, i.e. about 88,500.

So how has the return on investment changed? At the turn of February and March, each euro spent on campaigns brought the client about EUR 14.50 in revenue, which is a very good result anyway. Now, however, EUR 22 in costs corresponds to EUR 22 in revenue, an increase of 50%.

Channel bonding = more profits

By combining a Google Ads advertising campaign with positioning, you can significantly increase not only brand visibility and recognition, but also real profits in a relatively short time – this applies to both online stores and service industries. Can we promise that the profits for each company will be on a similar level? Of course not. Marketing campaigns on the Internet are governed by too many variables to take anything for granted, and we focus on honesty and transparency in contact with the client. Instead of empty promises, however, we offer experience, knowledge and commitment, and this in many cases, with good cooperation with the client, brings even higher profits than expected.

]]>
Core Web Vitals https://www.paraphrase-online.com/blog/seo/core-web-vitals/ Mon, 25 Jan 2021 06:30:43 +0000 https://www.paraphrase-online.com/blog/?p=1106 Continue readingCore Web Vitals]]> The internet is constantly changing. The websites from its origins barely resemble those that accompany us today. Not only their appearance and the form of providing information have changed, but also our expectations. The differences also include the way the website is assessed – today it is not enough that it simply IS. In this context, we need to consider what Core Web Vitals are and what they actually tell us.

Core Web Vitals – website quality comes first

For a long time, Google has placed great emphasis on the speed and readability of the presented content. One of the first elements of these changes was certainly the shift towards users of mobile devices (also known as “Mobilegeddon”), which in 2015 electrified SEOs and web developers. The next ones include the need to ensure the security of the website and its users, or SSL. Now it is turning to user experience.

In May 2021, page quality will become part of the Google algorithm and will have an impact on the site’s position in the search engine ranking. According to official sources, the quality of the website will not matter more than the content, but it will allow you to promote websites that convey them in a manner tailored to the user’s needs.

The quality of the website will be analyzed in terms of several factors mentioned earlier:
– adaptation to mobile devices (mobile-first indexing for all websites from March 2021);
– browsing safety – no harmful and misleading content, i.e. malware and phishing elements;
HTTPS;
– no full-screen ads that prevent access to website content.

In addition, there is also Core Web Vitals, known in USA as Basic Internet Indicators. The pointers are designed to help you gauge the user experience of page loading and availability during and after rendering.

Google emphasizes that work on the method of testing the quality of the website will be continued. In the coming years, the list of factors taken into account will certainly change and will be expanded with new points. We will be informed about the changes in advance, the search engine will also give us time to adapt our websites to the new requirements.

Basic Internet Indicators – let’s take a closer look at them

One of the more controversial seo-themes is website speed. In specialist groups, more than once or twice, we have witnessed long discussions about how a high PageSpeed Insights score improves a page’s position. It is true that the site’s rating in the tool does not translate directly into search engine rankings. It is known, however, that a website that has no problems with displaying its content correctly even with a weaker Internet connection will be welcomed warmly by users. This, in turn, will translate into the length of the visit, the willingness to share the link and the chance to achieve the conversion we set.

The research carried out by the tools I mentioned in the page speed article checks specific elements of the site, but does not provide full information about the experience of a less-one user. Their analysis is necessary, but it focuses on aspects of the website other than Basic Internet Indicators.

Core Web Vitals allow us to define the experiences that our website users experience. The data allowing to determine the website quality level is collected based on the actual visits to the website. Thanks to this, the behavior of the website is examined on a wide cross-section of cases – different speeds of network connections, sizes and technological capabilities of devices.

Core Web Vitals in the current version focuses on three elements:
– the largest rendering of the content,
– delay on first action,
– collective shift of the layout.

As I mentioned before, Google is planning to expand its website quality analysis and we can expect that this list will expand in the coming years. Google does not exclude the possibility of modifying current indicators, e.g. extending the scope of research.

LCP – Largest Contentful Paint – Largest content rendering

LCP determines the rendering time of the largest fragment of the page visible in the browser window (everything we see before scrolling). The elements that are taken into account are:
– graphics,
– video,
– an element containing background graphics, downloaded using the url (…) function in CSS,
– block elements (e.g. <p>, <div>, <ul>, <ol>, <hx>, etc.).

If the content of individual tags extends below the scroll line, these fragments are not taken into account in the analysis. In the case of graphics, the analysis is based on the smallest size reached by the element. If the graphic is rendered smaller than the values specified in the tag during rendering, its rendered size will be crucial for the analysis. When the graphic is stretched, the size of the sample taken for testing will depend on the values specified in the CSS.

How do you know which item is the largest? Web pages render in stages, and the HTML code is read in order – from the first line to the last. Therefore, as your site renders, the object marked “largest” will change. A given fragment of the page is defined as the “greatest rendering of content” only when it is fully loaded. Note – If the user moves the screen while rendering the page, further changes to “largest contentful paint” will not be taken into account.

What about items that load off-screen and only later appear within the user’s view? In most cases, these extracts will not be reported. In the opposite situation – when they render within the viewport and are pushed below the scroll line, they will still be taken into account and may be included in the final result.

Highest Content Render – Score

In PageSpeed Insight, values are provided in two ways:
– as a percentage, collected on the basis of data collected from users in the last 28 days. LCP was satisfactory in 39% of the cases,
– in seconds, based on tests performed during the page load simulation.

PageSpeed Insights analyzes data on the specified address. For more detailed information about pages within the entire domain, you may want to check out the Google Search Console and one of the newest tabs labeled “Basic Internet Indicators”. There are two reports available on the page displayed on mobile devices and on computers.

The report in GSC lists the pages where there is a problem with poor LCP results. In a situation where the tool has identified the same problem on many subpages – the table contains one address and an annotation about the number of pages on which the error repeats. The exact addresses are available in the extended view, available after clicking on the record.

Poor LCP Score – Effective Ways to Improve

The key to a good LCP score is the optimization of points that slow down the performance of the website. Slow server, the need to execute JS and CSS code, which blocks further rendering of the site and client-side rendering – these are the most common causes of poor results. Let’s take a look at a remedy for the most common problems:
– slow response from the server – the key value here will be TTFB – Time To First Byte (time from sending the query to receiving the first byte of the response). Slow response is reported e.g. in PageSpeed Insights, it can also be checked with other website speed analysis tools.

The primary way to improve the TTFB score is to optimize the server and eliminate processes that degrade its performance. For this purpose, it is worth looking at the advice of your hosting provider or using the help of a specialist.

– JavaScript and CSS rendering blocking – here are the tricks that are recommended for every website speed optimization – CSS minification, removing unused fragments and considering asynchronous loading of style files. In terms of JavaSript, the optimization will be similar – minification of JS files and removal of unused fragments.

Optimization of other elements, e.g. graphics compression, use of CDN and prioritization of loading certain resources, may also affect the improvement of the LCP score.

FID – First Input Delay – Delay on first input

FID refers to the interactivity of the website – it defines the time between the first action performed by the user and the moment when the browser starts handling this action. Clicks (in the case of mobile – taps) and button presses are taken into account. The FID score focuses on the analysis of response time, but does not examine the processing time of a user-initiated event.

The lag on the first action will not always be measured – some users choose not to interact with the page while it is loading. Interestingly, in the case of this indicator, the analysis takes place only during the actual use of the website – similar studies are not carried out during the simulation performed by Lighthouse.

Every Internet user has certainly encountered a situation in which he immediately wanted to move to another place after entering the website. However, clicking on the link or expanding the rest of the text was not possible immediately – the site’s reaction to our traffic was significantly delayed. Why is this happening? While the site is rendering, the browser is busy handling the files that make up the page. Only after completing this process is it able to respond to user behavior. The waiting time for the interactivity of the page translates directly into the subjective assessment by the user and determining it as “slow” or “fast”.

When describing the LCP, I mentioned that HTML is loaded line by line – so why is it not possible for the loaded fragments to be active earlier? Some tags, such as <input>, <a>, or <select>, require rendering of the main thread to be closed. As for the other obstacles – the <head> section of the .html file usually contains links to CSS and JS resources that the browser must parse before continuing to render the page. Hence the previous suggestions to minify these files (i.e. remove whitespace) and move to the rest of the code fragments unnecessary for rendering the first view. In the case of events that are registered by “event listeners” placed in JS, their operation is possible only after all the code has been executed.

Delay on first action – result

A satisfactory result is in the range of 0-100ms. Above this time it is worth introducing corrections, and above 300 ms – you have to.

The information from PageSpeed Insight is based on user input – in this case 82% of the cases are satisfactory.

More detailed data, along with data on specific addresses that need improvement, are included in the “Basic Internet Indicators” tab in Google Search Console.

Poor FID score – effective ways to improve

Ideally, the tools would report 95-99% of “green responses”. If our current results are not that high yet, making changes to the site will certainly do them good.

As in the case of LCP, first of all, it is recommended to optimize the code necessary to load the key elements of the website and move fragments with a lower priority to the rest of the file. It is also worth removing unused fragments and minifying.

The main reason for low FID results is usually the JS code. Optimizing files and breaking individual fragments into smaller, asynchronously executed batches should significantly speed up the rendering process and reduce resource locking.

Code coming from outside the website will also affect the browser response speed – limiting it will reduce the number of processes performed by the browser.

CLS – Cumulative Layout Shift

The CLS score is the sum of the changes due to an unexpected page layout shift. The change counts when the element changes its position at the time of rendering. CLS does not count as new elements appearing on the page or resizing existing ones, as long as they do not cause other page components to move. Contrary to the indicators listed above, CLS collects data during the entire visit to the page, not in the first moments of its rendering.

Jumps in the page layout are created, among others by asynchronously loading resources or by dynamically adding additional elements to the DOM structure. Such situations are noticeable, for example, when using pages on which graphics are loaded without the declared size – matching the appropriate size to the screen size may cause a sudden change in the appearance of the page, which causes the user to be moved to another part of it.

The resources downloaded from external sources, built on the basis of values other than the target page, and personalized content, generated without proper formatting, may also be problematic here. Unexpected offset can also be caused by extended font rendering.

Not every shift is bad, depending on the page design, some user actions will cause changes from the original layout of the elements. It is important that this type of shift is directly related to the action performed.

CLS is based on data collected from users and allows you to determine the actual experience on the site, which is not always available under the conditions in which the website developer works.

Cumulative Layout Offset Result

A good CLS score is in the range of 0-0.1. A score above 0.25 means that you need to make changes and improve the experience of its users. It is assumed that the website is stable when at least 75% of users receive a score below 0.1.

Data on the shift can be found in PageSpeed Insights. The percentage score is based on an analysis of user data – in this case, 90% get a good score. The PSI also provides information collected under laboratory conditions. Detailed information on more subpages can be obtained from Google Search Console.

Poor CLS Score – Effective Ways to Improve

A better CLS score can be achieved by adhering to general website building standards. The precise description of graphics and videos plays a significant role here – adding to them attributes related to the size of the element (width, height and the aspect ratio associated with them) makes it easier for the browser to correctly arrange the elements during rendering, without the need to change while loading subsequent pieces of content. In the case of dynamic added graphics, it is recommended to use placeholders that reserve space for the element added later.

For some time now, PageSpeed Insights has recommended the use of a css font-display, which is responsible for keeping the font on the page. Another possibility is to load the font in advance using the preload attribute placed in the <link> element. Preload causes the selected resources to be downloaded before rendering the page. Thanks to this, the problem of downtime and waiting for all files from the first part of the queue to be loaded is eliminated later in the page display.

Basic Internet Indicators – UX Matters!

If you look at the solutions that will have a positive impact on the score of individual indicators, it is easy to see that they are not new. Many of them were already recommended by Google during the first versions of PageSpeed Insights. However, indicators and their entry into the search engine algorithm are a good reason to take a closer look at your websites and actually make changes that will make it easier to use the websites.

]]>