To achieve this goal, it is essential to know which set of products sells best at what price and the actual value of a subscriber in a particular region of the world.
One of the tools that allow for quick answers to these questions is mobile app A/B testing. Modern software development is based on hypothesis testing systems. The best practices of large companies rely on creating a product based on data from A/B tests.
The first step is to define a hypothesis and establish specific goals. For example, the hypothesis could be that changing the design of the onboarding paywall will lead to an increase in the conversion rate by x%, or offering a free trial version will increase the conversion of paid users by y%. The goals should be clear and measurable, such as increase the purchase conversion from onboarding by 5%.
Here's a brief framework for defining a hypothesis:
Let's examine the stages of formulating a hypothesis in more detail.
First and foremost, it is necessary to conduct research and gather both quantitative and qualitative data to identify a specific problem that you will address within the experiment. For example, according to your statistics, when segmenting by paywall, you observe that Paywall 1 has five times more purchases compared to Paywall 2.
Once you have gathered all the data and noticed, for example, that users frequently request a trial version (or complain about high prices or a cumbersome onboarding process). As discussed earlier, within one experiment, we identify which problem we will address. Let's move on to the second step.
The next step in forming the hypothesis is to document the expected changes and their presumed impact. For example, based on your research, you assume that providing a trial version of the product on the onboarding screen should increase the purchase conversion rate from 3% to 7%.
The final step in forming the hypothesis is to determine the metrics by which we will measure the expected outcome. We recommend tracking at least these metrics: Views, Trials, Purchases, Proceeds, Refunds, ARPU, and ARPPU.
As you know, there are multiple factors that influence a user's decision to subscribe to a mobile application, including subscription price, availability of a free trial period, the appearance and information on the paywall, and other aspects.
Within an experiment, you can modify various paywall parameters:
Once you have formulated your hypothesis and determined the experiment's goal, you need to define the following parameters for your experiment:
Decide on the prices and products that will be featured in each tested paywall. For example, you can compare the current onboarding paywall with a variation that includes a product trial. Or you can compare a monthly subscription with a trial period against a 3-month or annual subscription, and so on.
You can run an a/b test with multiple variations of paywalls (multivariate testing), each featuring different products and prices. The number of variations depends on the complexity of the changes you want to test and the resources available for analysis.
Before conducting the test, it is necessary to decide which users will participate - new, existing, all users, a specific segment (users from a particular country, or those who had a specific event such as trial cancellation or subscription cancellation), etc.
Alternatively, you can target an audience of users who have previously shown interest in our product but did not become paying customers, as we assume they might be more receptive to our product. In this case, when selecting the audience, specify "Non-paying users" or create a custom audience by choosing users who had the event Paywall payment canceled.
It is also important to determine the allocation size. The most common approach is to test 3-5 variations with an equal distribution of the sample. For example, 33% + 33% + 34% or 25% + 25% + 25% + 25%. Alternatively, you can allocate a larger percentage of users to the variation that is expected to be more successful in order to gather more data and obtain a more accurate estimation of the effect. You could try allocations like 70% + 30% or 80% + 20%.
You can increase the duration of the test to obtain more accurate results. However, it is important to consider that conditions may change over time, which can impact the test outcomes. For example, seasonal increases or decreases in purchases within your industry.
Let's consider the following example for calculating the sample size. On average, 1,000 new users visit the onboarding process per day. If we conduct the test for 1 month, each variation of the onboarding paywall will be seen by approximately 15,000 potential users. This sample size should be sufficient for making objective decisions.
It is recommended to conduct the a/b test for at least three weeks to account for differences in user behavior across different days of the week and seasonal factors.
Once you have set all the parameters of the experiment, launch the test and wait for it to complete. Some users may notice that in the early stages, one variation significantly differs from the others in all metrics, and in such cases, some beginners make the mistake of stopping the test before the planned duration. However, at the end of the experiment, it may be revealed that there is no actual effect or the test result is negative. Therefore, it is important to wait until the planned period is completed.
After the completion of the mobile a/b testing, the data is analyzed to assess the presence of a significant difference between the variants and determine the likelihood of randomness. Based on the analysis, a decision is made: if one of the variants shows statistically significant improvements compared to the other variants, a decision is made to implement the changes from that variant to all users.
When conducting an A/B test, it's important to remember that the conversion to trial or purchase is a short-term indicator of changes. While one variation may have a better conversion rate at the initial stage of testing, another variation may emerge as the winner in long-term metrics such as ARPU/ARPPU (Average Revenue Per User/Average Revenue Per Paying User).
Let's analyze an experiment using a specific example.
From the analysis of the metrics, we can see that Variation B has a higher conversion rate to purchase, but Variation A has a higher average revenue per user (ARPU) compared to Variation B.
Additionally, it's important to consider the cost per install (CPI) and compare it to the ARPU. If the ARPU is lower than the CPI, we recommend analyzing historical data by geography, generating new hypotheses, and creating new variations in price, a list of products on the paywall, and the appearance of the paywall for retesting. If the new variations do not result in an increase in ARPU, we suggest reevaluating the hypotheses and trying again. Ultimately, you will find the optimal budget and rate to maximize profitability.
In conclusion, mobile A/B testing is a powerful tool for understanding user preferences and behavior, as well as optimizing software development.
This tool is an integral part of modern software development and can provide a significant competitive advantage in the market.
Through experiments, the following metrics can be improved:
Additionally, experiments can provide insights into:
A/B tests enable developers to better understand their audience and offer products that best meet their needs. By using A/B testing, companies can make data-driven decisions that contribute to increased revenue and project success.
To get more insights regarding subscription app revenue growth read the Apphud Blog.