Experimentation: Testing and validation#
In this section, we discuss a process for running experiments to test hypotheses on our business model canvas.
Experimentation#
Why experiment? We experiment to gain insights about customers, their problems, our solutions, and any other important assumptions we make about our product.
There are two types of experiment:
Generative/exploratory: These experiments help us to explore a space.
Evaluative: These are hypothesis-driven and aim to find data to support or refute a hypothesis.
There are two reasons that we experiment:
To evaluate our understanding of our customers and market fit.
To evaluate products.
The process of experimentation is summed up with this excellent process model from Steve Blank:

Fig. 18 The process model for experimentation. Right click and open in new tab to see a larger version. Adapted from: https://steveblank.com/2014/10/24/17577/#
Step 1: Extract the hypotheses from your business model canvas (or define them if you haven’t yet!)
Step 2: Prioritise the hypotheses from what you believe are the most valuable to the least valuable. Based your prioritisation on things like:
The risk to your product, where risk is probability times impact. If the impact on your product of the hypothesis being wrong is low and the probability of it being right is high, it is lower priority. If the impact is high and the probability low or unclear, then it is higher priority.
How much of a problem the hypothesis is for customers.
How much the customers would be willing to pay for the problem to be solved.
Step 3: Design tests to gather data for your hypotheses and capture these as test cards.
Step 4: Prioritise your tests. Base your prioritisation on things like:
The priority of the hypothesis.
How easily you can test the hypothesis.
Step 5: Run the tests!
Step 6: Capture the learnings in learning cards.
Step 7: Decide what to do based on what you have learnt.
Them repeat!
Hypotheses#
At this point, our canvases are full of hypotheses, not facts (yet!).
What but exactly is a hypothesis?
Definition – Hypothesis
In the context of innovation, a hypothesis is a falsifiable statement or prediction about a business idea, product, or customer that can be investigated through empirical observation and experimentation.
This is very similar to the definition of “hypothesis” in science.
Note this important property: falsifiable. This means that we must write the hypothesis in such a way that we are able to demonstrate it is false using empirical data.
“In so far as a scientific statement speaks about reality, it must be falsifiable: and in so far as it is not falsifiable, it does not speak about reality.” – Karl Popper, The Logic of Scientific Discovery.
Example – Non-falsifiable hypothesis
“This solution will help this segment of customers improve their health over the long term.”
This is not falsifiable for three (well, two and a half) reasons:
The timeframe “over the long term” is not specific enough to measure. If we find it has had no impact within 10 years, anyone could claim “10 years is not long term”.
The group “this segment” is not specific enough. If it improved 95% of customers health within a year, is this hypothesis supported or refuted? 95% is a lot, but the hypothesis states “this segment”, which could be intepreted as “everyone in this segment” or perhaps “a majority of people in this segement”.
It is not testable right now because it talks about the long term. If we mean 10 years, we cannot really test it for 10 years.
Example – Falsifiable hypothesis
“We believe that adding an inventory management function to this product will achieve a customer retention rate of 75% over the first year”
This is falsifiable:
It specifies what will happen: 75% retention rate.
This is measurable: we can measure retention by seeing which customers at the start of the year are still there at the end of the year.
The timeframe is clear: the first year.
Framing our assumptions like this treats innovation as a scientific process. Just like designing a scientific experiment, we design an experiment to test our hypotheses; and further, the aim of the experiment should be to try to prove that the hypothesis is wrong!
“The point is that, whenever we propose a solution to a problem, we ought to try as hard as we can to overthrow our solution, rather than defend it. Few of us, unfortunately, practice this precept; but other people, fortunately, will supply the criticism for us if we fail to supply it ourselves.” – Karl Popper, The Logic of Scientific Discovery.
Test cards#
Test cards, a product of Strategyzer, are a good way to provide some structure to the hypotheses and experiments that we run. The format is:

Fig. 19 The structure of a test card. Source: https://www.strategyzer.com/#
Each card has five parts:
“We believe that … “ is the hypothesis that we have — the assumption that we have about a customer segment, their problem, etc. – anything on the business model canvas!
“To verify that, we will … “ what we will do to test the hypothesis, which almost always involves finding data; e.g. talking to customers, finding existing data, etc.
“And measure … “ specifies exactly what we want to measure; e.g. number of customers who say X, or number of sales of existing product Y.
“We are right if …” specifies the criteria for either accepting or rejecting the hypothesis.
“Therefore, we will ask …” defines what interview questions we will ask our interviewees to test the hypothesis.
This final part (interview questions) is not typical to see on test cards, such as those used by Strategyzer. It was an innovation that one of the early COMP1100 teams added to their test cards that I thought brought value, so I adopted it.
Note
Note that we often gather data from sources other than interviews, but if we cannot, it helps to have a clear set of initial questions to ask.
Structuring hypotheses#
A good hypothesis statement (the “We believe that …” statement) should identify a target, a behaviour or observable phenomenon, and a reason for the behaviour/phenomenon:
We believe that [target] are [showing behaviour / display interest in ] [for reason Y].
Structuring verification statements#
A good verification statement (the “To verify that …” statement) should describe a particular behaviour that we will undertake to collect the right data to test the hypothesis:
To verify that, we will [undertake behaviour] to [identify behaviour].
Structuring measurement statements#
A good measurement statement (the “And measure …” statement) should identify exactly what data will be measured:
And measure [particular variable].
Structuring accept/reject statements#
A good accept/reject statement (the “We are right if …”) should have a clear and measurable statement that makes it easy to determine whether the hypothesis is accepted or rejected:
We are right if [some measurable statement].
Structuring interview question statements#
A good set of questions (the “Therefore, we will ask …” statement) should identify the questions that we will need answers to, in order to gather the data
And ask [specific questions].
We can see here that these statements make it clear to determine whether to accept or reject the hypothesis: we count the number of people who said something, and then calculate the percentage.
Steve Blank states that we should make bets with ourselves (or with each other if in teams) to help make the success/failure definition clear. By framing it as a bet, it helps us to make these concrete.
Example – Medical alert system test card: users
We believe that our end users are people over 65 who live on their own who would use technology to keep in touch with their family more often because they do not see them much.
To verify that, we will interview 50 people over 65 who live on their own.
And measure how many would like more regular digital interaction with their family.
We are right if 70% of older people say that they would like more regular digital interaction with their family.
Therefore, we will ask “how often would you like to interact with your family members?”
We can see above that we don’t need to use an actual card: we can just use standard text and Markdown to record these. The format of the card is what helps us to structure our thinking and to document our findings.
Learning cards for capture learnings#
Similar to test cards, we can structure the outcomes from our experiments in a format called learning cards:

Fig. 20 The structure of a learning card. Source: https://www.strategyzer.com/#
The structure for these is pretty clear: it reflects the language in the test card.
Example – Medical alert system learning card: users
We believed that our end users are people over 65 who live on their own who would like interact digitally with their family more often because they do not see them much.
We observed that 44 out of 50 people we interview confirmed that they would like more regular digital interaction with their family.
From this we learnt that our end users would consider using our product.
Therefore, we will accept this hypothesis.
Note in the above example that at some point during the interviews, some additional data (about using smartphones) was found. We haven’t yet tested this hypothesis formally – for now, it becomes a new hypothesis (which we eventually did test).
Types of experiments#
Experiments gather evidence for transforming hypotheses into facts – or into refuted hypotheses.
There are several ways to collect evidence:

Fig. 21 Types of experiments#
Tips for designing and executing experiments#
Experiments should:
Be as small and comprehensible as possible
Have clear expected outcomes.
Have a clear definition of success/failure, which everyone agrees to.
Be as brief as possible when interviewing to collect data; e.g. preferably less than 30 minutes.
We can test multiple hypotheses at once; e.g. in a single interview or prototype session, but we must do this carefully. We really want to focus on just a few things in each experiment.
We need to prepare for the experiment: research the market and other competitors, research a business (if they are a customer), learn about people in the segment from public data, find out a bit about an interviewee, etc.
Remember that we should:
Listen!
Observe people’s behaviour, tone, and reactions, not just what they say.
Ask “why” and “why not”
Ask who else we should talk to.
Ask if we can follow up with them later.
Remember that we should NOT:
“Correct” people (except to clarify if they have misunderstood a question).
Convince people or sell to them.
Learning from experiments#
Once we have executed an experiment, we need to gain insights.
If hypotheses have been supported, then great!
If not, what insights can we gain?
If people would not pay for this, what would they pay for?
If these are not our customers, then who is?
Did we see any outliers that could open up new ideas?
Pay attention to why our hypotheses are rejected.
What type of insights would we gain? It can be insights about:
Our target market.
Our value propositions.
Our market fit.
Our product features.
(sec:experimentation:accept_pivot_or_refine)
Accept, pivot or refine#
Once we have gained some insights, recall that there are a few things we can do for each hypothesis that we have:
accept the hypothesis as a fact;
refine our hypothesis; or
pivot our product entirely — for cases where it appears our value proposition does not have a good market fit.
Recall Eric Ries’s model for innovation from Fig. 5, re-produced below:

In cases where a pivot is required, this typically indicates a setback – we have rejected some hypotheses and have concluded that we do not have a market fit.
We can react to a setback in three ways:
Ignore what our testing showed us and press on with our idea. This is commonly done; but also, note that businesses and product commonly fail, and this is a key reason why. Why do people press on despite the facts telling them they are wrong? Because it is difficult to give up on an idea that we have invested so much time and effort on.
We can refine our product ideas. Typically, we can have a good understanding of our customers’ problems, but our product idea did not quite solve their problem; or at least, not enough that they would buy it or invest in it. However, we can fine-tune our product ideas, prototype these, and test them again.
We can pivot our strategy. This happens when we have fundamentally misunderstood what our customers’ problems were or what we thought could help solve them. Again, we use what we learnt from our test to improve our understanding and come up with new ideas.
Ries[1] notes that it is unusual for organisations to change their vision – this is usually the dream to which they are committed, and often the insights gained from data are sufficient.
Takeaways#
Takeaways
Experiments help us gain insights about our customers and products by testing our hypotheses.
Hypotheses must be falsifiable to be of value.
Test cards are a useful tool to structure our hypotheses.
Learning cards are a useful tool to structure our learnings from the data.
If a hypothesis is refuted, we can still gain insights by asking “why?” and paying attention to the data we gathered.