Oshi ResearchRESEARCH

Do loyalty programs actually work?

We went beyond the vanity metrics that every loyalty platform publishes and tested whether our program actually makes a difference.

About the Study

How we measured loyalty program effectiveness

We analyzed transaction data across seven merchants on the Oshi network: over 40,000 customers and tens of thousands of orders. These seven were selected because they had the largest customer bases, the longest program history, and enough data to produce reliable comparisons. They span food and drink, health and wellness, and specialty consumer goods.

At each merchant, customers who engage with the Bitcoin rewards program are classified as “enrolled.” Enrollment means the customer opted into marketing, received a reward notification, and actively claimed the reward. Depending on the merchant, roughly 19–38% of customers have done this. The rest either never engaged with the reward or opted out of marketing entirely.

That gives us two groups to compare at every store: customers who enrolled in the rewards program and customers who didn’t. That comparison is the starting point for every loyalty program case study in the industry. It’s also where most of them stop.

We didn’t stop there. Every loyalty platform publishes case studies showing that their members spend more, but none of them test whether the program actually drove that difference. We wanted to find out. So we applied six layers of analytical correction that no other platform uses and published everything.

To our knowledge, this is the most rigorous loyalty program effectiveness study published by any rewards platform. This is what we found.

This is an observational study. Customers who claim rewards self-selected into the program. Our analysis controls for lifecycle stage, calendar timing, and pre-enrollment spending, but does not claim that bitcoin rewards cause increased spending. This study measures the association between reward engagement and spending growth within a bitcoin rewards program. Whether the bitcoin element itself contributes to the effect, through novelty, appreciation potential, or attracting a different customer profile, is a question our data cannot answer separately.

1. The Vanity Metrics

We could tell you that our loyalty members spend 3x more. That they come back twice as often. That they generate over half of total revenue. And it would all be true.

It’s also exactly what every other loyalty platform will tell you. These are real figures from our own merchants, the same type of numbers you’ll see in any loyalty program case study:

3x

Higher Lifetime Value

members vs non-members

57%

of total revenue comes from loyalty members

Repeat Purchase Rate

66%

Members

27%

Non-Members

Purchases per Customer

3.7

Members

1.6

Non-Members

These are impressive numbers. And if we operated like other loyalty platforms, this is where we’d stop. We’d put these in a case study, slap on a client logo, and call it proof that our program drives results.

But these are vanity metrics. They describe a correlation between program membership and spending. They don’t tell you whether the program caused the spending, or whether high-spending customers were simply more likely to join. Your most engaged customers are the ones who sign up. They were already your best buyers. And because they shop more often, they see the reward offer more often, giving them more chances to enroll. The program gets credit for spending that most likely would have happened anyway.

So is the program actually changing behavior, or is it just identifying customers who were already great? The only way to know is to test it properly.

2. How We Tested It

Instead of comparing members to non-members, which compares fundamentally different groups of people, we compared the same customers to themselves. We looked at what each customer spent in the 90 days before they enrolled, then what they spent in the 90 days after. Same person, same store, different time periods.

But even a before-and-after comparison isn’t enough on its own. Customer spending naturally changes over time. Prices change, seasons change, promotions come and go. So we applied six layers of correction to isolate the signal from the noise.

Same customer, before and after

Each enrolled customer is their own baseline. We measure their spending in a 90-day window before enrollment and a 90-day window after. No cross-customer comparisons. The question is simple: did this individual customer spend more after enrolling than before?

A matched control group

Non-enrolled customers who reached the same purchase stage at the same time provide a natural baseline. If your third-time buyers typically increase spending by 10% regardless of any program, we need to know that. The control group tells us what would have happened without the reward.

The enrollment purchase is excluded

The reward is issued at the moment of purchase. That purchase cannot be influenced by a reward that was given during it. So we exclude it from both groups. Only purchases made after the first reward count toward the result.

Calendar-time matching

If a merchant raised prices, ran a holiday promotion, or changed their product mix, we need the control group to reflect those same conditions. We matched each enrolled customer’s control group to within 45 days of their enrollment date, so both groups are measured in the same market environment.

Pre-spend matching

Two customers at the same purchase stage can still have very different spending levels. A customer who spent $200 in the pre-period is not a fair comparison for one who spent $50. We added a spending similarity requirement so enrolled customers are compared to controls with similar pre-enrollment spending.

Pre-trend check

Were future claimers already on an upward spending trajectory before they claimed? If so, the post-enrollment lift might just be a continuation of a pre-existing trend. We checked. They weren’t. Both groups had similar spending trajectories leading into the enrollment moment.

No other loyalty platform applies any of these corrections. Most don’t even use a control group.

That doesn’t make their numbers wrong. It makes them incomplete. And if you’re making business decisions based on incomplete numbers, that can be expensive.

3. The Result

After every correction we could apply, the result was the same: positive. Enrolled customers showed meaningfully stronger spending growth than comparable non-enrolled customers at the same lifecycle stage and calendar period. After subtracting natural spending patterns from a matched control group, enrolled customers still spent significantly more.

$1$10+Every $1 spent on rewards was associated with $10 or more in additional customer spending

To put that in concrete terms: a merchant spending $500 per month on bitcoin rewards at a 1% rate would be associated with roughly $5,000 or more in additional customer spending beyond what similar non-enrolled customers generated naturally.

We ran five independent analytical methods with progressively stricter controls: calendar-matched, enriched matching, regression-adjusted, multi-merchant pooled, and a pre-trend falsification check. All five came back positive. Normalized to a standard 1% reward rate, the range across merchants was $5–$22 per $1.

Why the range varies

The $5–$22 range reflects real differences between merchants. A high-ticket brand where customers spend $200 per order will see a different dollar impact than a snack company with $50 average orders, even if the behavioral lift is similar. Purchase frequency, customer mix, and product category all affect the magnitude. The direction, however, was consistent everywhere we looked.

4. It Replicates Across Seven Merchants

A finding from one merchant could be a fluke. A finding from two could be a coincidence. But when the same pattern appears across all seven merchants, in different industries, with different customer bases and price points, it becomes very difficult to dismiss.

All seven merchants showed statistically significant positive results.

Estimated Incremental Revenue Per $1 of Rewards (at 1% rate)

Merchant AFood & drink
$22.14p = 0.003
Merchant BHealth & wellness
$21.60p < 0.001
Merchant CHealth & wellness
$17.71p < 0.001
Merchant DSkincare
$14.22p < 0.001
Merchant EFood & drink
$11.17p = 0.014
Merchant FFood & drink
$10.38p = 0.006
Merchant GFood & drink
$5.24p = 0.030

All figures normalized to a 1% reward rate for apples-to-apples comparison.

The seven merchants span food and wellness, health and nutrition, snacks, beverages, meat and provisions, and specialty food. Different products, different customers, different purchase cadences. Same positive direction.

The Fine Print

What we can and cannot claim

Proving that a loyalty program causes customers to spend more is genuinely difficult. The gold standard would be a randomized experiment where some customers are offered rewards and others are not, without either group knowing. That kind of study is extremely difficult to run in a real e-commerce environment, and to our knowledge, no loyalty platform has ever published one.

That means no one in this industry can claim causation with certainty. Not us, not any other platform. The difference is that most platforms don’t acknowledge this. They present vanity metrics as if they’re proof and move on.

What we can say is that we went as far as the data allows. After controlling for lifecycle stage, calendar timing, pre-enrollment spending, and multiple other factors, the association between reward engagement and subsequent spending growth remained positive across every method and every merchant we tested. We can’t prove causation, but we can show a strong, consistent signal that survives serious scrutiny.

What we can say

  • Enrolled customers showed stronger spending growth than matched non-enrolled customers
  • The result survived five independent analytical methods
  • It replicated across all seven merchants
  • No evidence of pre-existing trends explains the result

What we cannot say

  • The reward caused the spending change
  • Every enrolled customer benefits equally
  • These results generalize to all merchants

The Opportunity

What this means for your business

Most merchants evaluate loyalty programs based on vanity metrics because that’s all anyone gives them. The platforms don’t dig deeper, so the merchants can’t either. That means investment decisions are being made on numbers that look impressive but don’t hold up to scrutiny.

The merchants in our study spend 1–5% of purchase value on bitcoin rewards. At a 1% rate, that’s $1 for every $100 a customer spends. Even at the low end of our estimates, that’s a compelling return for a program that runs passively in the background.

More than retention

This study measures spending behavior among existing customers. But there’s a separate dimension worth considering: customer acquisition. Bitcoin enthusiasts are some of the most valuable customers in e-commerce. Our separate research found that they generate nearly 3x the lifetime value of other customers at the same stores. A Bitcoin rewards program doesn’t just retain the customers you have. It makes your store discoverable to a passionate, high-value community that most merchants aren’t reaching at all.

Why now

The number of merchants offering Bitcoin rewards is still small. Most stores are running the same traditional points programs that have been around for decades, backed by the same unverified metrics. That creates an opening for merchants willing to try something different and hold themselves to a higher standard of evidence.

As Bitcoin adoption continues to grow, more merchants will eventually move in this direction. The ones who start now build relationships with this customer base first. They establish trust, earn loyalty, and show up in rewards networks before the space gets crowded.

Oshi connects to your existing store and lets you offer Bitcoin rewards to every customer. No technical setup. No minimum commitment. Just better data and a better rewards experience.

FAQ

Frequently Asked Questions

Most loyalty platforms compare "members" to "non-members," groups that differ before the program exists. We compare the same customers to themselves, before and after enrollment, with a matched control group to account for natural spending changes. We then apply five additional corrections that no other platform uses.

The methodology applies to any loyalty reward. Bitcoin rewards have the additional property that customers can hold and grow the value of their reward over time, which may strengthen the retention signal. But this study measures spending behavior, not reward appreciation.

No. That would require a randomized experiment. What we can show is a strong, consistent positive association that survives every correction we applied, replicated across all seven merchants.

We selected seven merchants based on data availability: those with enough customers, enrollment history, and program runtime to produce statistically meaningful results. Every qualifying merchant is included, regardless of outcome.

Yes. An independent statistical analyst reviewed the methodology, suggested improvements which we implemented, and confirmed the analysis is suitable for publication.

Over 40,000 direct-channel customers across seven merchants. The pre/post analysis specifically covers enrolled customers with measurable pre-enrollment purchasing history. First-time enrollees are excluded because they have no baseline for comparison.

Merchants in our study typically set reward rates between 1% and 5% of purchase value. At a 1% rate, a $100 order costs the merchant $1 in Bitcoin rewards. There is no platform fee to use Oshi.

No. Customers receive a simple email notification when they earn a reward. They can claim it with one click. Many customers treat it like any other loyalty reward without thinking about the Bitcoin element at all.

Most loyalty platforms compare members to non-members and present the difference as proof that the program works. That comparison is misleading because engaged customers self-select into loyalty programs. Our study compares the same customers before and after enrollment, with a matched control group, calendar-time matching, pre-spend matching, and pre-trend checks. No other platform applies these corrections.

Our study covers seven merchants across food, health and wellness, and specialty goods. The results were consistent across all seven, but every business is different. We run the same analysis for every Oshi merchant so you can see your own data, not just an industry average.

Methodology

Data: Transaction-level records from merchant databases. All revenue in USD. Over 40,000 direct-channel customers across seven merchants.

Merchant selection: Seven merchants selected based on data sufficiency: those with enough customers, enrollment history, and program runtime for reliable analysis.

Marketplace exclusion: Customers who purchased through Amazon, Faire, TikTok Shop, or other marketplace integrations were excluded. These customers never had the opportunity to interact with the rewards program.

Outlier handling: Individual transactions above a per-merchant cap (Tukey method, $2,000 floor) were excluded from spend calculations to prevent single large orders from distorting results.

Treatment group: Customers who claimed at least one bitcoin reward, enrolled at least 90 days ago, with measurable spending in the 90 days before enrollment.

Control group: Opted-in customers who received the reward offer but never claimed. Never opted out of marketing. At the same purchase lifecycle stage and calendar period as treatment customers.

Matching: Three-dimensional matching on visit bucket (lifecycle stage), calendar period (±45 days), and pre-period spending similarity (±50%).

Anchor exclusion: The enrollment purchase (the transaction at which the first reward was issued) is excluded from both treatment and control post-windows. This purchase could not have been influenced by a reward issued at that same moment.

Inference: Per-customer paired effects with 10,000-resample nonparametric bootstrap. 95% confidence intervals from empirical percentiles.

Pre-trend check: Compared spending trajectories in the 180–90 day period versus the 90–0 day period before the anchor. No evidence that future claimers were already accelerating relative to matched controls.

Regression check: ANCOVA model controlling for pre-spend, pre-orders, prior AOV, recency, tenure, and visit bucket. Treatment coefficient remained positive and statistically significant.

External review: Full methodology reviewed by an independent statistical analyst. Reviewer feedback incorporated into the final design.

Calculation audit: Both headline figures ($55.75 calendar-matched, $37.91 enriched) were reproduced end-to-end from raw transaction data in an independent verification script.

Ready to see real data, not marketing metrics?

We run the same rigorous analysis for every Oshi merchant. No inflated numbers. No vanity metrics. Just what your data actually shows.