Expert Guide Series

What Should I Do When My Research Shows Conflicting Results?

Research conflicts happen on nearly every app project I've worked on, and honestly? They're one of the most frustrating parts of the discovery phase. You'll run user interviews where people tell you they desperately want feature A, then your analytics show nobody actually uses feature A when its available. Or you'll get survey results pointing in one direction whilst your competitor analysis suggests something completely different. It's a bit mad really—you're trying to make informed decisions about your app but the information itself seems to be fighting against you.

I remember working on a healthcare app where we had this exact problem; our focus groups kept saying they wanted detailed health tracking with graphs and data exports, but when we looked at similar apps in the market the most successful ones were dead simple with minimal features. The qualitative data said "give us more" whilst the quantitative market data screamed "keep it simple." We were stuck because both data sources seemed legitimate and well-researched, yet they contradicted each other completely. This wasn't a case of bad research—it was a case of needing to understand why the conflict existed in the first place.

The truth is that conflicting research data doesn't mean you've done something wrong; it usually means you're asking complex questions about human behaviour, and humans are messy and unpredictable.

What I've learned over the years is that these conflicts often contain the most valuable insights for your app. The contradiction itself is telling you something about your users or your market that wouldn't be obvious from clean, consistent data. The challenge isn't making the conflict disappear—its learning how to interpret what the conflict actually means and using that understanding to make better decisions for your mobile app.

Understanding Why Research Conflicts Happen

I've lost count of how many times I've sat in meetings where everyone's looking at research results that completely contradict each other. Its frustrating, I know, but here's the thing—conflicting research isn't a sign you've done something wrong; it's actually pretty normal when you're dealing with real people and complex behaviours. In fact, I'd be more worried if all your research lined up perfectly because that usually means you're only asking surface-level questions.

The main reason conflicts happen is because different research methods measure different things. When we ran user interviews for a healthcare app, patients told us they wanted detailed medical information and lots of features. Made sense, right? But when we looked at the analytics from the prototype, people were abandoning the app because there was too much information. They said they wanted depth, but their behaviour showed they actually needed simplicity. This happens all the time—what people say they want and what they actually use can be worlds apart.

Another big source of conflict comes from who you're asking and when you're asking them. Your research might show different results simply because you surveyed different user segments or caught people at different points in their journey. I worked on an e-commerce app where new users and existing customers had completely opposite opinions about the checkout process. New users wanted more guidance, existing ones wanted fewer steps... both were right for their situation.

Common Sources of Research Conflicts

  • Different research methods capturing different aspects of user behaviour (qualitative vs quantitative)
  • Timing differences—users change their minds or needs shift over time
  • Sample size problems where small groups don't represent your full user base
  • Environmental factors like testing in a lab versus real-world usage
  • Leading questions or biased research design that influences responses
  • Multiple user types with genuinely different needs using the same app

The thing about research conflicts is they often reveal something important about your users that you wouldn't have discovered otherwise. When a fintech client's A/B test results contradicted their focus group findings, we dug deeper and realised we were actually dealing with two distinct user personas who needed different experiences. That conflict led us to build a smarter onboarding flow that adapted to user type... and retention improved by 34%.

Checking Your Research Methods First

Before you start panicking about conflicting results, you need to look at how you collected your data in the first place. I've seen dozens of projects where the "conflict" wasn't actually about user behaviour or market trends—it was about dodgy research methods. And honestly, this happens more often than people want to admit.

The biggest mistake I see? Mixing research methods that measure completely different things. If you're running a survey asking users what features they want, then looking at analytics showing which features they actually use, those results are going to conflict. They're measuring intent versus behaviour, and those are two very different things. I worked on a fitness app where survey respondents said they'd use workout planning tools daily, but our analytics showed most people just tracked their runs and nothing else. The data wasn't conflicting—we were just asking the wrong questions.

Common Research Method Problems

Here's what usually causes false conflicts in app research:

  • Sample sizes that are too small or not representative of your actual user base
  • Leading questions in surveys that push people toward certain answers
  • Testing with friends and family who won't give honest feedback
  • Comparing data from different time periods without accounting for seasonal changes
  • Using analytics tools incorrectly or tracking the wrong events
  • Not segmenting your users properly, so you're lumping different behaviours together

Quick Ways to Check Your Methods

Start by documenting exactly how you collected each piece of data. What questions did you ask? Who did you ask them to? When did you collect it? I know this sounds boring, but its the only way to spot where things might have gone wrong. On a fintech project, we found our "conflicting" data about payment preferences was actually collected from two completely different user segments—one group was mostly under 30, the other mostly over 50. Of course their preferences didn't match up.

Write down your research method for each data source before comparing results. If you cant explain exactly how you got each piece of data, you probably shouldn't be comparing them at all.

Look at your sample sizes too. If one survey has 500 responses and another has 12, they don't carry equal weight. This seems obvious, but people ignore it all the time when they're desperate for insights. The smaller sample might be showing you an outlier, not a trend.

Spotting Good Data From Bad Data

The first thing I look at when reviewing research data is sample size—and I mean really look at it, not just glance at a number. I worked on a healthcare app where the initial user testing showed people loved a particular feature, but when I dug into the data I realised they'd only tested with 8 users. All from the same age group. All recruited from the same private clinic. That's not data you can trust; its just a snapshot of a very specific group's opinion that might not represent your actual user base at all.

Good data comes from diverse sources and reasonably sized groups. For consumer apps I generally want at least 30-50 users per segment before I start making firm decisions, though that number shifts depending on what you're testing. But here's the thing—quality matters more than quantity sometimes. I'd rather have 40 users who match my target demographic perfectly than 200 random people who downloaded the app by accident.

Red Flags That Your Data Might Be Dodgy

Over the years I've learned to spot certain warning signs that suggest the data in front of me isn't reliable. These aren't always dealbreakers but they should make you pause and dig deeper before you base major decisions on that information.

  • Leading questions in surveys that push users toward a particular answer
  • Research conducted by people with a vested interest in a specific outcome
  • Data thats too perfect or matches expectations exactly (real user behaviour is messy)
  • Missing context about how, when, or where the data was collected
  • Conflicting methodology between different data sets you're comparing
  • Self-reported data without any behavioural validation

I built an e-commerce app where the client insisted their internal survey showed users wanted more product categories, but when we looked at actual usage data the existing categories were barely being used. The survey questions had basically asked "would you like more options?" and surprise surprise, everyone said yes. Good data shows you what people actually do, not just what they say they want when you ask them directly.

When Your Users Say One Thing But Do Another

This is probably the most frustrating thing you'll encounter during research and honestly, it happens more often than you'd think. People will sit in an interview and tell you they desperately want feature X, then when you build it and watch the actual usage data... crickets. Nobody uses it. I've seen this play out dozens of times across different projects and its always a bit mad how consistent this pattern is.

The gap between stated preference and revealed preference is real—people genuinely believe what they're telling you in that moment, but their actual behaviour tells a different story. I worked on a fitness app where users told us they wanted detailed nutrition tracking with calorie counting and macro breakdowns. We built it. Barely 8% of users who requested it actually logged their meals more than three times. What they actually used? The simple workout timer and progress photos. Their behaviour showed us what they really valued, even though their words said something completely different.

What people say they want and what they actually use when nobody's watching are often two completely different things

The fix here is pretty straightforward but requires patience. You need to observe behaviour, not just collect opinions. When we're validating features now, I always push for behaviour-based testing—things like clickable prototypes where we can see what people actually interact with, or soft launches where we measure real usage patterns. For a banking app we developed, users said they wanted investment portfolio tracking, but when we added a simple spending breakdown feature almost as an afterthought, that became the most-used feature in the entire app. Users actions don't lie, even when their words do.

Making Sense of Numbers That Don't Match Up

I've lost count of how many times a client has come to me with two sets of data that completely contradict each other—and honestly, its more common than you'd think. Maybe your analytics show that 80% of users drop off at the registration screen, but your heat mapping tool says people are actively engaging with that exact page for an average of 3 minutes. Or your A/B test declares version A the winner with 95% confidence, but your user feedback is overwhelmingly in favour of version B. This stuff happens all the time, and learning to work through it is what separates good product decisions from bad ones.

The first thing I do when numbers don't match up is check whether we're actually measuring the same thing. Sounds obvious, right? But you'd be surprised how often this is the problem. I worked on a healthcare app where our backend logs showed 10,000 daily active users, but Firebase Analytics reported only 7,500. Turns out the backend was counting every API call (including automated health checks) while Firebase only tracked actual user sessions. We weren't comparing like with like at all.

Common Reasons Your Data Conflicts

Here are the most frequent culprits I've encountered when numbers don't line up:

  • Different tracking implementations—one tool fires on page load, another on user interaction
  • Time zone mismatches between your analytics platforms (this one catches people out constantly)
  • Sample sizes that are too small to be meaningful, especially in A/B tests
  • Cached data vs real-time data showing different snapshots
  • Users with ad blockers or privacy settings that prevent certain tracking scripts
  • Bot traffic inflating some metrics but not others

Another thing I've learned is that quantitative and qualitative data often conflict because they're answering different questions. On an e-commerce app, we saw that users spent 40% more time on product pages with video demos. Great, right? But in user interviews, people said the videos were "too long" and "annoying". The numbers showed engagement, but what users actually meant was they felt forced to watch the video to get the information that should have been in the description. The metrics weren't lying—we just weren't interpreting them correctly.

What to Do When You Can't Resolve the Conflict

Sometimes you'll do all the checks and the data still won't match up perfectly. That's when you need to decide which source is more reliable for your specific question. For a fintech app we built, we trusted transaction logs over user-reported data when it came to feature usage because people genuinely couldn't remember which payment method they'd used last week. But for understanding why people abandoned transactions? User interviews were gold, even though they contradicted what we initially thought from the numbers alone. Context matters more than perfect alignment, and being transparent about data limitations in your decision-making process is way better than pretending everything lines up neatly when it doesn't.

Testing Your Assumptions in the Real World

Look, you can spend months pouring over research data and trying to figure out which numbers tell the truth—but honestly? There's only one way to know for sure what works and that's putting it in front of actual users. I've built apps where the research pointed in one direction, we tested it with real people, and everything we thought we knew went out the window. It happens more than you'd think.

The way I approach this is through what I call focused prototype testing. You don't need a fully built app to test your assumptions; in fact, building too much before testing is one of the biggest mistakes I see. For a healthcare booking app we developed, research showed users wanted lots of filtering options for finding doctors—specialty, location, availability, insurance, the works. But when we put a clickable prototype in front of fifteen people and watched them try to book an appointment? They got overwhelmed and most just wanted to search by name or location. That's it. All those fancy filters we nearly spent weeks building would have been wasted effort.

Start small with your tests. Pick the assumption that has the most conflicting data and test just that one thing first. Use tools like InVision or Figma to create interactive mockups that feel real enough for users to react naturally. Then watch what they actually do, not what they say they'll do—theres a massive difference and I cannot stress that enough.

What to Test First

  • The core user flow that your conflicting data disagrees about most
  • Navigation patterns if users' stated preferences don't match usage data
  • Feature priority when surveys and analytics tell different stories
  • Onboarding steps where drop-off rates contradict user feedback

Test with 5-8 users at a time; you'll spot most usability issues by the fifth person and can iterate quickly without burning through your budget or timeline.

The beautiful thing about real-world testing is it cuts through all the noise. Numbers can be interpreted different ways, surveys can be misleading, but watching someone struggle with a button placement? That tells you everything you need to know right there.

Deciding Which Data Actually Matters

Right, so you've got conflicting research results and you need to actually make a decision. This is where years of shipping apps teaches you something that no research methodology handbook ever will—some data is just more valuable than the rest, and knowing which data to trust comes down to understanding what you're really trying to achieve.

I've learned (sometimes the hard way, honestly) that behavioural data almost always trumps what people say in interviews or surveys. When we built a fintech app, users told us they wanted detailed transaction histories front and centre on the home screen. Made sense, right? But when we tracked actual usage, most people barely looked at their full history—they just wanted to see their current balance and recent charges. The data from real behaviour showed us what users actually needed, not what they thought they wanted.

Here's how I prioritise conflicting data when making decisions:

The Data Priority Framework

  • Actual user behaviour in your existing app or prototype—this is gold because it shows what people do when nobody's watching
  • Usability test observations where you see people struggle or succeed with specific tasks—actions speak louder than words
  • Quantitative data from large sample sizes—the bigger your dataset, the more confident you can be about patterns
  • Feedback from your target users specifically—not just anyone who'll take your survey, but actual potential customers
  • Industry benchmarks and competitor analysis—useful context but shouldn't override your own user data
  • Stakeholder opinions and internal assumptions—these matter for business reasons but are least reliable for predicting user behaviour

The trick is matching your data to your specific decision. If you're trying to decide on feature priority, look at engagement metrics and user requests weighted by frequency. If its a design question? Usability test results and heatmaps will tell you more than any survey. When we worked on a healthcare app, we had conflicting data about whether users wanted appointment reminders via push notification or SMS—the analytics showed push notifications had terrible open rates, but SMS had higher costs. We let the business constraint (budget) and the behavioural data (actual opens) make the decision for us.

Building Confidence When Things Stay Unclear

Here's what I've learned after launching dozens of apps into markets where the research never fully lined up: you don't need perfect clarity to move forward. You just need enough confidence to take the next step. I worked on a healthcare booking app where our survey data said users wanted comprehensive doctor profiles with loads of information, but our analytics from a competitor's app showed people spent an average of 12 seconds on profile pages before booking. Completely contradictory, right? We couldn't resolve it through more research because both findings were valid in their own context.

The solution wasn't to keep digging for the "truth"—it was to build in phases with clear success metrics. We launched with medium-length profiles and instrumented everything to see what actually happened with our specific users. Turns out they wanted different amounts of info depending on whether it was a routine appointment or something more serious. The conflicting data was both right and both wrong, it just depended on the use case.

The goal isn't to eliminate all uncertainty before you start building; its to reduce uncertainty to a level where the risk of moving forward is smaller than the risk of standing still

Set yourself a decision threshold. I usually tell clients we need about 70% confidence to start development on a feature. Not 100%. Not 50%. Somewhere in that range where you've done your homework but you're not paralysed by contradictions. Document what you're uncertain about, build those areas to be flexible, and plan your first release as a learning exercise. Some things will only become clear once real users have your app in their hands, and that's completely normal in this industry.

Conclusion

Building apps when your research shows conflicting results isn't a problem to solve—its just part of the job really. I've shipped dozens of apps over the years and I can count on one hand the number of times all the data pointed in exactly the same direction. Usually there's some tension between what users say they want, what the analytics show they actually do, and what makes business sense to build. That's normal.

The key thing I've learned is that you don't need perfect clarity to move forward; you just need enough confidence to take the next step. Start with the data that comes from real behaviour rather than stated preferences, prioritise the findings that align with your core business goals, and be honest about what you simply don't know yet. Build small, test fast, and let actual usage tell you what to do next. I mean, we built a healthcare appointment booking feature based on conflicting survey data—half our users wanted calendar integration, the other half found it too complicated—so we launched with a simple toggle that let people choose. Turns out 80% enabled it within their first week, which settled the debate pretty quickly.

When research conflicts, you're not stuck... you're just at a decision point. Use what you know, acknowledge what you don't, and remember that shipping something you can learn from beats endlessly debating which dataset is "right." The best apps aren't built by people who had all the answers upfront—they're built by teams who knew how to make smart decisions with incomplete information and weren't afraid to adjust course when real users showed them a better way.

Frequently Asked Questions

How can I tell if my research data is reliable or just poorly collected?

Start by documenting exactly how you collected each piece of data—what questions you asked, who you asked, and when. If you can't explain your methodology clearly, or if your sample sizes are tiny (under 30 users per segment for most consumer apps), then you shouldn't trust those results for major decisions.

What should I do when users say they want one thing but the analytics show they use something completely different?

Trust the behavioural data over stated preferences—people genuinely believe what they tell you, but their actions reveal what they actually value. I've seen this countless times where users request complex features in interviews but then only use the simple ones when the app launches.

How do I know which conflicting data source to prioritise when making product decisions?

Actual user behaviour in your app or prototype trumps everything else, followed by usability test observations and large-sample quantitative data. Survey responses and stakeholder opinions are useful context but are the least reliable for predicting real user behaviour.

Is it normal to have conflicting research results, or does it mean I've made mistakes?

Conflicting research is completely normal when you're asking complex questions about human behaviour—I'd actually be more worried if all your data lined up perfectly. The contradictions often contain the most valuable insights about your users that you wouldn't discover from clean, consistent data.

How much certainty do I need before I can start building features with conflicting research?

You need about 70% confidence to move forward—not perfect clarity, but enough to reduce risk below the cost of inaction. Document what you're uncertain about, build those areas to be flexible, and plan your first release as a learning exercise rather than a final solution.

Should I do more research to resolve conflicts, or just pick one data source and move on?

Sometimes more research helps, but often you need to test assumptions with real users through clickable prototypes or soft launches. I've found that watching 5-8 people interact with a prototype cuts through conflicting survey data and analytics faster than collecting more opinions.

What's the fastest way to validate which version of conflicting research is actually correct?

Build focused prototypes testing just the conflicting element and observe real user behaviour—not what they say they'll do. For a healthcare app, users told us they wanted complex filtering but when we watched them use a prototype, they got overwhelmed and just searched by location.

How do I present conflicting research findings to stakeholders without losing credibility?

Be transparent about the conflicts and explain what each data source tells you, then recommend a testing approach to resolve the key uncertainties. Stakeholders respect honesty about limitations more than pretending everything aligns perfectly when it clearly doesn't.

Subscribe To Our Learning Centre