Every panel provider claims to deliver quality data. Fewer can define what that means in measurable terms. And almost none publish the numbers that would let buyers make an informed comparison. This piece is an attempt to change that, at least for the B2B segment where we operate.

The data below comes from three sources: our own study logs (n=1,840 studies, 2023–2024), a cross-provider comparison study we commissioned through an independent research firm (n=800 studies from eight providers), and publicly available quality data from the Insights Association's 2024 Panel Quality Report. Where we're describing our own performance, we'll say so explicitly.

Metric 1: Fraud rejection rate

This is the number most providers refuse to publish, because high rejection rates look like a product problem rather than a quality feature. The framing is backwards.

A 5% rejection rate doesn't mean 95% of your panel is high quality. It means the provider is running 5% of checks. Either their detection is weak, or they're not incentivized to reject—because rejected completions don't bill.

Provider Tier Avg Fraud Rejection Rate Range Checks Run
Mass-market panel providers5–9%2–14%3–6 checkpoints
Mid-market specialized providers12–18%8–24%8–12 checkpoints
Top-quartile providers20–28%15–35%14–20 checkpoints
QRCSurvey (2024 avg)34%22–51%23 checkpoints

The variation within each tier matters. A provider averaging 12% rejection rate might run 4% on a low-friction consumer study and 24% on a B2B executive study where fraud is higher. The average is less useful than the distribution—and the distribution requires access to study-level data that providers rarely share.

Metric 2: Demographic accuracy

B2B panel providers make demographic targeting claims that are hard to verify without post-hoc screening. We ran a validation study in Q3 2024: we purchased samples from four providers using identical quota specs (IT decision-makers, 500+ employee companies, North America and Western Europe), then ran a 12-question screener on the delivered data to verify actual job function, decision-making authority, and company size.

In our validation study, only 62% of "IT decision-makers" from the median provider actually had decision-making authority over technology purchases. That number was 91% for our own panel.

The gap isn't primarily about fraud—it's about profile verification. Self-reported titles are unreliable. "IT Manager" can mean someone who manages a team of 40 engineers or someone who manages the office printer. Without corroborating signals (firmographic data, professional registry matches, activity verification), you're buying titles, not roles.

Metric 3: Completion quality for open-ended questions

This is where the 2024 data shows the biggest year-over-year deterioration across the industry. AI-generated open-end responses have become materially more common in the past 18 months. They pass minimum-length filters. They're topically relevant. They feel coherent. And they contribute zero genuine signal.

In our cross-provider comparison study, we ran all open-end responses through an AI-generation likelihood model. The results:

The last point is important. Some providers track AI-generated response rates. Fewer actually exclude those responses from delivered data. We exclude them—which is why our delivered open-end acceptance rate runs lower than our total completion rate.

Metric 4: Quota accuracy and field pace

Two operational metrics that rarely appear in quality conversations but matter for research timelines: how accurately do providers hit your quota specs, and how predictably?

Quota accuracy—the percentage of delivered n that falls within your specified demographic targets—varied in our analysis from 84% (mass-market) to 97% (top-quartile). For a 500n study, that 13-percentage-point gap means 65 completions that don't fit your spec and need to be cleaned post-delivery. That's either a manual cleaning task or a distorted sample.

Field pace predictability matters for studies with hard deadlines. Providers who can't give you a reliable field timeline force you to build in buffer that costs money and extends project cycles. Our analysis found that top-quartile providers delivered within ±15% of projected timeline 91% of the time; mass-market providers hit that threshold 64% of the time.

What these benchmarks should change about your buying process

Four practical changes that follow from this data:

A note on what this analysis can't tell you

The benchmarks above describe population-level quality metrics. They don't predict what quality you'll get on any specific study. B2B fraud rates vary substantially by target profile: studies targeting C-suite executives in the US show different fraud patterns than studies targeting SMB owners in Southeast Asia. Market-specific quality data—not global averages—is what you actually need for study design decisions.

If you're running a study in a market or with a target profile where you have no baseline experience, the right first step is to ask your provider for market-specific rejection rate data for comparable studies. If they can't provide it, their quality monitoring isn't granular enough to be useful.

Want the full benchmark dataset?

The complete analysis—including market-level breakdowns and provider-anonymous comparisons—is available on request for qualified research teams.

Request Full Report