The mental model most researchers have of survey fraud is wrong. It's not primarily bots clicking through at machine speed, producing obviously nonsensical responses with sub-second completion times. That type of fraud is easy to catch and has largely been eliminated from reputable panel networks.
What's harder to catch—and what's gotten significantly more sophisticated in the past three years—is organized human fraud: coordinated groups of real people who know how to pass quality checks, understand what screeners are looking for, and have economic incentives to participate in every survey they can reach.
The anatomy of a modern fraud ring
The basic operating structure is straightforward. A fraud ring is a coordinated group—ranging from a few dozen to several thousand individuals—who share techniques for passing panel screenings and maximizing survey completion rates across multiple platforms simultaneously. They're typically organized through private messaging channels (Telegram groups being currently most common) and incentivized through revenue sharing on completed surveys.
What makes them hard to detect isn't their behavior within any single survey. It's the pattern across surveys and platforms. Individually, each participant looks like a legitimate if somewhat unremarkable respondent. Collectively, they share:
- Common screener strategies (how to pass specific types of qualification questions)
- VPN services optimized to pass geo-IP checks while spoofing location data
- Shared device configurations that defeat cookie-based deduplication
- Response templates for common open-ended question formats
- Timing guidelines to avoid speeding flags while maintaining high throughput
// Scale context
In 2024, the Insights Association estimated that organized fraud rings account for approximately 21% of all fraudulent completions in online panels—up from 9% in 2020. The majority of the remaining fraudulent completions come from disengaged legitimate panel members (random responding, straight-lining) rather than intentional fraud. Both types require detection, but they require different detection approaches.
Why standard quality checks fail against organized fraud
Standard quality checks were designed to catch the fraud pattern that existed five years ago: individual bad actors using obvious techniques. They work reasonably well for that use case. They're inadequate for organized fraud because the attack surface has shifted.
Geo-IP spoofing has become routine
Residential proxy services—as opposed to data center VPNs, which are relatively easy to block—route traffic through real residential IP addresses in legitimate locations. When your geo-IP check resolves a respondent to an address in Chicago, you can't tell whether they're actually in Chicago or routing through a residential proxy that happens to exit in Chicago. Data center IP blocklists don't catch residential proxies, and residential proxy networks have grown substantially more accessible and affordable in the last two years.
Our current approach uses a multi-signal location verification that combines IP, device GPS (where permitted), mobile carrier country code, and Wi-Fi SSID country hash. No single signal is decisive; the combination makes location spoofing substantially harder to execute without triggering flags. Even this approach isn't foolproof—we estimate we miss approximately 3–5% of sophisticated location fraud.
Cookie clearing defeats basic deduplication
Cookie-based device identification can be defeated in about 30 seconds by clearing browser cookies. This is widely known within fraud communities, and any provider relying primarily on cookie-based deduplication is providing essentially no protection against repeat participation by motivated actors.
Hardware-level device fingerprinting—based on device hardware characteristics that persist through cookie clearing, browser changes, and VPN switches—raises the bar substantially. A fraud actor needs to use physically different hardware to circumvent hardware fingerprinting. This is possible (cheap Android devices are abundant) but requires more infrastructure investment, which effectively filters out lower-sophistication fraud while pushing higher-sophistication actors toward approaches we describe below.
Timing calibration defeats simple speeder detection
Fraud ring communities share information about typical completion time distributions for different survey types. Members are coached to complete surveys in the 60th–80th percentile of typical completion time—fast enough to maintain throughput, slow enough to avoid the speeder flag threshold. Against a simple minimum-time-per-question check, this approach is undetectable.
Per-question timing analysis—scored against a distribution of legitimate completion patterns for that specific question type—is more robust. An experienced fraudulent respondent who knows to slow down overall might still show unusual timing patterns at the question level: pausing at easy questions, completing complex questions unusually quickly. These patterns are detectable when you have a reference distribution to compare against.
The open-end problem
The fastest-growing fraud vector in 2024 is AI-assisted open-end completion. Fraud ring participants increasingly use LLMs to generate contextually appropriate responses to open-ended questions, then paste them in. These responses:
- Pass minimum length filters
- Include topically relevant keywords
- Are grammatically correct
- Don't exhibit obvious machine-generation artifacts
What they share—and what makes detection tractable—is semantic similarity. When multiple participants in a fraud ring use the same LLM prompt or template structure to answer the same question, the resulting responses cluster in embedding space even when they differ in surface wording. Cosine similarity scoring across all open-end responses within a study catches this clustering pattern.
We also run an AI-generation likelihood score on each response independently. Neither signal is decisive alone—some legitimate responses are similar to each other (because respondents share genuine opinions), and some AI-generated text scores low on AI-detection models. But the combination of intra-study similarity and AI-generation likelihood provides reliable detection across most attack patterns we've seen.
What this means for B2B panels specifically
B2B studies are disproportionately targeted by organized fraud for a straightforward reason: they pay more. A study targeting IT decision-makers at enterprise companies typically offers 3–5× the per-complete incentive of a mass-consumer study. That economic incentive makes the investment in fraud sophistication—better VPN services, AI-assisted responses, device rotation—worthwhile in a way it wouldn't be for lower-value studies.
The implication is that quality protection that's adequate for consumer research is often insufficient for B2B research. The fraud actors targeting your C-suite or IT director study are better resourced and more sophisticated than those targeting a general population study.
"The incentive structure of B2B research creates a selection effect: the fraud targeting high-value B2B studies is, on average, more sophisticated than the fraud targeting lower-value consumer studies. Your quality controls need to match the sophistication of the fraud you're actually facing."
What we still can't reliably detect
Transparency requires acknowledging what our detection framework doesn't catch well.
The most sophisticated attack pattern—which we see in perhaps 1–2% of attempted fraud—involves genuine individuals who legitimately qualify for a study target profile but respond inattentively or randomly for economic reasons. A real IT director at a real 500-employee company who clicks through a survey in 8 minutes without reading the questions is using credentials our verification systems will accept, completing in time our timing systems won't flag, and producing responses our pattern-detection systems may not identify.
Embedded attention items are our primary defense against this pattern, and they work—but imperfectly. A respondent who fails two of three attention checks is rejected. A respondent who gets lucky on two and fails one is flagged but not automatically excluded. This is a design choice: aggressive exclusion on attention failures produces false positives among legitimate but fast respondents. We've calibrated to a position that we believe is right, but it's not zero-miss.
We publish our false positive rate (4.2% of manual review sample) because it's a meaningful quality metric. A provider who claims no false positives is claiming their detection is too weak to ever misclassify—which means it's too weak to catch much.
Learn how we apply this in practice
Our quality control documentation describes each detection checkpoint, the thresholds we use, and how we calibrate between sensitivity and false positive risk.
Read Quality Framework