Websites are tracking you via browser fingerprinting
Scope and goals of the research
- Commenters note fingerprinting has been known and deployed for over a decade, but prior work mostly showed scripts could fingerprint, not that it was actually used for ad tracking at scale.
- This paper’s claimed contribution (via FPTrace) is tying fingerprint changes to ad auction behavior, showing that ad systems really use fingerprints for targeting and to bypass consent/opt-outs (e.g. GDPR/CCPA), not just for fraud/bot detection.
How fingerprinting works and what’s collected
- Fingerprints combine many attributes: UA string, headers, fonts, screen size, GPU/CPU details, media capabilities, timezone/language, storage and permission state, sensors, WebGL/canvas behavior, and sometimes lower-level network or TLS signatures.
- Timing side channels (render speed, interrupts, TCP timestamps, human typing/mouse dynamics) are cited as additional long-lived signals.
- Modern privacy tests (EFF, amiunique, CreepJS, fingerprint.com) demonstrate how easily browsers become statistically unique, though some commenters question their methodology and traffic representativeness.
Persistence, uniqueness, and effectiveness
- Strong disagreement over “half-life of a few days”:
- One side argues many attributes (versions, window size) change quickly, making long-term tracking fragile.
- Others say many properties (hardware, fonts, GPU, sensors, stack behavior) are stable, and trackers can link evolving fingerprints via overlap and cookies.
- Important distinction: uniqueness vs persistence. Being “unique” in a niche test set doesn’t mean globally unique; randomized or spoofed fingerprints may look unique each visit, which actually reduces linkability.
- Several people think adtech’s real-world effectiveness is overstated and often resembles snake oil, though others point out 90%+ long-term match claims from commercial vendors.
IP/geo and cross-device behavior
- Multiple comments say large ad networks lean heavily on IP-based geo and “flood” an area, which explains household and cross-device ad effects.
- VPNs, CGNAT, iCloud Private Relay, mobile IPs, and geolocation drift add noise but often still allow neighborhood-level targeting; some ads obviously change when switching VPN countries.
Defenses, tradeoffs, and practical limits
- Common mitigations: disabling JavaScript, using Tor/Mullvad/Brave, Firefox’s resistFingerprinting and letterboxing, anti-detect browsers (mainly used for fraud/ban evasion), VPNs, adblockers, strict JS and storage controls.
- Tradeoffs are severe: many sites break without JS; aggressive privacy settings increase “weirdness” and can both aid fingerprinting and trigger bot defenses.
- Randomization and dummy data can defeat persistence but often cause privacy-test sites to label you “unique,” confusing users.
- Some argue the only robust strategy is drastically reducing exposed APIs and surface area; others think browsers are constrained by web compatibility and user expectations.
Browsers, standards, and regulation
- Criticism that mainstream browsers, especially those touting privacy, still leak excessive information (detailed UA, referer, fonts, battery, etc.) and move slowly to restrict APIs.
- Debate over whether open-source options (particularly Firefox and derivatives) remain meaningfully privacy-respecting given funding sources and recent ad-related features.
- Several call for stronger regulation and enforcement, since technical defenses alone create an endless cat-and-mouse game while tracking steadily improves.