this post was submitted on 21 Mar 2025
208 points (99.1% liked)
Linux
6661 readers
118 users here now
A community for everything relating to the GNU/Linux operating system
Also check out:
Original icon base courtesy of lewing@isc.tamu.edu and The GIMP
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Most bots and scrapers from what I've seen already are using (headless) full browsers, and hence are executing javascript, so I think anything that slows them down or increases their cost can reduce the traffic they bring.
Source? I strongly disagree, and it's not hard to change your browser characteristics to get a new canvas fingerprint every time, some browsers like firefox even have built-in options for it.
@refalo @sudo If Proof of Work gets widely adopted I foresee a future where bot running data-centers can out-compute humans to visit sites, while old devices of users in poorer countries struggle to compute the required task for hours … Or is that fear misguided?
Admins will always turn down the bot management when it starts blocking end users. At that point you cough up the money for the extra bandwidth and investigate different solutions.
That's not going to be the majority of your bot traffic by a long shot because it doesn't scale like using basic HTTP requests.
This is from personal experience. With PoW you just need any puppetted browser, maybe less. With Canvas finerprinting you need a heavily customized scraping browser, either one you made yourself or one you're paying for. If that's the case the cost of PoW is neglible. If you still want actual stats, I'd have to ask where you're getting any stats on PoW working.