Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Server-side analytics will tell you in most cases, and if they don’t, they’re easy to add in a way that makes them unlikely to be blocked?


Server side analytics are overwhelmed by bots and crawlers and other non humans in my experience


Maybe we should we whitelisting for analytics. Instead of perpetually trying to identify bots, Pick the 20% of your users who you are pretty sure are actually humans. Nielsen-style, if you will.

Maybe two cohorts. The people we're really sure about, and the ones we're pretty sure about. If they diverge, ask why.


> Instead of perpetually trying to identify bots, Pick the 20% of your users who you are pretty sure are actually humans.

Note that reCAPTCHA thinks you’re a bot if you block analytics…


Not that that matters to me since I have (re)captcha blocked, too...


Blocking reCAPTCHA has lost me hours of work, repeatedly. And I still do it.

The Textarea Cache Firefox extension https://addons.mozilla.org/en-US/firefox/addon/textarea-cach... is incredibly useful, but I keep forgetting to install it when I get to a new machine. I forget only once.


In general it's quite easy to filter out bots and crawlers from your basic access logs, as most bots and crawlers will identify themselves as such.

If you're running anything with an API, then unless somethings horribly wrong it's even easier: look at the number of requests being made to an API endpoint and spot check a few of the user identifiers (tokens, keys, whatever you're using) to see the variety of users.

All of this is assuming you're trying to merely investigate the volume of use of a feature, not trying to diagnose demographics. If you're trying to extract more fine-grained detail, I don't have as many answers; I hope others will chime in with constructive ways to get things like geographic demographics via server logs.


A very sizable portion of bot traffic does not identify itself as such. I don’t know if it’s a majority now, but it could be.


Many bots and crawlers are designed to be indistinguishable from humans.


That sounds like a server-side problem, not a client-side one. Don't expect me to solve it for you at my end.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: