Bots detection #23
Replies: 7 comments
-
While filtering based on the User-Agent in request headers is possible, it can be easily bypassed. |
Beta Was this translation helpful? Give feedback.
-
I am using hits to avoid Google Analytics :P reasons
|
Beta Was this translation helpful? Give feedback.
-
what if we assume that crawlers are well-behaving and do not want to bypass protections? e.g. anything like this is possible? any well-known protocols/standards to talk-to/detect crawlers? (one way I can imagine is to have reverse-proxy at hits.sh end with robots.txt that disallows going further. and then somehow making another HTTP request to your backend, no without robots who dropped at |
Beta Was this translation helpful? Give feedback.
-
That sounds interesting. But if I do that, the |
Beta Was this translation helpful? Give feedback.
-
Hmm.. |
Beta Was this translation helpful? Give feedback.
-
something like this probably will work instead of 304 can also use meta html tag: https://www.w3.org/TR/WCAG20-TECHS/H76.html |
Beta Was this translation helpful? Give feedback.
-
some other method people say |
Beta Was this translation helpful? Give feedback.
-
Is there way (or ideas?) on how to detect real humans vs bots that crawl webpages?
Beta Was this translation helpful? Give feedback.
All reactions