This Python script can be used to check if there is any differences in responses of an application when the request comes from a search engine's crawler.
- 30 crawler's user agent strings.
- JSON export with
- Auto-detecting responses that stands out.
$ ./crawlersuseragents.py -h [~] Access web pages as web crawlers User-Agents, v1.1 usage: crawlersuseragents.py [-h] [-v] [-t THREADS] [-x PROXY] [-k] [-L] [-j JSONFILE] url This Python script can be used to check if there is any differences in responses of an application when the request comes from a search engine's crawler. positional arguments: url e.g. https://example.com:port/path optional arguments: -h, --help show this help message and exit -v, --verbose arg1 help message -t THREADS, --threads THREADS Number of threads (default: 5) -x PROXY, --proxy PROXY Specify a proxy to use for requests (e.g., http://localhost:8080) -k, --insecure Allow insecure server connections when using SSL (default: False) -L, --location Follow redirects (default: False) -j JSONFILE, --jsonfile JSONFILE Save results to specified JSON file.
Auto-detecting responses that stands out
Results are sorted by uniqueness of their response's length. This means that the results with unique response length will be on top, and results with response's length occurring multiple times at the bottom:
|Two different result lengths||Four different result lengths|
Pull requests are welcome. Feel free to open an issue if you want to add other features.