Skip to content

Generates a random robots.txt deny list to throw script kiddies off the scent.

License

Notifications You must be signed in to change notification settings

james-see/random-robots-txt

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

random-robots-txt

Generates a random robots.txt deny list to throw script kiddies off the scent.

Requirements

OS

OSX or Debian-based Linux

SOFTWARE

Python 3.6+
Check that /usr/share/dict/words exists for your os and shuf or gshuf commands work.
If not, in ubuntu install sudo apt install wamerican coreutils or in OSX brew install coreutils

CONFIG

Change the global variable numberofwords in each .py file to the number of deny paths to generate in the robots.txt file.

RUN

To run on OSX, simply python3 gen-robots-osx.py or python3.6 gen-robots-osx.py if you have multiple pythons installed.

To run on Debian/Ubuntu, simply python3 gen-robots-ubuntu.py or python3.6 gen-robots-ubuntu.py if you have multiple pythons installed.

Regardless of version, a robots.txt file will be generated in the same folder where you execute the script from.

About

Generates a random robots.txt deny list to throw script kiddies off the scent.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages