Offline web browsing

Here is a new version of a simple script I have written to create local copies of websites suitable for browsing offline. We have been using the program successfully at the university to schedule downloads of websites during off peak hours of internet usage for reading the following day.

The program uses the *nix wget utility to do its magic. My code is simply a wrapper around wget which sets the proper command line arguments for wget to create a mirror website. The script uses conservative settings by default for fetching sites in order to be respectful to website owners and other users of the network. Once a site is downloaded the program automatically zips the file in a tar gz for you. You will need python and wget installed in order to run.