I messed around a little bit with QuadSucker and it isn't too complicated. In summary: It works when you set proper link depth and proper start page. But it has some serious problems and won't work well.
The main problem with is that it will download a lot of junk. For instance, it won't just download the threads, but also download multiple versions of the index page (an index page sorted by user name, an index page sorted by last post, etc.). It will also download multiple versions of the same thread so there will be a LOT of bloating and mess.
Here's how I used the program to save some files.
You want to start with a link that indexes all the oldest threads on a given site. So you could use this one:
http://www.bladeforums.com/forums/f...=1&pp=25&sort=lastpost&order=asc&daysprune=-1
or this one, if you want to index more threads at a time (recommended):
http://www.bladeforums.com/forums/f...1&pp=200&sort=lastpost&order=asc&daysprune=-1
Apparently you can't show more than 200 threads on a single page.
Now you will need to go into the settings (under the menu Settings, click Configuration Options).
Under Directory Structure, use "Mirror the Website" (should be default anyway)
Under Download Directory put whatever you want.
Under Spidering, I unchecked the "off-site images" and "off-site pages" options. This ought to reduce the volume of what you are downloading.
Now set the "link depth" to 1. This is so you only download threads on the HI forum and not from every single forum. In specific, setting the link depth to 1 will force QuadSucker to only download stuff which is linked directly from the index page.
Under "Link Relativiser" I think you need to make sure that "Relativize on-site links" is checked. Otherwise the links on the page will not refer to what is on your hard drive, but rather, what is on the Internet. This appears to be the problem that ferguson had.
On the other hand it was working fine for me without relativized links. I think this might be due to the directory structure that is used.
Now, ferguson, you still have the important data on your hard drive, but it is just a bit disorganized. It should still be possible to find any given thread. If you look at the list of files that QuadSucker downloaded, there are probably a bunch of files with names like: "showthread.php_t=198282&page=1". This is the actual content, but the problem is separating, searching, sorting, and distributing this content. I'll have to think on this a little more.
back to the tutorial.
Close the configuration window. Go to the Settings menu again, and click on Priority Keywords. Add the keyword "showthread.php". This makes Quadsucker realize that the threads are the most important content on the page. It will download threads before anything else.
Now when you click "go" the program should start downloading all the threads on the index page you gave it. When it is done you will have to give it another index page. If you put 200 threads per page there are at most 50-some pages that you will have to download to get the entire contents of the HI forum.
But like I said, there will be a great deal of mess and bloat. After testing it out, this program seems less than ideal for this specific application. I'll see if I can find something else.
Here's a list of various ones which may or may not work with Windows:
http://www.manageability.org/blog/stuff/open-source-web-crawlers-java/view