ClioSport.net

Register a free account today to become a member!
Once signed in, you'll be able to participate on this site by adding your own topics and posts, as well as connect with other members through your own private inbox!

  • When you purchase through links on our site, we may earn an affiliate commission. Read more here.

archiving a website



there's a website I would like to acrhive, the only problem is that most of it's pages aren't linked from anywhere.

the format for the pages is: p-001.htm to p-123435435 (or some very high number)

does anyone know of any software that's out there which will download and store the pages using an incremental system?

or does anyone know of anything that will generate a large wordlist into a .txt file with certain parameters (p-NUMBER.htm) ?

any help appreciated!

(p.s this is all legal and innocent)
 

KDF

  Audi TT Stronic
Code:
<?php

$max = 123435435;
$saveto = "/savefilehere";

while ($i = 0; $i < $max; $i++)
{
    if($i < 100)
        {
            $number = str_pad((int) $i, 3, "0", STR_PAD_LEFT);
        }
        else
        {
           $number = $i;
        }
    
    $file = file_get_contents('http://website.com/p-'.$number.'.htm');
    file_put_contents($file, $saveto)
}

?>

Knocked up in 2 min. I could also write it in C, C++, C#, Asp, Python, Bash, vbasic etc.. if Php doesn't do it for you.. ;)
 


Top