It happened again....
This time a vital machine at work went down - this happened a couple of years ago with another machine, so I built a backup system using rsync that created daily backups of important data. Both ends of the backup system would email if there was a problem i.e client was unable to backup to the server or that the server had not seen machine x backup for y number of days.
Brilliant, it worked fantastic.
That is until you put a human into the equation and they happen to just ignore the emails saying that "machine x has not backed up for 230 days".
Woke up in the middle of the night in a panic because there was critical data on that machine.
Managed to get somebody to put the drive into another machine this morning and *luckily* I was able to copy out the important data from the drive. The drive is buggered, there's a physical fault on it, but fortunately not on the important data. Bizarrely enough, the rsync backup scripts do a sync locally before doing it across the network and the backup script had been happily backing up data locally! There were a couple of years worths of daily backups, I grabbed those as well just incase the data was buggered, far easier to recover from data thats a couple of days/weeks old than 12 months.
Moral of the story, always assume that humans will cover up the big red flashing light with something rather than act on it!
Backups people, double check!
This time a vital machine at work went down - this happened a couple of years ago with another machine, so I built a backup system using rsync that created daily backups of important data. Both ends of the backup system would email if there was a problem i.e client was unable to backup to the server or that the server had not seen machine x backup for y number of days.
Brilliant, it worked fantastic.
That is until you put a human into the equation and they happen to just ignore the emails saying that "machine x has not backed up for 230 days".
Woke up in the middle of the night in a panic because there was critical data on that machine.
Managed to get somebody to put the drive into another machine this morning and *luckily* I was able to copy out the important data from the drive. The drive is buggered, there's a physical fault on it, but fortunately not on the important data. Bizarrely enough, the rsync backup scripts do a sync locally before doing it across the network and the backup script had been happily backing up data locally! There were a couple of years worths of daily backups, I grabbed those as well just incase the data was buggered, far easier to recover from data thats a couple of days/weeks old than 12 months.
Moral of the story, always assume that humans will cover up the big red flashing light with something rather than act on it!
Backups people, double check!