Hi,
I think the simplest way is to use dd. I made a backup with:
dd if=/dev/sdb bs=2048 count=700000 | gzip -v > backup_cd1.img.gz
dd if=/dev/sdb bs=2048 skip=700000 | gzip -v > backup_cd2.img.gz
To put it back, mount the cd's and:
gunzip -cd backup_cd1.img.gz | dd of=/dev/sdb bs=2048 count=700000
gunzip -cd backup_cd2.img.gz | dd of=/dev/sdb bs=2048 seek=700000
Hope it helps,
Greetz,
Sebastiaan
On 8 Feb 2001, David Bishop wrote:
> I'm trying to make backups to CD, and of course, have 800-900 megs worth of
> data, compressed. What is the best way to split up large tar or cpio files,
> that will allow them to easily be put back together, booting off of a rescue
> floppy or the like? I don't need any scripts or direct-to-the-burner
> solutions, neither of the computers have cdr's in them. Just a simple way
> for my boss, who is only very slightly linux savvy, to be able to run
> restores.
>
> TIA and HAND,
>
> D.A.Bishop
>
>
> --
> To UNSUBSCRIBE, email to [EMAIL PROTECTED]
> with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]
>
>
--
To UNSUBSCRIBE, email to [EMAIL PROTECTED]
with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]