[ale] Need a big external drive quick. Suggestions?

Greg Freemyer greg.freemyer at gmail.com
Fri Jun 2 14:50:20 EDT 2006


On 6/2/06, Bob Toxen <transam at verysecurelinux.com> wrote:
> You probably should do that dd on a per partition basis, i.e.:
>
>      tcsh
>      foreach i (1 2 3 4 5)
>         echo Doing $i
>         dd bs=10240k if=/dev/hda$i of=/dev/hdb$i
>      end
>
> I find that sometimes just doing:
>
>      dd bs=10240k if=/dev/hda of=/dev/hdb
>
> doesn't work, probably due to architecture.
>
> Another advantage to doing it by partition is that after each partition
> is copied you can mount the new partition on /dev/hdb and test it to
> see if it worked.  You don't want to discover Sunday morning that something
> didn't work.
>
> Of course be REAL careful that you copy in the right direction.  First
> backing up the old drive (if possible) or, at least, its most critical
> data is a real good idea.
>
> Bob Toxen

Thanks Bob, but doing forensics I do this sort of thing all the time.

The only big deal here is the shear size.  Never had to capture a
700GB volume before.  Have always been able to simply buy a big enough
single disk to hold the image.

Sounds like you do dd captures fairly often as well.  On the dd, you
should add conv=noerror,sync.  That allows dd to continue in the
presence of a disk error.

We actually normally use dcfldd (DOD Computer Forensic Lab DD).  It
generates a md5 as it goes.  We have it generate a md5 for each 2 GB
segment and keep the 2 GB segments in seperate files.  We then verify
both the source and the dest have the same md5 for each 2 GB segment
as we recorded in the capture pass.  It surprises me how often we get
a single 2 GB segment md5 disagree.  We then just reperform the
capture for any segments that fail.

Obviously most of the above is scripted.

Greg
-- 
Greg Freemyer
The Norcross Group
Forensics for the 21st Century



More information about the Ale mailing list