[OSX-Users] Re: reconciling duplicated files/directories/backups
Marcus Cobden
mc08r at ecs.soton.ac.uk
Wed Dec 14 15:25:38 GMT 2011
Ah right.
This'd be too complex to do in a shell, but you could build a matrix 'relevant' of information about each file (basename, timestamps, md5, etc) and then throw it through some kind of clustering algorithm.
That might be overkill :P
Also I can't see an easy way to incorporate directory structures into that, but that might not be needed.
MArcus
On 14 Dec 2011, at 13:12, Les A Carr wrote:
> That's fine if you have two candidate duplicate directories, but I want to identify all the files across various hard disks that may be the same as any others!
> --
> les
>
>
> On 14 Dec 2011, at 13:02, Marcus Cobden wrote:
>
>> There's always the Unix diff command
>>
>> diff -r dir1 dir2
>>
>> I prefer to add in the -u option, and also view a colorised version.
>> Also I'm pretty sure you can get it to only tell you that files differ, and not how.
>>
>> Finally, the timestamps on the files might be useful, so if you move/copy them around, take care to preserve them.
>>
>> Marcus
>>
>> On 14 Dec 2011, at 12:08, Les A Carr wrote:
>>
>>> I am writing some simple tools to help me reconcile multiple copies of files/directories from various attempts to back up or duplicate information across machines for safekeeping.
>>>
>>> Before I go too far down this road (shell scripts and md5 checksums!) does anyone know of any software that already does this?
>>> --
>>> Les
>>
>>
>
>
More information about the Osx-users
mailing list