I did a search for duplicate finder programs, and one seemed to come out quite well in some reviews - though not necessarily all of them. That was the Gemini program available from the Apple App store (for Mac).
indeed it does seem to work, and I've done a few tests to see if it will detect obvious duplicates, and other less obvious ones, and also not get confused if a small edit is made to an existing file.
I then scanned the whole of my main drive on this machine, and it appears to have found around 100 Gbytes of duplicates - whether that really means 100 GBytes of spare space which can be claimed back I'm not quite sure yet. 50 x 1 Gbytes file duplicated would represent 50 Gbytes which could be cleared off, while 25 x 1 Gbyte files quadruplicated would represent 75 Gbytes which could be cleared.
Having identified some files which appear to be duplicates, I then move large ones to a folder called something like PENDING DELETE, and leave them there for a while. That should ensure that they get backed up to a TIM backup. Some while later I delete the files which reclaims the space.
Obviously this will mean that the TM backup could get duplicates - but at present the TM disc is substantially larger than the main drive.
This strategy should work for a while, at least, and should ensure that files do get spread around different drives, but that duplicates on any one drive (except the TM drive) are avoided.
I think that later on a more structured approach will be needed - actually classifying each file - music, photos, videos etc. and grouping them onto different media/devices.
indeed it does seem to work, and I've done a few tests to see if it will detect obvious duplicates, and other less obvious ones, and also not get confused if a small edit is made to an existing file.
I then scanned the whole of my main drive on this machine, and it appears to have found around 100 Gbytes of duplicates - whether that really means 100 GBytes of spare space which can be claimed back I'm not quite sure yet. 50 x 1 Gbytes file duplicated would represent 50 Gbytes which could be cleared off, while 25 x 1 Gbyte files quadruplicated would represent 75 Gbytes which could be cleared.
Having identified some files which appear to be duplicates, I then move large ones to a folder called something like PENDING DELETE, and leave them there for a while. That should ensure that they get backed up to a TIM backup. Some while later I delete the files which reclaims the space.
Obviously this will mean that the TM backup could get duplicates - but at present the TM disc is substantially larger than the main drive.
This strategy should work for a while, at least, and should ensure that files do get spread around different drives, but that duplicates on any one drive (except the TM drive) are avoided.
I think that later on a more structured approach will be needed - actually classifying each file - music, photos, videos etc. and grouping them onto different media/devices.
Comment