Page 1 of 1
Performance improvement
Posted: Wed Jun 08, 2011 1:01 pm
by St�phane BARIZIEN
Process Monitor reveals you are reading files 4KB at a time.
Reading with larger e.g. 64KB or even 1MB buffers would probably increase performance significantly for large files.
Just my �0.02
Posted: Sun Jun 12, 2011 9:51 am
by DV
Thanks, will experiment with this.
Re: Performance improvement
Posted: Wed Jul 06, 2011 4:32 am
by Myth
It's likely that Duplicate Cleaner already employs the following optimisation, but here goes anyway:-
The idea is to reduce the number of files checked for time consuming criteria by eliminating obvious mismatches through less intensive criteria first - typically using multiple passes, with the most intensive searches occuring in the final pass.
Pass 1: Compare for search criteria that can be obtained from the master file table (name, size, attributes etc). Size is particularly handy as two files with the same content are going to be the same size, so it follows that a mismatch in file sizes means there cannot be a match in content. Pass 1 should result in a preliminary list of matching file groups.
Pass 2: Step through pass 1's list of file groups, checking each group for more intensive criteria (such as content matching).