Calculating the MD5 digest (or any other checksum) of a file can take quite a bit of time, especially if the file is large.
And there is no point of computing the MD5 of two files to see whether they're identical if their size isn't the same. And, of course, finding the size of a file is much faster.
So I would suggest that a possible performance enhancement is to compute the MD5 of files only for files that have the same size.
In reply to Re: List Duplicate Files in a given directory
by Laurent_R
in thread List Duplicate Files in a given directory
by pr33
| For: | Use: | ||
| & | & | ||
| < | < | ||
| > | > | ||
| [ | [ | ||
| ] | ] |