Let’s say you have a folder with 5000 MP3 files you want to check for duplicates. Or a directory containing thousands of EPUB files, all with different names but you have a hunch some of them might be duplicates. You can cd your way in the console up to that particular folder and then do a
find -not -empty -type f -printf “%s\n” | sort -rn | uniq -d | xargs -I{} -n1 find -type f -size {}c -print0 | xargs -0 md5sum | sort | uniq -w32 --all-repeated=separate
This will output a list of files that are duplicates, according tot their HASH signature.
Another way is to install fdupes and do a
fdupes -r ./folder > duplicates_list.txt
The -r is for recursivity. Check the duplicates_list.txt afterwards in a text editor for a list of duplicate files.
![Digg](http://1.bp.blogspot.com/_lxBSX0YJV58/SwQiO9krTcI/AAAAAAAAAhc/80Z-NLrZ7wU/s1600/digg.png)
![Google Bookmarks](http://1.bp.blogspot.com/_lxBSX0YJV58/SwQitItySOI/AAAAAAAAAhs/RQQh6Du9HrA/s1600/google.png)
![reddit](http://1.bp.blogspot.com/_lxBSX0YJV58/SwQitlmYSuI/AAAAAAAAAh8/mTtVHQxL_mE/s1600/reddit.png)
![Mixx](http://2.bp.blogspot.com/_lxBSX0YJV58/SwQitawG7PI/AAAAAAAAAh0/vF4HOlj8JH8/s1600/mixx.png)
![StumbleUpon](http://2.bp.blogspot.com/_lxBSX0YJV58/SwQit6IUS5I/AAAAAAAAAiE/gfbBSYnbjnU/s1600/stumbleupon.png)
![Technorati](http://4.bp.blogspot.com/_lxBSX0YJV58/SwQiuP6h64I/AAAAAAAAAiM/x_aGzBKYCgc/s1600/technorati.png)
![Yahoo! Buzz](http://1.bp.blogspot.com/_lxBSX0YJV58/SwQi2FtKY5I/AAAAAAAAAiU/ByTK_Wci_Uo/s1600/yahoo.png)
![DesignFloat](http://4.bp.blogspot.com/_lxBSX0YJV58/SwQiOmnjWDI/AAAAAAAAAhU/nMIbaOaCDic/s1600/designfloat.png)
![Delicious](http://4.bp.blogspot.com/_lxBSX0YJV58/SwQiOHLlI8I/AAAAAAAAAhM/6JP39Dh6sF8/s1600/delicious.png)
![BlinkList](http://3.bp.blogspot.com/_lxBSX0YJV58/SwQiN859bNI/AAAAAAAAAhE/E_Uh6XTQmpo/s1600/blinklist.png)
![Furl](http://3.bp.blogspot.com/_lxBSX0YJV58/SwQiPG9Y6xI/AAAAAAAAAhk/YzmjQUPdnrg/s1600/furl.png)