pep8 tunning

This commit is contained in:
Darko Poljak 2014-01-28 08:53:51 +01:00
parent 8cb421c6e4
commit cf0ff95b36
1 changed files with 9 additions and 7 deletions

View File

@ -15,16 +15,17 @@ Options:
file's content [default: 4096] file's content [default: 4096]
-d <hashalg>, --digest-alg=<hashalg> secure hash algorithm [default: md5] -d <hashalg>, --digest-alg=<hashalg> secure hash algorithm [default: md5]
-a <action>, --action=<action> action on duplicate files (pprint, -a <action>, --action=<action> action on duplicate files (pprint,
print, remove, move) [default: pprint] print, remove, move)
-remove removes duplicate files except [default: pprint]
first found -remove removes duplicate files
-move moves duplicate files to except first found
-move moves duplicate files to
duplicates driectory, except first duplicates driectory, except first
found found
-print prints result directory where -print prints result directory where
keys are hash values and values are keys are hash values and values are
list of duplicate file paths list of duplicate file paths
-pprint prints sets of duplicate file -pprint prints sets of duplicate file
paths each in it's line where sets paths each in it's line where sets
are separated by blank newline are separated by blank newline
-m <directory>, --move=<directory> move duplicate files to directory -m <directory>, --move=<directory> move duplicate files to directory
@ -114,9 +115,10 @@ def mv_file_dups(topdirs=['./'], hashalg='md5', block_size=4096,
shutil.move(f, dest_dir) shutil.move(f, dest_dir)
def iter_file_dups(topdirs=['./'], rethash=False, hashalg='md5', block_size=4096): def iter_file_dups(topdirs=['./'], rethash=False, hashalg='md5',
block_size=4096):
"""Yield duplicate files when found in specified directory list. """Yield duplicate files when found in specified directory list.
If rethash is True then tuple hash value and duplicate paths list is If rethash is True then tuple hash value and duplicate paths list is
returned, otherwise duplicate paths list is returned. returned, otherwise duplicate paths list is returned.
""" """
dups = file_dups(topdirs, hashalg, block_size) dups = file_dups(topdirs, hashalg, block_size)