我正在处理一个内存问题,当我在一个文件夹中搜索一个百万的文件。有人知道如何克服这种情况吗?有什么方法可以限制glob搜索的文件吗?这样它就可以成批执行了?
Traceback (most recent call last):
File "./lb2_lmanager", line 533, in <module>
main(sys.argv[1:])
File "./lb2_lmanager", line 318, in main
matched = match_files(policy.directory, policy.file_patterns)
File "./lb2_lmanager", line 32, in wrapper
res = func(*args, **kwargs)
File "./lb2_lmanager", line 380, in match_files
listing = glob.glob(directory)
File "/usr/lib/python2.6/glob.py", line 16, in glob
return list(iglob(pathname))
File "/usr/lib/python2.6/glob.py", line 43, in iglob
yield os.path.join(dirname, name)
File "/usr/lib/python2.6/posixpath.py", line 70, in join
path += '/' + b
MemoryError发布于 2015-10-19 14:43:42
尝试使用generators而不是lists。
要理解生成器是什么,请阅读this
import glob
dir_list = glob.iglob(YOUR_DIRECTORY)
for file in dir_list:
print file将YOUR_DIRECTORY更改为要列出的目录。
https://stackoverflow.com/questions/33217062
复制相似问题