我有一个由7列和900 K行组成的数据集。所有列都是非唯一的,所有的值都是整数.
过滤的两个重要条件:
。
举个例子,这里是用于基准测试性能的SQL查询:
SELECT DISTINCT
col_2
FROM dataset
WHERE
c_1 in (1,9,5,6,8,18,14,7,15) AND
c_3 in (1) AND
c_4 in (61) AND
c_5 in (3) AND
c_6 in (0) AND
c_7 in (0)我尝试的第一种方法是使用SQLite索引的SQL,这并不是很糟糕,但是由于过滤器返回了很多行,性能下降了。
然后,我尝试了Python中的普通列表理解。性能比SQL中的性能要差一些。
有什么更好的方法吗?我考虑的方向是numpy,也许使用比lists和SQL表更有效的数据结构?
我对这里的速度和性能非常感兴趣,但效率却不那么高。
欢迎任何建议!
发布于 2011-01-07 12:20:05
按照您所说的,每列大约有20个不同的值,除了一个有400的值。如果内存和加载时间不是问题,那么我建议为每列创建一个值集。
这里有一些东西可以生成一个样本数据集。
#!/usr/bin/python
from random import sample, choice
from cPickle import dump
# Generate sample dataset
value_ceiling = 1000
dataset_size = 900000
dataset_filename = 'dataset.pkl'
# number of distinct values per column
col_distrib = [400,20,20,20,20,20,20]
col_values = [ sample(xrange(value_ceiling),x) for x in col_distrib ]
dataset = []
for _ in xrange(dataset_size):
dataset.append(tuple([ choice(x) for x in col_values ]))
dump(dataset,open(dataset_filename,'wb'))这里有一些内容可以加载数据集并创建每个列的每个值的查找集、一个搜索方法和创建示例搜索。
#/usr/bin/python
from random import sample, choice
from cPickle import load
dataset_filename = 'dataset.pkl'
class DataSearch(object):
def __init__(self,filename):
self.data = load(open(filename,'rb'))
self.col_sets = [ dict() for x in self.data[0] ]
self.process_data()
def process_data(self):
for row in self.data:
for i,v in enumerate(row):
self.col_sets[i].setdefault(v,set()).add(row)
def search(self,*args):
# args are integers, sequences of integers, or None in related column positions.
results = []
for i,v in enumerate(args):
if v is None:
continue
elif isinstance(v,int):
results.append(self.col_sets[i].get(v,set()))
else: # sequence
r = [ self.col_sets[i].get(x,set()) for x in v ]
r = reduce(set.union,r[1:],r[0])
results.append(r)
#
results.sort(key=len)
results = reduce(set.intersection,results[1:],results[0])
return results
def sample_search(self,*args):
search = []
for i,v in enumerate(args):
if v is None:
search.append(None)
else:
search.append(sample(self.col_sets[i].keys(),v))
return search
d = DataSearch(dataset_filename)并使用它:
>>> d.search(*d.sample_search(1,1,1,5))
set([(117, 557, 273, 437, 639, 981, 587), (117, 557, 273, 170, 53, 640, 467), (117, 557, 273, 584, 459, 127, 649)])
>>> d.search(*d.sample_search(1,1,1,1))
set([])
>>> d.search(*d.sample_search(10,None,1,1,1,1))
set([(801, 334, 414, 283, 107, 990, 221)])
>>> d.search(*d.sample_search(10,None,1,1,1,1))
set([])
>>> d.search(*d.sample_search(10,None,1,1,1,1))
set([(193, 307, 547, 549, 901, 940, 343)])
>>> import timeit
>>> timeit.Timer('d.search(*d.sample_search(10,None,1,1,1,1))','from __main__ import d').timeit(100)
1.7874310016632081.8秒完成100次搜索足够快吗?
发布于 2011-01-07 12:08:41
我想出的是:
513 $ cat filtarray.py
#!/usr/bin/python2
#
import numpy
import itertools
a = numpy.fromiter(xrange(7*900000), int)
a.shape = (900000,7)
# stuff a known match
a[33][0] = 18
a[33][2] = 1
a[33][3] = 61
# filter it, and make list, but that is not strictly necessary.
res = list(itertools.ifilter(lambda r: r[0] in (1,9,5,6,8,18,14,7,15) and r[2] == 1 and r[3] == 61, a))
print res在Intel E8400上运行它:
512 $ time python filtarray.py
[array([ 18, 232, 1, 61, 235, 236, 237])]
python filtarray.py 5.36s user 0.05s system 99% cpu 5.418 total这样更快吗?
发布于 2011-01-07 12:20:42
这是一个矮胖的版本,大约需要1秒。
x = numpy.random.randint(0, 100, (7, 900000))
def filter(data, filters):
indices = []
for i, filter in enumerate(filters):
indices.append(numpy.any([data[i] == x for x in filter], 0))
indices = numpy.all(indices, 0)
return data[indices]
# Usage:
filter(x, [(1,9,5,6,8,18,14,7,15), (1,), (61,), (3,), (0,), (0,)])
%timeit filter(x, [(1,9,5,6,8,18,14,7,15), (1,), (61,), (3,), (0,), (0,)])
1 loops, best of 3: 903 ms per loophttps://stackoverflow.com/questions/4624129
复制相似问题