在one answer to: Is shared readonly data copied to different processes for multiprocessing?中,给出了一个用于numpy数组的共享内存的可行解决方案。
如果使用pandas DataFrame会是什么样子?
背景:我希望能够在多重处理期间写入DataFrame,并且希望能够在多重处理完成之后进一步处理它。
答案 0 :(得分:1)
如果您不想使用dask,则可以使用共享内存共享熊猫数据帧,方法是先将其转换为numpy数组,然后在子进程中对其进行重构。
from multiprocessing import shared_memory
def create_shared_block(to_share, dtypes):
# float64 can't be pickled
for col, dtype in to_share.dtypes.items():
if dtype == 'float64':
to_share[col] = pd.to_numeric(to_share[col], downcast='float')
# make the dataframe a numpy array
to_share.reset_index(inplace=True)
# drop the index if named index
to_share = to_share.drop('index', axis=1)
# get the dtypes in the same order as the dataframe columns and make sure the types are correct for numpy
dtypes_sorted = sort_dtypes(to_share, dtypes)
# get the dataframe values in the format expected by numpy
values = [tuple(x) for x in to_share.values.tolist()]
# create a numpy array
to_share = np.array(values, dtype=(dtypes_sorted))
# create a shared memory of the size of the array
shm = shared_memory.SharedMemory(create=True, size=to_share.nbytes)
# now create a NumPy array backed by shared memory
np_array = np.ndarray(to_share.shape, dtype=dtypes_sorted, buffer=shm.buf)
# Copy the original data into shared memory
np_array[:] = to_share[:]
return shm, np_array, dtypes_sorted
def sort_dtypes(df, dtypes):
# category is a pandas dtype, not numpy
string_types = ('category', 'object', '|S')
dtypes = [(x, '|S{}'.format(df[x].str.len().max())) if y in string_types else (x, y) for x, y in dtypes if
x in df.columns]
# build a lookup
dtypes_dict = {x: y for x, y in dtypes}
# fix the order
dtypes_sorted = [(x, dtypes_dict[x]) for x in df.columns]
return dtypes_sorted
# ------PARENT PROCESS-------#
# create your shared memory
to_share = pd.DataFrame([['obstacle','obstacle',2,3],['obstacles','obstacle',2,np.nan]],columns=['w1','w2','d1','d2'])
dtypes = [('w1','str'),('w2','|S'),('d1','f'),('d2','f')]
shm, arr, dtypes_sorted = create_shared_block(to_share, dtypes)
# then pass these values to your child processes
shared = (shm.name, arr.shape, dtypes_sorted)
# ------CHILD PROCESS-------#
# assuming you have passed to the child process in a variable called shared, you can reconstruct the dataframe as follows
shared_memory = shared_memory.SharedMemory(name=shared[0])
np_array = np.ndarray(shared[1], dtype=shared[2], buffer=shared_memory.buf)
columns = [x for x, y in shared[2]]
df = pd.DataFrame(np_array, columns=columns)
这在共享10万行数据帧时在我的应用程序中节省了一些内存,但可能不及使用dask等已建立的库保存的内存。而且我不太确定重新创建pandas数据框所涉及的开销-我想认为它只是引用共享的numpy数组,并在顶部添加了一些额外的内容以使其成为df。