我正在尝试在生成器中实现 numpy.memmap 方法,以使用 keras 训练神经网络,以免超过内存 RAM 限制。我将这篇文章用作参考,但没有成功。这是我的尝试:
def My_Generator(path, batch_size, tempo, janela):
samples_per_epoch = sum(1 for line in np.load(path))
number_of_batches = samples_per_epoch/batch_size
#data = np.memmap(path, dtype='float64', mode='r+', shape=(samples_per_epoch, 18), order='F')
data = np.load(path)
# create a memmap array to store the output
X_output = np.memmap('output', dtype='float64', shape=(samples_per_epoch, 96, 100, 17), mode='r+', order='F')
y_output = np.memmap('output', dtype='float64', shape=(samples_per_epoch, 1), mode='r+', order='F')
holder = np.zeros([batch_size, 18], dtype='float64')
counter=0
while 1:
holder[:] = data[counter:batch_size+counter]
X, y = input_3D(holder, tempo, janela)
lenth_X = len(X)
lenth_y = len(y)
print(lenth_X, lenth_y)
y = y.reshape(-1, 1)
X_output[0:lenth_X, :] = X
y_output[0:lenth_y, :] = y
counter += 1
yield X_output[0:lenth_X, :].reshape(-1, 96, 10, 10, 17), y_output[0:lenth_y, :]
#restart counter to yeild data in the next epoch as well
if counter >= number_of_batches:
counter = 0
尽管如此,它仍然将块保存在 RAM 内存中,因此在一些时期后它会超过其限制。
谢谢