用于预测数字序列的 Keras 模型

Mos*_*osu 5 machine-learning neural-network lstm keras recurrent-neural-network

我正在尝试训练 Keras LSTM 模型来预测序列中的下一个数字。

  1. 下面我的模型有什么问题,当模型没有学习时我如何调试
  2. 我如何决定使用哪些图层类型
  3. 编译时我应该根据什么选择损失和优化器参数

我的输入训练数据的形状 (16000, 10) 如下所示

[
    [14955 14956 14957 14958 14959 14960 14961 14962 14963 14964]
    [14731 14732 14733 14734 14735 14736 14737 14738 14739 14740]
    [35821 35822 35823 35824 35825 35826 35827 35828 35829 35830]
    [12379 12380 12381 12382 12383 12384 12385 12386 12387 12388]
    ...
]
Run Code Online (Sandbox Code Playgroud)

相应的输出训练数据的形状 (16000, 1) 如下所示

[[14965] [14741] [35831] [12389] ...]
Run Code Online (Sandbox Code Playgroud)

正如 LSTM 所抱怨的,我重塑了训练/测试数据

X_train = X_train.reshape(X_train.shape[0], X_train.shape[1], 1)
X_test = X_test.reshape(X_test.shape[0], X_test.shape[1], 1)
Run Code Online (Sandbox Code Playgroud)

这是最终的训练/测试数据形状

Total Samples: 20000
X_train: (16000, 10, 1)
y_train: (16000, 1)
X_test: (4000, 10, 1)
y_test: (4000, 1)
Run Code Online (Sandbox Code Playgroud)

这是我的模型

# Model configuration
epochs = 2
batch_size = 32
hidden_neurons = 100
output_size = 1

# Create the model
model = Sequential()
model.add(LSTM(hidden_neurons, input_shape=(X_train.shape[1], X_train.shape[2])))
model.add(Dense(output_size))
model.compile(loss='mean_squared_error', optimizer='rmsprop', metrics=['accuracy'])
print(model.summary())
model.fit(X_train, y_train, epochs=epochs, batch_size=batch_size)

scores = model.evaluate(X_test, y_test, batch_size=batch_size, verbose=0)
print("Model Accuracy: %.2f%%" % (scores[1]*100))
Run Code Online (Sandbox Code Playgroud)

这是我的输出

_________________________________________________________________
Layer (type)                 Output Shape              Param #   
=================================================================
lstm_3 (LSTM)                (None, 100)               40800     
_________________________________________________________________
dense_3 (Dense)              (None, 1)                 101       
=================================================================
Total params: 40,901
Trainable params: 40,901
Non-trainable params: 0
_________________________________________________________________
None
Epoch 1/2
16000/16000 [==============================] - 11s - loss: 533418575.3600 - acc: 0.0000e+00    
Epoch 2/2
16000/16000 [==============================] - 10s - loss: 532474289.7280 - acc: 6.2500e-05    
Model Accuracy: 0.00%
Run Code Online (Sandbox Code Playgroud)

Pad*_*ddy 2

试试这个代码:

epochs = 30
batch_size = 64
hidden_neurons = 32
output_size = 1

# Create the model
model = Sequential()
model.add(LSTM(hidden_neurons, input_shape=(X_train.shape[1], X_train.shape[2])))
model.add(Dense(output_size, activation = 'elu'))

model.compile(loss='mean_squared_error', optimizer='adam', metrics=['accuracy'])
print(model.summary())
model.fit(X_train, y_train, epochs=epochs, batch_size=batch_size)

scores = model.evaluate(X_test, y_test, batch_size=batch_size, verbose=0)
print("Model Accuracy: %.2f%%" % (scores[1]*100))
Run Code Online (Sandbox Code Playgroud)

一般来说,这真的很难帮助你,因为我们需要一种可以测试的可重现的例子。不过,我的建议如下:

调整神经网络的超参数,例如:激活函数、选择函数、层数、学习率等。

更新:

强烈建议首先标准化您的数据。