如何在Tensorflow平台上实现早期停止和降低学习率?

2024-04-27 15:33:52 发布

您现在位置:Python中文网/ 问答频道 /正文

我想为使用tensorflow构造的神经网络模型实现两个回调EarlyStopping和{}。(我没有使用Keras

下面的示例代码是我在编写的脚本中如何实现早期停止的,我不知道它是否正确。在

# A list to record loss on validation set
val_buff = []
# If early_stop == True, then terminate training process
early_stop = False

while icount < maxEpoches:

    '''Shuffle the training set'''
    '''Update the model by using Adam optimizer over the entire training set'''

    # Evaluate loss on validation set
    val_loss = self.sess.run(self.loss, feed_dict = feeddict_val)
    val_buff.append(val_loss)

    if icount % ep == 0:

        diff = np.array([val_buff[ind] - val_buff[ind - 1] for ind in range(1, len(val_buff))])
        bad = len(diff[diff > 0])
        if bad > 0.5 * len(diff):
            early_stop = True

        if early_stop:
            self.saver.save(self.sess, 'model.ckpt')
            raise OverFlow()
        val_buff = []

    icount += 1

当我训练模型并跟踪验证集上的损失时,我发现损失会上下波动,所以很难判断模型何时开始过度拟合。在

既然Earlystopping和{}非常相似,我如何修改上面的代码来实现ReduceLearningRateOnPlateau?在


Tags: the模型selfleniftrainingdiffval
1条回答
网友
1楼 · 发布于 2024-04-27 15:33:52

振荡误差/损耗很常见。实施早期停止或学习率降低规则的主要问题是验证损失的计算相对滞后。为了解决这个问题,我建议下一条规则:当最好的验证错误至少超过N个时代时停止训练。在

max_stagnation = 5 # number of epochs without improvement to tolerate
best_val_loss, best_val_epoch = None, None

for epoch in range(max_epochs):
    # train an epoch ...
    val_loss = evaluate()
    if best_val_loss is None or best_val_loss < val_loss:
        best_val_loss, best_val_epoch = val_loss, epoch
    if best_val_epoch < epoch - max_stagnation:
        # nothing is improving for a while
        early_stop = True
        break  

相关问题 更多 >