import tensorflow as tf
RANDOM_SEED_CONSTANT = 42 # FOR_REPRODUCIBILITY
tf.random.set_seed(RANDOM_SEED_CONSTANT)
# Prevent NHWC errors https://www.nuomiphp.com/eplan/en/50125.html
from tensorflow.keras import backend as K
K.set_image_data_format("channels_last")
from tensorflow import keras
from tensorflow.keras import datasets, layers, models
(train_images, train_labels), (test_images, test_labels) = datasets.cifar10.load_data()
train_images, test_images = train_images / 255.0, test_images / 255.0 # Normalize pixel values to be between 0 and 1
# Create a simple CNN
model = models.Sequential()
model.add(layers.Conv2D(32, (3, 3), activation='relu', input_shape=(32, 32, 3)))
model.add(layers.MaxPooling2D((2, 2)))
model.add(layers.Conv2D(64, (3, 3), activation='relu'))
model.add(layers.MaxPooling2D((2, 2)))
model.add(layers.Conv2D(64, (3, 3), activation='relu'))
model.add(layers.Flatten())
model.add(layers.Dense(64,
activation='relu',
kernel_initializer=tf.keras.initializers.HeNormal(seed=RANDOM_SEED_CONSTANT)))
model.add(layers.Dense(10,
kernel_initializer=tf.keras.initializers.HeNormal(seed=RANDOM_SEED_CONSTANT)))
print(model.summary())
model.compile(optimizer='adam',
loss=tf.keras.losses.SparseCategoricalCrossentropy(from_logits=True),
metrics=['accuracy'])
model.save_weights('myweights.h5')
# Run1
history = model.fit(train_images, train_labels, epochs=1,
shuffle=False,
validation_data=(test_images, test_labels))
# Run2
model.load_weights('myweights.h5')
history = model.fit(train_images, train_labels, epochs=1,
shuffle=False,
validation_data=(test_images, test_labels))
# Run3
model.load_weights('myweights.h5')
history = model.fit(train_images, train_labels, epochs=1,
shuffle=False,
validation_data=(test_images, test_labels))
上面的3个model.fit()调用给我以下结果:
1563/1563 [==============================] - 7s 4ms/step - loss: 1.4939 - accuracy: 0.4543 - val_loss: 1.2516 - val_accuracy: 0.5567
1563/1563 [==============================] - 6s 4ms/step - loss: 1.6071 - accuracy: 0.4092 - val_loss: 1.3857 - val_accuracy: 0.4951
1563/1563 [==============================] - 7s 4ms/step - loss: 1.5538 - accuracy: 0.4325 - val_loss: 1.3187 - val_accuracy: 0.5294
产生这种差异的原因是什么?我试图了解可能阻碍从模型再现结果的来源。除了随机种子,密集层初始化之外,我还缺少什么?
您测试可重复性的方法不正确。您需要关闭程序并重新运行,以查看结果是否相同。否则,运行2取决于运行1期间发生的事件,而运行3取决于运行1和2期间发生的事件。
原因是Tensorflow维护一个内部计数器来随机生成,如tf.random.set_seed
(重点是我的)的文档中所述:
print(tf.random.uniform([1], seed=1)) # generates 'A1' print(tf.random.uniform([1], seed=1)) # generates 'A2'
我们在上面的第二次tf.random.uniform调用中获得'A2'而不是'A1'的原因是因为TensorFlow使用相同的tf.random.uniform内核(即内部表示)对所有具有相同参数的调用,内核维护一个内部计数器,该计数器在每次执行时都会增加,从而产生不同的结果。
如果我只运行程序的第一次运行两次,然后在每次运行之间关闭程序(在这种情况下为IPython),则会得到:
In [1]: run program.py
1563/1563 [==============================] - 13s 8ms/step - loss: 1.4997 - accuracy: 0.4540 - val_loss: 1.2528 - val_accuracy: 0.5494
{'loss': [1.4996991157531738], 'accuracy': [0.4540199935436249], 'val_loss': [1.2527965307235718], 'val_accuracy': [0.5493999719619751]}
In [2]: run program.py
1563/1563 [==============================] - 12s 8ms/step - loss: 1.4997 - accuracy: 0.4540 - val_loss: 1.2528 - val_accuracy: 0.5494
{'loss': [1.4996991157531738], 'accuracy': [0.4540199935436249], 'val_loss': [1.2527965307235718], 'val_accuracy': [0.5493999719619751]}
减去执行计算所需的时间(可能会有所不同,具体取决于计算机上的负载),结果是完全相同的。
本文收集自互联网,转载请注明来源。
如有侵权,请联系[email protected] 删除。
我来说两句