keras 源碼中下載MNIST的方式是 path = get_file(path, origin='https://s3.amazonaws.com/img-datasets/mnist.npz'),數(shù)據(jù)源是通過 url = https://s3.amazonaws.com/img-datasets/mnist.npz 進(jìn)行下載的。訪問該 url 地址被墻了,導(dǎo)致 MNIST 相關(guān)的案例都卡在數(shù)據(jù)下載的環(huán)節(jié)。本文主要提供解決方案,讓需要的讀者可以跑案例的代碼感受一下。
下載 mnist.npz 數(shù)據(jù)集
本文使用的 mnist.npz 數(shù)據(jù)集是通過一個(gè) japan 的服務(wù)器下載得到的,在此免費(fèi)分享給大家。如果下載有問題的話,可以留言哈。
下載鏈接:https://pan.baidu.com/s/1jH6uFFC密碼: dw3d
下載后文件放在*.py所在目錄下
改造原來(lái)的文件
#load data
#(X_train, y_train), (X_test, y_test) = mnist.load_data()
import numpy as np
path='./mnist.npz'
f = np.load(path)
X_train, y_train = f['x_train'], f['y_train']
X_test, y_test = f['x_test'], f['y_test']
f.close()
運(yùn)行效果如下所示:
45568/48000 [===========================>..] - ETA: 4s - loss: 0.0574 - acc: 0.9818
45696/48000 [===========================>..] - ETA: 4s - loss: 0.0576 - acc: 0.9817
45824/48000 [===========================>..] - ETA: 4s - loss: 0.0576 - acc: 0.9817
45952/48000 [===========================>..] - ETA: 3s - loss: 0.0576 - acc: 0.9817
46080/48000 [===========================>..] - ETA: 3s - loss: 0.0576 - acc: 0.9817
46208/48000 [===========================>..] - ETA: 3s - loss: 0.0577 - acc: 0.9817
46336/48000 [===========================>..] - ETA: 3s - loss: 0.0576 - acc: 0.9817
46464/48000 [============================>.] - ETA: 2s - loss: 0.0578 - acc: 0.9817
46592/48000 [============================>.] - ETA: 2s - loss: 0.0577 - acc: 0.9817
46720/48000 [============================>.] - ETA: 2s - loss: 0.0578 - acc: 0.9817
46848/48000 [============================>.] - ETA: 2s - loss: 0.0578 - acc: 0.9817
46976/48000 [============================>.] - ETA: 1s - loss: 0.0579 - acc: 0.9817
47104/48000 [============================>.] - ETA: 1s - loss: 0.0580 - acc: 0.9816
47232/48000 [============================>.] - ETA: 1s - loss: 0.0580 - acc: 0.9816
47360/48000 [============================>.] - ETA: 1s - loss: 0.0580 - acc: 0.9817
47488/48000 [============================>.] - ETA: 0s - loss: 0.0579 - acc: 0.9817
47616/48000 [============================>.] - ETA: 0s - loss: 0.0578 - acc: 0.9817
47744/48000 [============================>.] - ETA: 0s - loss: 0.0578 - acc: 0.9817
47872/48000 [============================>.] - ETA: 0s - loss: 0.0577 - acc: 0.9818
48000/48000 [==============================] - 100s 2ms/step - loss: 0.0577 - acc: 0.9817 - val_loss: 0.0447 - val_acc: 0.9862
Test score: 0.03886812744811614
Test accuracy: 0.9869
完整代碼如下:
# -*- coding: utf-8 -*-
__author__ = "TF大Q"
import numpy as np
np.random.seed(1337)
from keras.datasets import mnist
from keras.models import Sequential
from keras.layers import Dense, Dropout, Activation, Flatten
from keras.layers import Convolution2D, MaxPooling2D
from keras.utils import np_utils
from keras import backend as K
# 全局變量
batch_size = 128
nb_classes = 10
epochs = 5
# input image dimensions
img_rows, img_cols = 28, 28
# 卷積濾波器的數(shù)量
nb_filters = 32
# size of pooling area for max pooling
pool_size = (2, 2)
# convolution kernel size
kernel_size = (3, 3)
#load data
#(X_train, y_train), (X_test, y_test) = mnist.load_data()
import numpy as np
path='./mnist.npz'
f = np.load(path)
X_train, y_train = f['x_train'], f['y_train']
X_test, y_test = f['x_test'], f['y_test']
f.close()
#from sklearn.cross_validation import train_test_split
from sklearn.model_selection import? train_test_split
#x為數(shù)據(jù)集的feature熟悉,y為label.
X_train, X_valid, y_train, y_valid = train_test_split(X_train, y_train, test_size = 0.2)
# 根據(jù)不同的backend定下不同的格式
if K.image_dim_ordering() == 'th':
? ? X_train = X_train.reshape(X_train.shape[0], 1, img_rows, img_cols)
? ? X_test = X_test.reshape(X_test.shape[0], 1, img_rows, img_cols)
? ? X_valid = X_valid.reshape(X_valid.shape[0], 1, img_rows, img_cols)
? ? input_shape = (1, img_rows, img_cols)
else:
? ? X_train = X_train.reshape(X_train.shape[0], img_rows, img_cols, 1)
? ? X_test = X_test.reshape(X_test.shape[0], img_rows, img_cols, 1)
? ? X_valid = X_valid.reshape(X_valid.shape[0], img_rows, img_cols,1)
? ? input_shape = (img_rows, img_cols, 1)
# 類型轉(zhuǎn)換
X_train = X_train.astype('float32')
X_test = X_test.astype('float32')
X_valid = X_valid.astype('float32')
X_train /= 255
X_test /= 255
X_valid /= 255
print('X_train shape:', X_train.shape)
print(X_train.shape[0], 'train samples')
print(X_test.shape[0], 'test samples')
print(X_valid.shape[0], 'valid samples')
# 轉(zhuǎn)換為one_hot類型
Y_train = np_utils.to_categorical(y_train, nb_classes)
Y_test = np_utils.to_categorical(y_test, nb_classes)
Y_valid = np_utils.to_categorical(y_valid, nb_classes)
#構(gòu)建模型
model = Sequential()
model.add(Convolution2D(nb_filters, (kernel_size[0], kernel_size[1]),
? ? ? ? ? ? ? ? ? ? ? ? padding='same',
? ? ? ? ? ? ? ? ? ? ? ? input_shape=input_shape)) # 卷積層1
model.add(Activation('relu')) #激活層
model.add(Convolution2D(nb_filters, (kernel_size[0], kernel_size[1]))) #卷積層2
model.add(Activation('relu')) #激活層
model.add(MaxPooling2D(pool_size=pool_size)) #池化層
model.add(Dropout(0.25)) #神經(jīng)元隨機(jī)失活
model.add(Flatten()) #拉成一維數(shù)據(jù)
model.add(Dense(128)) #全連接層1
model.add(Activation('relu')) #激活層
model.add(Dropout(0.5)) #隨機(jī)失活
model.add(Dense(nb_classes)) #全連接層2
model.add(Activation('softmax')) #Softmax評(píng)分
#編譯模型
model.compile(loss='categorical_crossentropy',
? ? ? ? ? ? ? optimizer='adadelta',
? ? ? ? ? ? ? metrics=['accuracy'])
#訓(xùn)練模型
model.fit(X_train, Y_train, batch_size=batch_size, epochs=epochs,
? ? ? ? ? verbose=1, validation_data=(X_valid, Y_valid))
#評(píng)估模型
score = model.evaluate(X_test, Y_test, verbose=0)
print('Test score:', score[0])
print('Test accuracy:', score[1])