我正在使用Keras和Python进行分类然后对象检测。我已经对猫/狗进行了80%以上的准确度分类,我现在的结果还不错。我的问题是如何从输入图像中检测猫或狗?我完全糊涂了。我想使用自己的高度,而不是来自互联网的预训练。
这是我目前的代码:
from keras.preprocessing.image import ImageDataGenerator
from keras.models import Sequential
from keras.layers import Convolution2D, MaxPooling2D
from keras.layers import Activation, Dropout, Flatten, Dense
import numpy as np
import matplotlib.pyplot as plt
import matplotlib
from keras.preprocessing.image import ImageDataGenerator, array_to_img, img_to_array, load_img
#########################################################################################################
#VALUES
# dimensions of our images.
img_width, img_height = 150, 150
train_data_dir = 'data/train'
validation_data_dir = 'data/validation'
nb_train_samples = 2000 #1000 cats/dogs
nb_validation_samples = 800 #400cats/dogs
nb_epoch = 50
#########################################################################################################
#MODEL
model = Sequential()
model.add(Convolution2D(32, 3, 3, input_shape=(3, img_width, img_height)))
model.add(Activation('relu'))
model.add(MaxPooling2D(pool_size=(2, 2)))
model.add(Convolution2D(32, 3, 3))
model.add(Activation('relu'))
model.add(MaxPooling2D(pool_size=(2, 2)))
model.add(Convolution2D(64, 3, 3))
model.add(Activation('relu'))
model.add(MaxPooling2D(pool_size=(2, 2)))
model.add(Flatten())
model.add(Dense(64))
model.add(Activation('relu'))
model.add(Dropout(0.5))
model.add(Dense(1))
model.add(Activation('sigmoid'))
model.compile(loss='binary_crossentropy',
optimizer='rmsprop',
metrics=['accuracy'])
# this is the augmentation configuration we will use for training
train_datagen = ImageDataGenerator(
rescale=1./255,
shear_range=0.2,
zoom_range=0.2,
horizontal_flip=True)
##########################################################################################################
#TEST AUGMENTATION
img = load_img('data/train/cats/cat.0.jpg') # this is a PIL image
x = img_to_array(img) # this is a Numpy array with shape (3, 150, 150)
x = x.reshape((1,) + x.shape) # this is a Numpy array with shape (1, 3, 150, 150)
# the .flow() command below generates batches of randomly transformed images
# and saves the results to the `preview/` directory
i = 0
for batch in train_datagen.flow(x, batch_size=1,
save_to_dir='data/TEST AUGMENTATION', save_prefix='cat', save_format='jpeg'):
i += 1
if i > 20:
break # otherwise the generator would loop indefinitely
##########################################################################################################
# this is the augmentation configuration we will use for testing:
# only rescaling
test_datagen = ImageDataGenerator(rescale=1./255)
#PREPARE TRAINING DATA
train_generator = train_datagen.flow_from_directory(
train_data_dir, #data/train
target_size=(img_width, img_height), #RESIZE to 150/150
batch_size=32,
class_mode='binary') #since we are using binarycrosentropy need binary labels
#PREPARE VALIDATION DATA
validation_generator = test_datagen.flow_from_directory(
validation_data_dir, #data/validation
target_size=(img_width, img_height), #RESIZE 150/150
batch_size=32,
class_mode='binary')
#START model.fit
history =model.fit_generator(
train_generator, #train data
samples_per_epoch=nb_train_samples,
nb_epoch=nb_epoch,
validation_data=validation_generator, #validation data
nb_val_samples=nb_validation_samples)
model.save_weights('savedweights.h5')
# list all data in history
print(history.history.keys())
#ACC VS VAL_ACC
plt.plot(history.history['acc'])
plt.plot(history.history['val_acc'])
plt.title('model accuracy ACC VS VAL_ACC')
plt.ylabel('accuracy')
plt.xlabel('epoch')
plt.legend(['train', 'test'], loc='upper left')
plt.show()
# summarize history for loss
#LOSS VS VAL_LOSS
plt.plot(history.history['loss'])
plt.plot(history.history['val_loss'])
plt.title('model loss LOSS vs VAL_LOSS')
plt.ylabel('loss')
plt.xlabel('epoch')
plt.legend(['train', 'test'], loc='upper left')
plt.show()
model.load_weights('first_try.h5')
所以现在,因为我对猫和狗进行了分类,我需要做什么以及如何输入图像并通过它来找到带有边界框的猫或狗?我对这个全新,甚至不确定我是否以正确的方式解决这个问题? 谢谢。
更新 嗨,很抱歉发布这么晚的结果,几天都无法解决这个问题。 我导入一个图像并将其重塑为1,3,150,150形状,因为150,150形状带来错误:
Exception: Error when checking : expected convolution2d_input_1 to have 4 dimensions, but got array with shape (150L, 150L)
导入图片:
#load test image
img=load_img('data/prediction/cat.155.jpg')
#reshape to 1,3,150,150
img = np.arange(1* 150 * 150).reshape((1,3,150, 150))
#check shape
print(img.shape)
然后我将def predict_function(x)更改为:
def predict_function(x):
# example of prediction function for simplicity, you
# should probably use `return model.predict(x)`
# random.seed(x[0][0])
# return random.random()
return model.predict(img)
现在我跑的时候:
best_box = get_best_bounding_box(img, predict_function)
print('best bounding box %r' % (best_box, ))
我将输出作为最佳边界框:无
所以我只跑了:
model.predict(img)
并得到以下内容:
model.predict(img)
Out[54]: array([[ 0.]], dtype=float32)
所以它根本不检查它是猫还是狗......有什么想法吗?
注意:当def预测时)函数(x)正在使用:
random.seed(x[0][0])
return random.random()
我确实得到输出,它复选框并给出最好的输出。
答案 0 :(得分:17)
您构建的机器学习模型和您尝试实现的任务不尽相同。模型尝试解决分类任务,同时您的目标是检测图像中的对象,即object detection task。
分类有一个布尔问题,而检测问题有两个以上的答案答案。
我可以建议你尝试三种可能性:
定义尺寸的裁剪框(例如从20X20到160X160)并使用滑动窗口。对于每个窗口,尝试预测它的概率,最后采用你预测的最大窗口。
这将为边界框生成多个候选项,您将使用最高概率选择边界框。
这可能很慢,因为我们需要预测数百个+样本。
另一个选择是尝试在您的网络上实施RCNN(another link)或Faster-RCNN网络。这些网络基本上减少了候选边界框窗口的数量。
以下代码演示了如何进行滑动窗口算法。你可以改变参数。
import random
import numpy as np
WINDOW_SIZES = [i for i in range(20, 160, 20)]
def get_best_bounding_box(img, predict_fn, step=10, window_sizes=WINDOW_SIZES):
best_box = None
best_box_prob = -np.inf
# loop window sizes: 20x20, 30x30, 40x40...160x160
for win_size in window_sizes:
for top in range(0, img.shape[0] - win_size + 1, step):
for left in range(0, img.shape[1] - win_size + 1, step):
# compute the (top, left, bottom, right) of the bounding box
box = (top, left, top + win_size, left + win_size)
# crop the original image
cropped_img = img[box[0]:box[2], box[1]:box[3]]
# predict how likely this cropped image is dog and if higher
# than best save it
print('predicting for box %r' % (box, ))
box_prob = predict_fn(cropped_img)
if box_prob > best_box_prob:
best_box = box
best_box_prob = box_prob
return best_box
def predict_function(x):
# example of prediction function for simplicity, you
# should probably use `return model.predict(x)`
random.seed(x[0][0])
return random.random()
# dummy array of 256X256
img = np.arange(256 * 256).reshape((256, 256))
best_box = get_best_bounding_box(img, predict_function)
print('best bounding box %r' % (best_box, ))
示例输出:
predicting for box (0, 0, 20, 20)
predicting for box (0, 10, 20, 30)
predicting for box (0, 20, 20, 40)
...
predicting for box (110, 100, 250, 240)
predicting for box (110, 110, 250, 250)
best bounding box (140, 80, 160, 100)
您可以查看pascal dataset(examples here),其中包含20个班级,其中两个是猫狗。
数据集包含作为Y目标的对象的位置。
最后但并非最不重要的是,您可以重复使用现有的网络,甚至可以进行知识转移" (keras示例)为您的特定任务。
查看以下convnets-keras
lib。
所以选择最好的方法来更新我们的结果。