Keras图层要求的形状与摘要中的形状不同

时间:2019-05-29 21:25:53

标签: keras deep-learning conv-neural-network

我正在喀拉拉邦编写一个U-net CNN,并尝试使用fit_generator进行培训。为了使它起作用,我使用了一个生成器脚本,该脚本可以为我的网络提供图像和标签(简单的拟合功能正在工作,但我想训练一个无法容纳到内存中的大数据集)。

我的问题是,在模型摘要中,它正确地说输出层具有以下形状:(无,288,512,4) https://i.imgur.com/69xG8pO.jpg

但是当我尝试实际训练时,出现此错误: https://i.imgur.com/j7H6sHX.jpg

我不知道为什么在摘要中keras希望(288,512,1)想要(288,512,4)

我用自己的unet代码进行了尝试,还从github复制了一个工作代码,但是它们都有完全相同的问题,这使我相信生成器脚本是弱链接。下面是我使用的代码(当我在以前的CNN中将它们与“ fit”一起使用时,此处使用的图像和标签数组功能已经起作用):

def generator(img_path, label_path, batch_size, height, width, num_classes):

    input_pairs = get_pairs(img_path, label_path) # rewrite if param name changes
    random.shuffle(input_pairs)

    iterate_pairs = itertools.cycle(input_pairs)

    while True:
        X = []
        Y = []
        for _ in range(batch_size):
            im, lab = next(iterate_pairs)

            appended_im = next(iter(im))
            appended_lab = next(iter(lab))

            X.append(input_image_array(appended_im, width, height))
            Y.append(input_label_array(appended_lab, width, height, num_classes, palette))

        yield (np.array(X), np.array(Y))

我尝试了生成器,提供的批次的形状为(批次大小为15): (15,288,512,3) (15,288,512,4) 所以我真的不知道这里可能是什么问题。

编辑:这是我使用的模型代码:

def conv_block(input_tensor, n_filter, kernel=(3, 3), padding='same', initializer="he_normal"):
    x = Conv2D(n_filter, kernel, padding=padding, kernel_initializer=initializer)(input_tensor)
    x = BatchNormalization()(x)
    x = Activation("relu")(x)
    x = Conv2D(n_filter, kernel, padding=padding, kernel_initializer=initializer)(x)
    x = BatchNormalization()(x)
    x = Activation("relu")(x)
    return x

def deconv_block(input_tensor, residual, n_filter, kernel=(3, 3), strides=(2, 2), padding='same'):
    y = Conv2DTranspose(n_filter, kernel, strides, padding)(input_tensor)
    y = concatenate([y, residual], axis=3)
    y = conv_block(y, n_filter)
    return y

# NETWORK - n_classes is the desired number of classes, filters are fixed
def Unet(input_height, input_width, n_classes=4, filters=64):

    # Downsampling
    input_layer = Input(shape=(input_height, input_width, 3), name='input')

    conv_1 = conv_block(input_layer, filters)
    conv_1_out = MaxPooling2D(pool_size=(2, 2))(conv_1)

    conv_2 = conv_block(conv_1_out, filters*2)
    conv_2_out = MaxPooling2D(pool_size=(2, 2))(conv_2)

    conv_3 = conv_block(conv_2_out, filters*4)
    conv_3_out = MaxPooling2D(pool_size=(2, 2))(conv_3)

    conv_4 = conv_block(conv_3_out, filters*8)
    conv_4_out = MaxPooling2D(pool_size=(2, 2))(conv_4)
    conv_4_drop = Dropout(0.5)(conv_4_out)

    conv_5 = conv_block(conv_4_drop, filters*16)
    conv_5_drop = Dropout(0.5)(conv_5)

    # Upsampling
    deconv_1 = deconv_block(conv_5_drop, conv_4, filters*8)
    deconv_1_drop = Dropout(0.5)(deconv_1)

    deconv_2 = deconv_block(deconv_1_drop, conv_3, filters*4)
    deconv_2_drop = Dropout(0.5)(deconv_2)

    deconv_3 = deconv_block(deconv_2_drop, conv_2, filters*2)
    deconv_3 = deconv_block(deconv_3, conv_1, filters)

    # Output - mapping each 64-component feature vector to number of classes
    output = Conv2D(n_classes, (1, 1))(deconv_3)
    output = BatchNormalization()(output)
    output = Activation("softmax")(output)

    # embed into functional API
    model = Model(inputs=input_layer, outputs=output, name="Unet")
    return model

1 个答案:

答案 0 :(得分:0)

将损失更改为categorical_crossentropy

  

使用sparse_categorical_crossentropy亏损时,您的目标   应该是整数目标。