Python 依赖于批量大小的Keras损失函数

Python 依赖于批量大小的Keras损失函数,python,machine-learning,keras,Python,Machine Learning,Keras,我试图在Keras中构造一个损失函数,在这个函数中,我惩罚预测和一组给定值之间的最小距离。问题是我需要计算预测值和给定值之间的距离 示例代码 def custom_loss(y_pred,y_test): #Given values centers=K.constant([[-2.5,-1],[-1.25,-2],[.5,-1],[1.5,.25]]) num_centers=K.int_shape(centers)[0] #Begin construct

我试图在Keras中构造一个损失函数,在这个函数中,我惩罚预测和一组给定值之间的最小距离。问题是我需要计算预测值和给定值之间的距离

示例代码

def custom_loss(y_pred,y_test):


    #Given values
    centers=K.constant([[-2.5,-1],[-1.25,-2],[.5,-1],[1.5,.25]])
    num_centers=K.int_shape(centers)[0]


    #Begin constructing distance matrix
    height=K.int_shape(y_pred)[0]
    i=0
    current_center=K.reshape(K.repeat(K.reshape(centers[i,:],[1,-1]),height),[height,2])
    current_dist=K.sqrt(K.sum(K.square(y_pred-current_center),axis=1))


    #Values of distance matrix for first center
    Distance=K.reshape(current_dist,[height,1])


    for i in range(1,num_centers):
        current_center=K.reshape(K.repeat(K.reshape(centers[i,:],[1,-1]),height),[height,2])
        current_dist=K.sqrt(K.sum(K.square(y_pred-current_center),axis=-1))
        current_dist=K.reshape(current_dist,[height,1])


        #Iteratively concatenate distances of y_pred from remaining centers
        Distance=K.concatenate([Distance,current_dist],axis=-1)

    #Determine minimum distance from each predicted value to nearest center
    A=K.min(A,axis=1)


    #Return average minimum distance as loss
    return K.sum(A)/float(height)
但是,我无法消除函数对y_pred的第一维的依赖性,它是可变的。我使用数组广播来计算y_pred和每个给定值之间的差异,但我使用批大小显式广播,因为我不知道如果不在Keras中使用批大小,怎么做。但是,这会产生一个错误,因为在构造计算图时,批大小不明确


我如何避免直截了当的广播?有没有更有效的方法来计算这个距离矩阵,因为当前的方法非常笨拙?

您的损失函数可以使用隐式广播实现,如下所示:

import keras.backend as K


def custom_loss(y_true, y_pred):
    centers = K.constant([[-2.5, -1], [-1.25, -2], [.5, -1], [1.5, .25]])

    # Expand dimensions to enable implicit broadcasting
    y_pred_r = y_pred[:, None, :]  # Shape: (batch_size, 1, 2)
    centers_r = centers[None, :, :]  # Shape: (1, nb_centers, 2)

    # Compute minimum distance to centers for each element
    distances = K.sqrt(K.sum(K.square(y_pred_r - centers_r), axis=-1))  # Shape=(batch_size, nb_centers)
    min_distances = K.min(distances, axis=-1)  # Shape=(batch_size,)

    # Output average of minimum distances
    return K.mean(min_distances)

注:未测试。

y_pred的尺寸为[?,2],其中第一个尺寸表示未指定的批次尺寸。