I am using python 3 with anaconda, and tensorflow 1.12 with eager eval.
I am using it to create a triplet loss function for a siamese network, and need to calculate distance between different data samples.
I created a function in order to create the distance calculation, but no matter what I do, when I try to calculate it's gradient with respect to the networks output, It keeps giving me all nan gradient.
This is the code:
def matrix_row_wise_norm(matrix):
import tensorflow as tf
tensor = tf.expand_dims(matrix, -1)
tensor = tf.transpose(tensor, [0, 2, 1]) - tf.transpose(tensor, [2, 0, 1])
norm = tf.norm(tensor, axis=2)
return norm
In the loss function I am using
def loss(y_true, p_pred):
with tf.GradientTape() as t:
t.watch(y_pred)
distance_matrix = matrix_row_wise_norm(y_pred)
grad = t.gradient(distance_matrix, y_pred)
And the grad is all nan
s.
I checked that y_pred
is made of legit values - and it does.
I tried to create a gradient of y_pred * 2
with respect to itself and got legitimate gradient values.
What am I missing here? Is the indexing in the creation of the distance matrix problematic?
edit:
the dtype of both y_pred
and loss
is tf.float32
edit: found an open bug report in tf - could this be the issue?
edit:
When I change the norm axis to 0 or 1, I am getting legitimate values and nothing goes to nan
. The operation I am getting using norm with axis=2
is the pairwise distance between the pairs of rows in the matrix, I suspected this might have something to do with 0 distance between a row to itself, so I clipped the values with min value of 1e-7 without any luck.
Thanks
Seems that tf.norm suffers from numeric instability as explained here
They also suggest using l2 norm that is more numeric stable, So I tried that, also getting nan values, thanks to 0 gradients. So I used those together with gradient clipping, so far so good, the loss function is working and manages to converge.
def last_attempt(y_true, y_pred):
import tensorflow as tf
import numpy as np
loss = tf.zeros(1)
for i in range(y_pred.shape[0]):
dist = tf.gather(y_pred, [i], axis=0)
y = y_true.numpy().squeeze()
norm = tf.map_fn(tf.nn.l2_loss, dist-y_pred)
d = norm.numpy()
d[np.where(y != y[i])] = 0.0
max_pos = tf.gather(norm, np.argmax(d))
d = norm.numpy()
d[np.where(y == y[i])] = np.inf
min_neg = tf.gather(norm, np.argmin(d))
loss += tf.clip_by_value(max_pos - min_neg + tf.constant(1, dtype=tf.float32),
1e-8, 1e1)
return loss
There is much room for optimizing that function, here is a reference to my other SO question - working on that.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With