I have the following code for a convolutional layer on TensorFlow. This layer is part of a larger computational graph.
# Define the shape of the filter
filter_shape = [1,
config.char_filter_size,
config.dim_char,
config.dim_char]
# Define the convolutional layer weights and biases
W_conv = tf.Variable(tf.truncated_normal(filter_shape, stddev=0.1),
name="W_conv")
b_conv = tf.Variable(tf.constant(0.1, shape=[config.dim_char]),
name="b_conv")
# Do 2d convolution
conv = tf.nn.conv2d(char_embeddings,
W_conv,
strides=[1, 1, 1, 1],
padding="VALID",
name="conv")
# Apply nonlinearity
# h_conv has the same shape as conv
h_conv = tf.nn.relu(tf.nn.bias_add(conv, b_conv),
name="conv_relu")
# Maxpooling h_conv over dim 2 (char dim)
# ERROR HERE
conv_pooled = tf.nn.max_pool(h_conv,
ksize=[1, 1, tf.shape(h_conv)[-2], 1],
strides=[1, 1, 1, 1],
padding='VALID',
name="conv_max_pool")
When trying to run, I get the error:
TypeError: Expected int for argument 'ksize' not tf.Tensor shape=() dtype=int32.
is tf.nn.max_pool
unable to handle dynamic ksize
?
It seems like you simply want to find the largest value over one of the dimensions which may be of dynamic size.
If that is the case, you are probably better off using the tf.reduce_max()
function instead of tf.nn.max_pool()
.
tf.reduce_max(
h_conv,
axis=2,
keep_dims=True
)
I set keep_dims=True
because it corresponds to what you would get if max pooling worked, but it is probably easier to work with the result if you set keep_dims=False
.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With