Example #1
0
 def test_ids_dtype(self):
     params = [[[1.0, 1.0, 1.0, 1.0]], [[2.0, 2.0, 2.0]]]
     ids32 = tf.constant([[0], [1]], dtype=tf.int32)
     ids64 = tf.constant([[0], [1]], dtype=tf.int64)
     transforms = [lambda embed: _transform_embedding(4, embed)] * len(params)
     embeddings32 = adaptive_embedding_lookup(params, ids32, transforms)
     embeddings64 = adaptive_embedding_lookup(params, ids64, transforms)
     self.assertAllEqual(self.evaluate(embeddings32), self.evaluate(embeddings64))
Example #2
0
    def test_adaptive_shape(self):
        params = [[[1.0, 1.0, 1.0, 1.0]], [[2.0, 2.0, 2.0]]]
        ids = [[0], [1]]
        transforms = [lambda embed: _transform_embedding(4, embed)] * len(params)
        embeddings = adaptive_embedding_lookup(params, ids, transforms)
        self.assertAllEqual(embeddings.shape, [2, 1, 4])

        actual_shape = tf.shape(embeddings)
        self.assertAllEqual(self.evaluate(actual_shape), [2, 1, 4])
Example #3
0
 def test_variables(self):
     p = [
         tf.Variable(tf.zeros(shape=[100, 100], dtype=tf.float32)),
         tf.Variable(tf.zeros(shape=[100, 100], dtype=tf.float32)),
     ]
     ids = tf.constant([0, 1, 1, 17], dtype=tf.int32)
     transforms = [lambda embed: _transform_embedding(100, embed)] * 2
     self.evaluate(tf.compat.v1.global_variables_initializer())
     self.evaluate(adaptive_embedding_lookup(p, ids, transforms))
Example #4
0
    def test_max_norm_nontrivial(self):
        params = [tf.constant([[2.0, 4.0], [3.0, 1.0]]), tf.constant([[2.0, 4.0], [3.0, 1.0]])]
        ids = tf.constant([0, 1], dtype=tf.int32)
        transforms = [lambda embed: _transform_embedding(2, embed)] * len(params)
        embeddings = adaptive_embedding_lookup(params, ids, transforms, max_norm=2.0)

        norms = tf.math.sqrt(tf.math.reduce_sum(embeddings * embeddings, axis=1))
        normalized = embeddings / tf.stack([norms, norms], axis=1)
        self.assertAllClose(self.evaluate(embeddings), 2 * self.evaluate(normalized))
Example #5
0
    def test_unknown_shape(self):
        params = [tf.constant([[1.0, 1.0, 1.0, 1.0]]), tf.constant([[2.0, 2.0, 2.0]])]
        params[0]._shape_val = tf.TensorShape([None, 4])

        ids = tf.constant([[0], [1]])
        transforms = [lambda embed: _transform_embedding(4, embed)] * len(params)
        embeddings = adaptive_embedding_lookup(params, ids, transforms)
        self.assertAllEqual(embeddings.shape, [2, 1, 4])

        actual_shape = tf.shape(embeddings)
        self.assertAllEqual(self.evaluate(actual_shape), [2, 1, 4])
Example #6
0
 def test_adaptive_error(self):
     with self.assertRaisesRegexp(ValueError, '2 variable'):
         self.evaluate(adaptive_embedding_lookup(
             [[0.]],
             [0],
             [lambda embed: _transform_embedding(2, embed)]
         ))
     with self.assertRaisesRegexp(ValueError, 'corresponding transform'):
         self.evaluate(adaptive_embedding_lookup([[[0.]], [[0.]]], [0], []))
     with self.assertRaisesRegexp(ValueError, 'should be callable'):
         self.evaluate(adaptive_embedding_lookup(
             [[[0.]], [[0.]]],
             [0],
             [lambda embed: _transform_embedding(2, embed), None]
         ))
     with self.assertRaisesRegexp(tf.errors.InvalidArgumentError, 'id should be less'):
         self.evaluate(adaptive_embedding_lookup(
             [[[1.0, 1.0, 1.0, 1.0]], [[2.0, 2.0, 2.0]]],
             [3],
             [lambda embed: _transform_embedding(4, embed)] * 2
         ))
Example #7
0
    def call(self, inputs):
        dtype = tf.keras.backend.dtype(inputs)
        if dtype not in {'int32', 'int64'}:
            inputs = tf.cast(inputs, 'int32')

        if isinstance(self.embeddings[0], sharded_variable.ShardedVariable):
            embeddings = [e.variables for e in self.embeddings]
        else:
            embeddings = self.embeddings

        out = adaptive_embedding_lookup(embeddings, inputs, self.projections)

        if self._dtype_policy.compute_dtype != self._dtype_policy.variable_dtype:
            # Instead of casting the variable as in most layers, cast the output, as
            # this is mathematically equivalent but is faster.
            out = tf.cast(out, self._dtype_policy.compute_dtype)

        return out
Example #8
0
    def test_adaptive_ragged(self):
        params = [
            [[1.0, 1.0, 1.0, 1.0],
             [1.1, 1.1, 1.1, 1.1],
             [1.2, 1.2, 1.2, 1.2],
             [1.3, 1.3, 1.3, 1.3]],
            [[2.0, 2.0, 2.0],
             [2.1, 2.1, 2.1],
             [2.2, 2.2, 2.2]],
            [[3.0, 3.0],
             [3.1, 3.1]],
            [[4.0]]
        ]
        ids = tf.ragged.constant([[0, 8], [3]])

        transforms = [lambda embed: _transform_embedding(4, embed)] * len(params)
        embeddings = adaptive_embedding_lookup(params, ids, transforms)
        expected = [
            [[1.0, 1.0, 1.0, 1.0],
             [3.1, 3.1, 0.0, 0.0]],
            [[1.3, 1.3, 1.3, 1.3]]
        ]
        self.assertAllClose(self.evaluate(embeddings), expected)
Example #9
0
 def test_max_norm_constant(self):
     params = [tf.constant([[2.0]]), tf.constant([[2.0]])]
     ids = tf.constant([0], dtype=tf.int32)
     transforms = [lambda embed: _transform_embedding(1, embed)] * len(params)
     embeddings = adaptive_embedding_lookup(params, ids, transforms, max_norm=1.0)
     self.assertAllEqual(self.evaluate(embeddings), [[1.0]])