def _test_op_slice_sequence(input_data, slice_params, expected_result, device_id, precision): input_data = AA(input_data, dtype=PRECISION_TO_TYPE[precision]) t = Axis.new_unique_dynamic_axis('t') sample_shape = input_data.shape[1:] a = I(shape=sample_shape, dtype=sanitize_dtype_cntk(PRECISION_TO_TYPE[precision]), needs_gradient=True, dynamic_axes=[Axis.default_batch_axis(), t], name='a') result = C.sequence.slice(a, begin_index=slice_params[0], end_index=slice_params[1]) def grad_slice(x, beg_index, end_index): res = np.zeros_like(x) res[beg_index:end_index] = 1 return res expected_gradient = grad_slice(np.asarray(input_data), *slice_params) expected_forward = AA([expected_result], dtype=PRECISION_TO_TYPE[precision]) expected_backward = { a: [grad_slice(np.asarray(input_data), *slice_params)] } # create batch input_data.shape = (1,) + input_data.shape forward_input = {a: input_data} unittest_helper(result, forward_input, expected_forward, expected_backward, device_id=device_id, precision=precision)
def _graph_dict(): # This function creates a graph that has no real meaning other than # providing something to traverse. d = {} batch_axis = Axis.default_batch_axis() input_seq_axis = Axis('ia') input_dynamic_axes = [batch_axis, input_seq_axis] d['i1'] = input_variable( shape=(2, 3), dynamic_axes=input_dynamic_axes, name='i1') d['c1'] = constant(shape=(2, 3), value=6, name='c1') d['p1'] = parameter(shape=(3, 2), init=7, name='p1') d['op1'] = plus(d['i1'], d['c1'], name='op1') d['op2'] = times(d['op1'], d['p1'], name='op2') #d['slice'] = slice(d['c1'], Axis.default_dynamic_axis(), 0, 3) #label_sentence_start = sequence.first(raw_labels) # no name d['p2'] = parameter(shape=(2, 2)) # duplicate names d['op3a'] = plus(d['op2'], d['p2'], name='op3') d['op3b'] = plus(d['op3a'], d['p2'], name='op3') d['first'] = sequence.first(d['op3b'], name='past') d['root'] = d['first'] return d
def test_op_broadcast_as(device_id, precision): from .. import sequence a_data = [ AA([1], dtype=PRECISION_TO_TYPE[precision]), AA([2], dtype=PRECISION_TO_TYPE[precision]), AA([3], dtype=PRECISION_TO_TYPE[precision]) ] b_data = [ AA([[2]], dtype=PRECISION_TO_TYPE[precision]), AA([[2], [3]], dtype=PRECISION_TO_TYPE[precision]), AA([[2], [3], [4]], dtype=PRECISION_TO_TYPE[precision]) ] a = I(shape=(1, ), dtype=sanitize_dtype_cntk(PRECISION_TO_TYPE[precision]), name='a', dynamic_axes=[Axis.default_batch_axis()]) b = I(shape=(1, ), dtype=sanitize_dtype_cntk(PRECISION_TO_TYPE[precision]), name='b') broadcast_a_as_b = sequence.broadcast_as(a, b) res = broadcast_a_as_b.eval({a: a_data, b: b_data}) assert np.array_equal(res[0], np.asarray([[1.]])) assert np.array_equal(res[1], np.asarray([[2.], [2.]])) assert np.array_equal(res[2], np.asarray([[3.], [3.], [3.]]))
def test_op_broadcast_as(device_id, precision): from .. import sequence a_data = [AA([1], dtype=PRECISION_TO_TYPE[precision]), AA([2], dtype=PRECISION_TO_TYPE[precision]), AA([3], dtype=PRECISION_TO_TYPE[precision])] b_data = [AA([[2]], dtype=PRECISION_TO_TYPE[precision]), AA([[2], [3]], dtype=PRECISION_TO_TYPE[precision]), AA([[2], [3], [4]], dtype=PRECISION_TO_TYPE[precision])] a = I(shape=(1,), dtype=sanitize_dtype_cntk(PRECISION_TO_TYPE[precision]), name='a', dynamic_axes=[Axis.default_batch_axis()]) b = I(shape=(1,), dtype=sanitize_dtype_cntk(PRECISION_TO_TYPE[precision]), name='b') broadcast_a_as_b = sequence.broadcast_as(a, b) res = broadcast_a_as_b.eval({a: a_data, b: b_data}) assert np.array_equal(res[0], np.asarray([[1.]])) assert np.array_equal(res[1], np.asarray([[2.], [2.]])) assert np.array_equal(res[2], np.asarray([[3.], [3.], [3.]]))
def test_op_broadcast_as_in_loop(device_id): from .. import sequence, placeholder_variable, past_value a_data = [AA([1]), AA([2]), AA([3])] b_data = [AA([[2]]), AA([[2], [3]]), AA([[2], [3], [4]])] a = I(shape=(1, ), name='a', dynamic_axes=[Axis.default_batch_axis()]) b = I(shape=(1, ), name='b') out_placeholder = placeholder_variable() out_delayed = past_value(out_placeholder, time_step=5) out_delayed_plus_b = out_delayed + b out = sequence.broadcast_as(a, out_delayed_plus_b) out.replace_placeholder(out) res = out.eval({a: a_data, b: b_data}) assert np.array_equal(res[0], np.asarray([[1.]])) assert np.array_equal(res[1], np.asarray([[2.], [2.]])) assert np.array_equal(res[2], np.asarray([[3.], [3.], [3.]]))
def test_op_broadcast_as_in_loop(device_id): from .. import sequence, placeholder_variable, past_value a_data = [AA([1]), AA([2]), AA([3])] b_data = [AA([[2]]), AA([[2], [3]]), AA([[2], [3], [4]])] a = I(shape=(1,), name='a', dynamic_axes=[Axis.default_batch_axis()]) b = I(shape=(1,), name='b') out_placeholder = placeholder_variable() out_delayed = past_value(out_placeholder, time_step=5) out_delayed_plus_b = out_delayed + b out = sequence.broadcast_as(a, out_delayed_plus_b) out.replace_placeholder(out) res = out.eval({a: a_data, b: b_data}) assert np.array_equal(res[0], np.asarray([[1.]])) assert np.array_equal(res[1], np.asarray([[2.], [2.]])) assert np.array_equal(res[2], np.asarray([[3.], [3.], [3.]]))
def __init__(self, in_shape, output_shape, device_id=None, learning_rate=0.00025, momentum=0.9, minibatch_size=32, update_interval=10000, n_workers=1, visualizer=None): """ Q Neural Network following Mnih and al. implementation and default options. The network has the following topology: Convolution(32, (8, 8)) Convolution(64, (4, 4)) Convolution(64, (2, 2)) Dense(512) :param in_shape: Shape of the observations perceived by the learner (the neural net input) :param output_shape: Size of the action space (mapped to the number of output neurons) :param device_id: Use None to let CNTK select the best available device, -1 for CPU, >= 0 for GPU (default: None) :param learning_rate: Learning rate (default: 0.00025, as per Mnih et al.) :param momentum: Momentum, provided as momentum value for averaging gradients without unit gain filter Note that CNTK does not currently provide an implementation of Graves' RmsProp with momentum. It uses AdamSGD optimizer instead. (default: 0, no momentum with RProp optimizer) :param minibatch_size: Minibatch size (default: 32, as per Mnih et al.) :param n_workers: Number of concurrent worker for distributed training. (default: 1, not distributed) :param visualizer: Optional visualizer allowing the model to save summary data (default: None, no visualization) Ref: Mnih et al.: "Human-level control through deep reinforcement learning." Nature 518.7540 (2015): 529-533. """ assert learning_rate > 0, 'learning_rate should be > 0' assert 0. <= momentum < 1, 'momentum should be 0 <= momentum < 1' QModel.__init__(self, in_shape, output_shape) CntkModel.__init__(self, device_id, False, n_workers, visualizer) self._nb_actions = output_shape self._steps = 0 self._target_update_interval = update_interval self._target = None # Input vars self._environment = input(in_shape, name='env', dynamic_axes=(Axis.default_batch_axis())) self._q_targets = input(1, name='q_targets', dynamic_axes=(Axis.default_batch_axis())) self._actions = input(output_shape, name='actions', dynamic_axes=(Axis.default_batch_axis())) # Define the neural network graph self._model = self._build_model()(self._environment) self._target = self._model.clone( CloneMethod.freeze, {self._environment: self._environment}) # Define the learning rate lr_schedule = learning_rate_schedule(learning_rate, UnitType.minibatch) # AdamSGD optimizer m_schedule = momentum_schedule(momentum) vm_schedule = momentum_schedule(0.999) l_sgd = adam(self._model.parameters, lr_schedule, momentum=m_schedule, unit_gain=True, variance_momentum=vm_schedule) if self.distributed_training: raise NotImplementedError('ASGD not implemented yet.') # _actions is a sparse 1-hot encoding of the actions done by the agent q_acted = reduce_sum(self._model * self._actions, axis=0) # Define the trainer with Huber Loss function criterion = huber_loss(q_acted, self._q_targets, 1.0) self._learner = l_sgd self._trainer = Trainer(self._model, (criterion, None), l_sgd)
def __init__(self, in_shape, output_shape, device_id=None, learning_rate=0.00025, momentum=0.9, minibatch_size=32, update_interval=10000, n_workers=1, visualizer=None): """ Q Neural Network following Mnih and al. implementation and default options. The network has the following topology: Convolution(32, (8, 8)) Convolution(64, (4, 4)) Convolution(64, (2, 2)) Dense(512) :param in_shape: Shape of the observations perceived by the learner (the neural net input) :param output_shape: Size of the action space (mapped to the number of output neurons) :param device_id: Use None to let CNTK select the best available device, -1 for CPU, >= 0 for GPU (default: None) :param learning_rate: Learning rate (default: 0.00025, as per Mnih et al.) :param momentum: Momentum, provided as momentum value for averaging gradients without unit gain filter Note that CNTK does not currently provide an implementation of Graves' RmsProp with momentum. It uses AdamSGD optimizer instead. (default: 0, no momentum with RProp optimizer) :param minibatch_size: Minibatch size (default: 32, as per Mnih et al.) :param n_workers: Number of concurrent worker for distributed training. (default: 1, not distributed) :param visualizer: Optional visualizer allowing the model to save summary data (default: None, no visualization) Ref: Mnih et al.: "Human-level control through deep reinforcement learning." Nature 518.7540 (2015): 529-533. """ assert learning_rate > 0, 'learning_rate should be > 0' assert 0. <= momentum < 1, 'momentum should be 0 <= momentum < 1' QModel.__init__(self, in_shape, output_shape) CntkModel.__init__(self, device_id, False, n_workers, visualizer) self._nb_actions = output_shape self._steps = 0 self._target_update_interval = update_interval self._target = None # Input vars self._environment = input(in_shape, name='env', dynamic_axes=(Axis.default_batch_axis())) self._q_targets = input(1, name='q_targets', dynamic_axes=(Axis.default_batch_axis())) self._actions = input(output_shape, name='actions', dynamic_axes=(Axis.default_batch_axis())) # Define the neural network graph self._model = self._build_model()(self._environment) self._target = self._model.clone( CloneMethod.freeze, {self._environment: self._environment} ) # Define the learning rate lr_schedule = learning_rate_schedule(learning_rate, UnitType.minibatch) # AdamSGD optimizer m_schedule = momentum_schedule(momentum) vm_schedule = momentum_schedule(0.999) l_sgd = adam(self._model.parameters, lr_schedule, momentum=m_schedule, unit_gain=True, variance_momentum=vm_schedule) if self.distributed_training: raise NotImplementedError('ASGD not implemented yet.') # _actions is a sparse 1-hot encoding of the actions done by the agent q_acted = reduce_sum(self._model * self._actions, axis=0) # Define the trainer with Huber Loss function criterion = huber_loss(q_acted, self._q_targets, 1.0) self._learner = l_sgd self._trainer = Trainer(self._model, (criterion, None), l_sgd)