forked from mpezeshki/RNN_Experiments
-
Notifications
You must be signed in to change notification settings - Fork 0
/
test_fork_lookup.py
66 lines (53 loc) · 1.74 KB
/
test_fork_lookup.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
import theano
from theano import tensor
from bricks import LookupTable
from blocks.bricks import FeedforwardSequence
from blocks.bricks.parallel import Fork
from blocks import initialization
from dataset import get_minibatch_char
from utils import parse_args
def build_fork_lookup(vocab_size, args):
x = tensor.lmatrix('features')
virtual_dim = 6
time_length = 5
mini_batch_size = 2
skip_connections = True
layers = 3
# Build the model
output_names = []
output_dims = []
for d in range(layers):
if d > 0:
suffix = '_' + str(d)
else:
suffix = ''
if d == 0 or skip_connections:
output_names.append("inputs" + suffix)
output_dims.append(virtual_dim)
print output_names
print output_dims
lookup = LookupTable(length=vocab_size, dim=virtual_dim)
lookup.weights_init = initialization.IsotropicGaussian(0.1)
lookup.biases_init = initialization.Constant(0)
fork = Fork(output_names=output_names, input_dim=time_length,
output_dims=output_dims,
prototype=FeedforwardSequence(
[lookup.apply]))
# Return list of 3D Tensor, one for each layer
# (Batch X Time X embedding_dim)
pre_rnn = fork.apply(x)
fork.initialize()
f = theano.function([x], pre_rnn)
return f
if __name__ == "__main__":
args = parse_args()
dataset = args.dataset
mini_batch_size = 2
time_length = 5
# Prepare data
train_stream, valid_stream, vocab_size = get_minibatch_char(
dataset, mini_batch_size, time_length, args.tot_num_char)
f = build_fork_lookup(vocab_size, args)
data = next(train_stream.get_epoch_iterator())[1]
print(data)
print(f(data))