def create_cnn(self):
		self.cnn = ConvolutionalNeuralNetwork()
		self.cnn.initialize_mnist()	
		if self.task_id != 1:	
			self.cnn.load_all_weights(self.task_id-1)	
		self.cnn.create_model_functions()
		return self.cnn
	def create_cnn(self):
		self.cnn = ConvolutionalNeuralNetwork()
		# self.cnn.trX, self.cnn.teX, self.cnn.trY, self.cnn.teY = mnist(onehot = True)	
		self.cnn.initialize_mnist()	
		if self.task_id != 0:	
			self.cnn.load_all_weights(self.task_id-1)	
		self.cnn.create_model_functions()
		return self.cnn
	def __init__(self):
		self.data_list=[]
		self.net_list = []
		self.task_id=1
		self.epochs=20
		self.batch_size=10
		self.data_set={}
		self.cnn = ConvolutionalNeuralNetwork()
		self.cnn.initialize_mnist()	
		self.trX_spf, self.trY_spf = self.split_data (self.cnn.trX, self.cnn.trY, 0)
		self.data_set[0] = [self.trX_spf, self.trY_spf]
class multi_net(object):

	def __init__(self):
		self.data_list=[]
		self.net_list = []
		self.task_id=1
		self.epochs=20
		self.batch_size=10
		self.data_set={}
		self.cnn = ConvolutionalNeuralNetwork()
		self.cnn.initialize_mnist()	
		self.trX_spf, self.trY_spf = self.split_data (self.cnn.trX, self.cnn.trY, 0)
		self.data_set[0] = [self.trX_spf, self.trY_spf]


	def create_cnn(self):
		self.cnn = ConvolutionalNeuralNetwork()
		self.cnn.initialize_mnist()	
		if self.task_id != 1:	
			self.cnn.load_all_weights(self.task_id-1)	
		self.cnn.create_model_functions()
		return self.cnn

	def shuffle_data(self, x, y):
		c=zip(x,y)
		random.shuffle(c)
		x, y=zip(*c)
		return x, y 	

	def get_sub_test_data(self):
		teX = self.cnn.teX
		teY = self.cnn.teY
		X = []
		Y = []
		for i in range(len(teX)):
			index = np.argmax(teY[i])
			if index <= self.task_id:
				X.append(teX[i,:,:,:])
				Y.append(teY[i])
		return X, Y


	def process_data(self):
		self.trX_spf, self.trY_spf = self.split_data (self.cnn.trX, self.cnn.trY, self.task_id)
		self.data_set[self.task_id] = [self.trX_spf, self.trY_spf]
		trX, trY, teX, teY = self. check_pre()
		trX, trY = self.shuffle_data(trX, trY)
		teX, teY = self.shuffle_data(teX, teY)
		return trX, trY, teX, teY

	def split_data(self, dataset, label, task_id):
		matching = np.nonzero(np.argmax(label, axis = 1) == task_id)[0]
		return dataset[matching], label[matching] 

	def run(self):
		for self.task_id in range(1,10):
			# if self.task_id == 0:
			# 	self.cnn = self.create_cnn()
			# 	trX, trY, teX, teY=self.process_data()
			# 	self.train(self.cnn, trX, trY, teX, teY)
			# 	self.cnn.save_all_weights(self.task_id)
			if len(self.net_list)-1 != self.task_id:
				self.cnn=self.create_cnn()
				trX, trY, teX, teY=self.process_data()
				self.train(self.cnn, trX, trY, teX, teY)
				self.cnn.save_all_weights(self.task_id)
			if len(self.net_list)-1 == self.task_id:
				pass
			self.net_list.append(self.cnn)
		# self.test()


	def check_pre(self):
		trX, trY, = self.data_set[self.task_id]
		# task_number=[self.task_id]
		if self.task_id >= 1:
			trX_1, trY_1 = self.data_set[self.task_id-1]
			trX = self.crossing_share_X(trX, trX_1, .8, trX)
			trY = self.crossing_share_Y(trY, trY_1, .8, trY)
			# task_number.append(self.task_id-1)

		if self.task_id >= 3:
			trX_2, trY_2 = self.data_set[self.task_id-3]
			trX = self.crossing_share_X(trX, trX_2, .6, trX)
			trY = self.crossing_share_Y(trY, trY_2, .6, trY)
			# task_number.append(self.task_id-2)

		if self.task_id >= 5:
			trX_3, trY_3 = self.data_set[self.task_id-5]
			trX = self.crossing_share_X(trX, trX_3, .4, trX)
			trY = self.crossing_share_Y(trY, trY_3, .4, trY)

		if self.task_id >= 7:
			trX_4, trY_4 = self.data_set[self.task_id-7]
			trX = self.crossing_share_X(trX, trX_4, .4, trX)
			trY = self.crossing_share_Y(trY, trY_4, .4, trY)

		if self.task_id >= 9:
			trX_4, trY_4 = self.data_set[self.task_id-9]
			trX = self.crossing_share_X(trX, trX_4, .2, trX)
			trY = self.crossing_share_Y(trY, trY_4, .2, trY)
			# task_number.append(self.task_id-3)
		# trX_neg, trY_neg = self.get_neg_train_data(task_number)
		# trX = self.crossing_share_X(trX, trX_neg, .99, trX_neg)
		# trY = self.crossing_share_Y(trY, trY_neg, .99, trY_neg)
		teX, teY = self.get_sub_test_data()
		# self.data_set[self.task_id-1]=[trX, trY, teX, teY]
		return trX, trY, teX, teY

	def crossing_share_X(self, a, b, rate, c):
		index = rate*len(c)
		b = b[:index,:,:,:]
		a = np.concatenate((a, b), axis=0)
		return a

	def crossing_share_Y(self, a, b, rate, c):
		index = rate*len(c)
		b = b[:index,:]
		a = np.concatenate((a, b), axis=0)
		return a

	def train(self, cnn, trX, trY, teX, teY):
		for i in range(self.epochs):
			for start, end in zip(range(0, len(trX), self.batch_size), range(self.batch_size, len(trX), self.batch_size)):
				cnn.cost = cnn.train(trX[start:end], trY[start:end])
			accuracies = []
		for start, end in zip(range(0, len(teX), self.batch_size), range(self.batch_size, len(teX), self.batch_size)):
			accuracy, num_net, betw_unit = self.calculate_accuracy(teX[start:end], teY[start:end])
			accuracies.append(accuracy)
		print("num_net: {0} step: {1} accuracy: {2}".format(num_net, betw_unit, np.mean(accuracies)))
			
	def calculate_accuracy(self, teX, teY):
		result = []

		for i in range(2):
			predict = self.cnn.predict(teX)


			if self.task_id <= 5:
				self.betw_unit = 1
				start = 2
			elif self.task_id > 5 and self.task_id < 10:
				self.betw_unit = 2	
				start = 3

			if self.task_id == 2:
				cnn = self.net_list[0]
				predict += cnn.predict(teX)	
			else:
				for i in range(start, self.task_id-1, self.betw_unit):
					cnn = self.net_list[self.task_id-i]
					predict += cnn.predict(teX)


		predict = np.argmax(predict, axis = 1)
		accuracy = np.mean(np.argmax(teY, axis = 1) == predict)
		return accuracy, self.task_id, self.betw_unit

	def test(self):
		pass
class multi_net(object):

	def __init__(self):
		self.data_list=[]
		self.net_list = []
		self.task_id=1
		self.epochs=10
		self.batch_size=40
		self.data_set={}
	# def binarize(self, classifications, task_id):
	# 	return np.asarray(np.asarray(classifications) == task_id, dtype = np.uint8)



	def create_cnn(self):
		self.cnn = ConvolutionalNeuralNetwork()
		# self.cnn.trX, self.cnn.teX, self.cnn.trY, self.cnn.teY = mnist(onehot = True)	
		self.cnn.initialize_mnist()	
		if self.task_id != 0:	
			self.cnn.load_all_weights(self.task_id-1)	
		self.cnn.create_model_functions()
		return self.cnn

	def shuffle_data(self, x, y):
		c=zip(x,y)
		random.shuffle(c)
		x, y=zip(*c)
		return x, y 	

	def get_sub_test_data(self):
		teX = self.cnn.teX
		teY = self.cnn.teY
		teX = teX[:1000,:,:,:]
		teY = teY[:1000,:]
		X = []
		Y = []
		for i in range(1000):
			index = np.nonzero(teY[i])[0][0]
			if index <= self.task_id:
				X.append(teX[i,:,:,:])
				Y.append(teY[i])
		return X, Y

	# def get_neg_train_data(self, task_number):	
	# 	trX = self.cnn.trX
	# 	trY = self.cnn.trY
	# 	trX = trX[:1000,:,:,:]
	# 	trY = trY[:1000, :]
	# 	for i in range(1000):
	# 		index = np.nonzero(trY[i])[0][0]
	# 		if  index not in list(task_number):
	# 			trY[i][index] = 0
	# 	return trX, trY

	def process_data(self):
		self.trX_spf, self.trY_spf = self.split_data (self.cnn.trX, self.cnn.trY, self.task_id)
		self.data_set[self.task_id] = [self.trX_spf, self.trY_spf]
		trX, trY, teX, teY = self. check_pre()
		trX, trY = self.shuffle_data(trX, trY)
		teX, teY = self.shuffle_data(teX, teY)
		return trX, trY, teX, teY

	def split_data(self, dataset, label, task_id):
		matching = np.nonzero(np.argmax(label, axis = 1) == task_id)[0]
		# nonmatching = np.nonzero(label != task_id)[0]
		return dataset[matching], label[matching] 

	def run(self):
		for self.task_id in range(0,10):
			if self.task_id == 0:
				self.cnn = self.create_cnn()
				trX, trY, teX, teY=self.process_data()
				self.train(self.cnn, trX, trY, teX, teY)
				self.cnn.save_all_weights(self.task_id)
			if len(self.net_list)-1 != self.task_id:
				self.cnn=self.create_cnn()
				trX, trY, teX, teY=self.process_data()
				self.train(self.cnn, trX, trY, teX, teY)
				self.cnn.save_all_weights(self.task_id)
			if len(self.net_list)-1 == self.task_id:
				pass
			self.net_list.append(self.cnn)
		# self.test()

	def check_pre(self):
		trX, trY, = self.data_set[self.task_id]
		# task_number=[self.task_id]
		if self.task_id-1 >= 0:
			trX_1, trY_1 = self.data_set[self.task_id-1]
			trX = self.crossing_share_X(trX, trX_1, .5, trX)
			trY = self.crossing_share_Y(trY, trY_1, .5, trY)
			# task_number.append(self.task_id-1)

		if self.task_id-2 >= 0:
			trX_2, trY_2 = self.data_set[self.task_id-2]
			trX = self.crossing_share_X(trX, trX_2, .3, trX)
			trY = self.crossing_share_Y(trY, trY_2, .3, trY)
			# task_number.append(self.task_id-2)

		if self.task_id-3 >= 0:
			trX_3, trY_3 = self.data_set[self.task_id-3]
			trX = self.crossing_share_X(trX, trX_3, .1, trX)
			trY = self.crossing_share_Y(trY, trY_3, .1, trY)
			# task_number.append(self.task_id-3)
		# trX_neg, trY_neg = self.get_neg_train_data(task_number)
		# trX = self.crossing_share_X(trX, trX_neg, .99, trX_neg)
		# trY = self.crossing_share_Y(trY, trY_neg, .99, trY_neg)
		teX, teY = self.get_sub_test_data()
		# self.data_set[self.task_id-1]=[trX, trY, teX, teY]
		return trX, trY, teX, teY

	def crossing_share_X(self, a, b, rate, c):
		index = rate*len(c)
		b = b[:index,:,:,:]
		a = np.concatenate((a, b), axis=0)
		return a

	def crossing_share_Y(self, a, b, rate, c):
		index = rate*len(c)
		b = b[:index,:]
		a = np.concatenate((a, b), axis=0)
		return a

	def train(self, cnn, trX, trY, teX, teY):
		for i in range(self.epochs):
			for start, end in zip(range(0, len(trX), self.batch_size), range(self.batch_size, len(trX), self.batch_size)):
				cnn.cost = cnn.train(trX[start:end], trY[start:end])
			if self.task_id-6 >= 0:
				self.calculate_accuracy("cnn_6", teX, teY)
			elif self.task_id-3 >= 0:
				self.calculate_accuracy("cnn_3", teX, teY)
			else:
				self.calculate_accuracy("cnn", teX, teY)
	
				
	def calculate_accuracy(self, flag, teX, teY):
		accuracies = []
		T.argmax(self.cnn.predict(teX))
		predict = self.cnn.predict(teX)
		if flag == "cnn_3":
			cnn_3 = self.net_list[self.task_id-3]
			if cnn_3.predict(teX) > predict:
				predict = cnn_3.predict(teX)
		elif flag == "cnn_6":
			cnn_6 = self.net_list[self.task_id-6]
			if cnn_6.predict > predict:
				predict = cnn_6.predict(teX)

		# predict = T.argmax(predict, axis = 1)
		# accuracy = np.mean(np.argmax(teY, axis = 1) == predict)
		# accuracies.append(accuracy)
		print("{0} for task_id {1} accuracy is {2}".format(flag, self.task_id, predict[0]))

	def test(self):
		pass
class multi_net(object):

	def __init__(self):
		self.data_list=[]
		self.net_list = []
		self.task_id=1
		self.epochs=1
		self.batch_size=10
		self.data_set={}
		self.cnn = ConvolutionalNeuralNetwork()
		self.cnn.initialize_mnist()	
		self.trX_spf, self.trY_spf = self.split_data (self.cnn.trX, self.cnn.trY, 0)
		self.data_set[0] = [self.trX_spf, self.trY_spf]
		self.signal = True


	def create_cnn(self):
		self.cnn = ConvolutionalNeuralNetwork()
		self.cnn.initialize_mnist()	
		if self.task_id != 1:	
			self.cnn.load_all_weights(self.task_id-1)	
		self.cnn.create_model_functions()
		return self.cnn

	def shuffle_data(self, x, y):
		c=zip(x,y)
		random.shuffle(c)
		x, y=zip(*c)
		return x, y 	

	def get_sub_test_data(self):
		teX = self.cnn.teX
		teY = self.cnn.teY
		X = []
		Y = []
		for i in range(len(teX)):
			index = np.argmax(teY[i])
			if self.signal:
				if index <= self.task_id:
					X.append(teX[i,:,:,:])
					Y.append(teY[i])
			else:
				if index > self.task_id:
					X.append(teX[i,:,:,:])
					Y.append(teY[i])
		return X, Y


	def process_data(self):
		self.trX_spf, self.trY_spf = self.split_data (self.cnn.trX, self.cnn.trY, self.task_id)
		self.data_set[self.task_id] = [self.trX_spf, self.trY_spf]
		trX, trY, teX, teY = self. check_pre()
		trX, trY = self.shuffle_data(trX, trY)
		teX, teY = self.shuffle_data(teX, teY)
		return trX, trY, teX, teY

	def split_data(self, dataset, label, task_id):
		matching = np.nonzero(np.argmax(label, axis = 1) == task_id)[0]
		return dataset[matching], label[matching] 

	def run(self):
		for self.task_id in range(1,3):
			# if self.task_id == 0:
			# 	self.cnn = self.create_cnn()
			# 	trX, trY, teX, teY=self.process_data()
			# 	self.train(self.cnn, trX, trY, teX, teY)
			# 	self.cnn.save_all_weights(self.task_id)
			if len(self.net_list)-1 != self.task_id:
				self.cnn=self.create_cnn()
				trX, trY, teX, teY=self.process_data()
				self.train(self.cnn, trX, trY, teX, teY)
				self.cnn.save_all_weights(self.task_id)
			if len(self.net_list)-1 == self.task_id:
				pass
			self.net_list.append(self.cnn)
		# self.test()

	def check_pre(self):
		trX, trY, = self.data_set[self.task_id]
		# task_number=[self.task_id]
		if self.task_id >= 1:
			trX_1, trY_1 = self.data_set[self.task_id-1]
			trX = self.crossing_share_X(trX, trX_1, .8, trX)
			trY = self.crossing_share_Y(trY, trY_1, .8, trY)
			# task_number.append(self.task_id-1)

		if self.task_id >= 3:
			trX_2, trY_2 = self.data_set[self.task_id-3]
			trX = self.crossing_share_X(trX, trX_2, .6, trX)
			trY = self.crossing_share_Y(trY, trY_2, .6, trY)
			# task_number.append(self.task_id-2)

		if self.task_id >= 5:
			trX_3, trY_3 = self.data_set[self.task_id-5]
			trX = self.crossing_share_X(trX, trX_3, .4, trX)
			trY = self.crossing_share_Y(trY, trY_3, .4, trY)

		if self.task_id >= 7:
			trX_4, trY_4 = self.data_set[self.task_id-7]
			trX = self.crossing_share_X(trX, trX_4, .4, trX)
			trY = self.crossing_share_Y(trY, trY_4, .4, trY)

		if self.task_id >= 9:
			trX_4, trY_4 = self.data_set[self.task_id-9]
			trX = self.crossing_share_X(trX, trX_4, .2, trX)
			trY = self.crossing_share_Y(trY, trY_4, .2, trY)

		trX = np.asarray(trX)
		trY = np.asarray(trY)
		teX, teY = self.get_sub_test_data()
		teX = np.asarray(teX)
		teY = np.asarray(teY)
		print teX.shape
		print teY.shape
		print trX.shape
		print trY.shape
		return trX, trY, teX, teY

	def crossing_share_X(self, a, b, rate, c):
		index = rate*len(c)
		b = b[:index,:,:,:]
		a = np.concatenate((a, b), axis=0)
		return a

	def crossing_share_Y(self, a, b, rate, c):
		index = rate*len(c)
		b = b[:index,:]
		a = np.concatenate((a, b), axis=0)
		return a

	def train(self, cnn, trX, trY, teX, teY):
		signals=[]
		for i in range(self.epochs):
			for start, end in zip(range(0, len(trX), self.batch_size), range(self.batch_size, len(trX), self.batch_size)):
				cnn.cost = cnn.train(trX[start:end], trY[start:end])


			for start, end in zip(range(0, len(teX), self.batch_size), range(self.batch_size, len(teX), self.batch_size)):
				signal = cnn.predict(teX[start:end], teY[start:end])
				signals.extend(signal)
	
		if self.signal:

			print "{0}_in_range".format(self.task_id)
			np.save("{0}_in_range.npy".format(self.task_id),signals)
		else:
			print "{0}_out_range".format(self.task_id)
			np.save("{0}_out_range.npy".format(self.task_id),signals)