Exemple #1
0
def get_sim_barrier():

    if os.path.exists('exp_results/sim_barrier.csv'):
        os.remove('exp_results/sim_barrier.csv')

    x0 = dict()
    with open('exp_results/resfd10' + '/info.csv', 'r') as f:
        data = f.readlines()
        for line in data:
            if 'x' in line:
                ldata = line.split(',')
                role = ldata[0][1:]
                x0[role] = np.array([float(ldata[1]), float(ldata[2])])
            if 'vd' in line:
                vd = float(line.split(',')[-1])
            if 'vi' in line:
                vi = float(line.split(',')[-1])

    if vd < vi:
        game = SlowDgame(LineTarget(), exp_dir='resfd10/')
        lb0, ub0 = .1, .8
    else:
        game = FastDgame(LineTarget(), exp_dir='resfd10/')
        lb0, ub0 = -.1, .4

    xbs, ybs = [], []
    with open('exp_results/sim_barrier.csv', 'a') as f:
        for xI in np.linspace(-.6, 0, 30):
            # for xI in [-.2]:
            lb, ub = lb0, ub0
            print(xI)
            while abs(ub - lb) > 0.005:
                # print(ub, lb)
                yI = .5 * (lb + ub)
                x0['I0'] = np.array([xI, yI])
                # print(x0)
                game.reset(x0)
                _, xs, info = game.advance(20.)
                print(yI, info)
                # game.plotter.reset()
                # game.plotter.plot({'play':xs}, 'play', game.pstrategy, fname=None)
                if info == 'captured':
                    ub = yI
                elif info == 'entered':
                    lb = yI
            xbs.append(xI)
            ybs.append(.5 * (lb + ub))
            f.write(','.join(map(str, [xI, .5 * (lb + ub)])) + '\n')

    return xbs, ybs
Exemple #2
0
def generate_data_for_exp(S,
                          T,
                          gmm,
                          D,
                          delta,
                          offy=0.2,
                          ni=1,
                          nd=2,
                          param_file='traj_param_test.csv'):
    game = SlowDgame(LineTarget())
    xs_ref = game.generate_analytic_traj(S,
                                         T,
                                         gmm,
                                         D,
                                         delta,
                                         offy=offy,
                                         file=param_file)
    game.reset({
        'D0': xs_ref['D0'][0, :],
        'I0': xs_ref['I0'][0, :],
        'D1': xs_ref['D1'][0, :]
    })
    ts, xs_play, _ = game.advance(10.)

    # for role in xs_ref:
    # 	xs_ref[role] = game.rotate_to_exp(xs_ref[role])
    xplot = {'play': xs_play, 'ref': xs_ref}

    fname = '_'.join([strategy for role, strategy in game.pstrategy.items()])
    figid = param_file.split('.')[0].split('_')[-1]
    game.plotter.plot(xs=xplot, geox='play', ps=game.pstrategy, traj=True)
Exemple #3
0
def sim_barrier(r, vd):

	x0 = {'D0': np.array([-.85, .2]), 'I0': np.array([-.2, 1.2]), 'D1': np.array([.85, .2])}
	if 1. < r:
		game = SlowDgame(LineTarget())
		lb0, ub0 = .0, .5
	else:
		game = FastDgame(LineTarget())
		lb0, ub0 = -.1, .4

	vi = r*vd
	game.reset(x0)
	game.set_vi(r*vd)
	game.set_vd(vd)

	xbs, ybs = [], []

	if os.path.exists('sim_revision1/sim_barrier_%.2f.csv'%((vd/vi)*100)):	
		os.remove('sim_revision1/sim_barrier_%.2f.csv'%((vd/vi)*100))

	with open('sim_revision1/sim_barrier_%.2f.csv'%((vd/vi)*100), 'a') as f:			
		for xI in np.linspace(-.6, 0, 13):
		# for xI in [-.6]:
			lb, ub = lb0, ub0
			print(xI)
			while abs(ub - lb) > 0.0005:
				# print(ub, lb)
				yI = .5*(lb + ub)
				x0['I0'] = np.array([xI, yI])
				# print(x0)
				game.reset(x0)
				# print('!!!!!!!!!! reseted x0 !!!!!!!!!!!!!')
				_, xs, info = game.advance(5000.)
				print(yI, info)
				# game.plotter.reset()
				# game.plotter.plot({'play':xs}, 'play', game.pstrategy, fname=None)
				if info == 'captured':
					ub = yI
				elif info =='entered':
					lb = yI
			xbs.append(xI)
			ybs.append(.5*(lb + ub))
			f.write(','.join(map(str, [xI, .5*(lb + ub)]))+'\n')
	
	# game.plotter.plot_sim_barrier_line()

	return xbs, ybs
Exemple #4
0
def sim_traj(rs, vd):

	x0 = {'D0': np.array([-.85, .2]), 'I0': np.array([-0.5, .75]), 'D1': np.array([.85, .2])}
	trajs = []
	# rs = []

	for r in rs:
		if 1. < r:
			game = SlowDgame(LineTarget())
		else:
			game = FastDgame(LineTarget())

		game.reset(x0)
		game.set_vi(r*vd)
		game.set_vd(vd)

		_, traj, _ = game.advance(8.)
		trajs.append(traj)
		# rs.append(vd/vi)

	game.plotter.plot_traj_compare(trajs, rs)
Exemple #5
0
def play_fastD_game(xd0, xi, xd1, ni=1, nd=2, param_file='traj_param_100.csv'):
    game = FastDgame(LineTarget())
    x0 = {'D0': xd0, 'I0': xi, 'D1': xd1}
    game.reset(x0)
    game.record_data(x0, file=param_file)

    ts_play, xs_play, _ = game.advance(8.)
    fname = '_'.join([strategy for role, strategy in game.pstrategy.items()])
    figid = param_file.split('.')[0].split('_')[-1]
    # game.plotter.plot(xs={'play': xs_play}, geox='play', ps=game.pstrategy, traj=True, fname='traj_'+fname+'_'+figid+'.png')
    game.plotter.plot(xs={'play': xs_play},
                      geox='play',
                      ps=game.pstrategy,
                      traj=True)
Exemple #6
0
def replay_exp(res_dir='res1/', ni=1, nd=2):
    x0s = dict()
    pdict = dict()
    with open('exp_results/' + res_dir + '/info.csv', 'r') as f:
        data = f.readlines()
        for line in data:
            if 'vd' in line:
                vd = float(line.split(',')[-1])
            if 'vi' in line:
                vi = float(line.split(',')[-1])
            if 'x' in line:
                ldata = line.split(',')
                role = ldata[0][1:]
                x0s[role] = np.array([float(ldata[1]), float(ldata[2])])

    if vd < vi:
        game = SlowDgame(LineTarget(), exp_dir=res_dir, ni=ni, nd=nd)
    else:
        game = FastDgame(LineTarget(), exp_dir=res_dir, ni=ni, nd=nd)

    game.reset({role: x for role, x in x0s.items()})
    # ts_ref, xs_ref, _ = game.advance(8.)
    ts_exp, xs_exp, ps_exp = game.replay_exp()
    for role, x in xs_exp.items():
        x0s[role] = x[0]
    game.reset({role: x for role, x in x0s.items()})
    ts_ref, xs_ref, _ = game.advance(8.)

    game.plotter.animate(ts_exp, xs_exp, game.pstrategy, xrs=xs_ref)
    game.plotter.plot({
        'ref': xs_ref,
        'exp': xs_exp
    },
                      'exp',
                      game.pstrategy,
                      dr=False,
                      fname='replay_traj.png')
Exemple #7
0
    },
                      'exp',
                      game.pstrategy,
                      dr=False,
                      fname='replay_traj.png')
    # game.plotter.plot({'ref':xs_ref, 'exp':xs_exp}, 'exp', pdict, dr=False, fname='replay_traj.png')


if __name__ == '__main__':

    # t0 = time.clock()
    # generate_data_for_exp(-.89, 2.8, acos(25/27)+0.5, 0.1, 0.3999999, param_file='traj_param_append01.csv')
    # play_fastD_game(np.array([-.85, 0.2]), np.array([-0.2, 0.4]), np.array([.85, 0.2]), param_file='traj_param_tst0.csv')
    # play_a_game(np.array([-.85, 0.2]), np.array([-0.2, 0.6]), np.array([.85, 0.2]), param_file='traj_param_tst0.csv')

    # replay_exp(res_dir='resfd24002/')
    # replay_exp(res_dir='resfd25001/')
    # replay_exp(res_dir='ressd26002/')
    # replay_exp(res_dir='ressd032/')

    # t1 = time.clock()
    # print(t1 - t0)
    # game = FastDgame(LineTarget(), exp_dir='ressd26002/')
    game = FastDgame(LineTarget(), exp_dir='ressd032/')
    # game = SlowDgame(LineTarget(), exp_dir='resfd24002/')
    # game.players['D0'].x = np.array([-0.85, 0.2])
    # game.players['D1'].x = np.array([ 0.85, 0.2])
    game.plotter.plot_barrier(dr='fd')

    # game.plotter.plot_velocity()
Exemple #8
0
from Games import FastDgame, SlowDgame
from geometries import LineTarget

with open('config.csv', 'r') as f:
    data = f.readlines()
    for line in data:
        if 'vd' in line:
            vd = float(line.split(',')[-1])
        if 'vi' in line:
            vi = float(line.split(',')[-1])
sim_dir = 'res5/'
xplot = dict()

if vd >= vi:
    game = FastDgame(LineTarget(), sim_dir=sim_dir)
    x0 = {
        'D0': np.array([-.8, 0.2]),
        'I0': np.array([-.1, .4]),
        'D1': np.array([.8, 0.2])
    }
else:
    game = SlowDgame(LineTarget(), sim_dir=sim_dir)
    # rgame = SlowDgame(LineTarget(), sim_dir=sim_dir)
    # for role in rgame.pstrategy:
    # 	rgame.pstrategy[role] = 'nn'

    xref = game.generate_analytic_traj(.0,
                                       5,
                                       acos(1 / 1.5) + 0.2,
                                       0,
Exemple #9
0
# from tensorflow.keras.models import Sequential
# from tensorflow.keras.layers import Dense
# from tensorflow.keras.models import load_model
from keras.models import Sequential
from keras.layers import Dense
from keras.models import load_model
from tensorflow.keras.losses import MeanSquaredError

# from envelope import envelope_traj, envelope_policy, w
from Games import slowDgame
from geometries import LineTarget

BARRIER_DIR = 'BarrierFn'
POLICY_DIR = 'Policies/PolicyFn'

game = slowDgame(LineTarget())


def network(n_in, ns, act_fns):
    model = Sequential()
    for i, (n, fn) in enumerate(zip(ns, act_fns)):
        if i == 0:
            model.add(Dense(n, input_shape=(n_in, ), activation=fn))
        else:
            model.add(Dense(n, activation=fn))
    model.compile(optimizer='Adam',
                  loss='mean_squared_error',
                  metrics=['accuracy'])
    return model

Exemple #10
0
def plot_sim_barrier():

	ratios = [1/0.8, 1/.97, 1/1.2, 1/1.56]
	# ratios = [25/24]
	linestyles=[(0, ((ratio-.6)*8, (ratio-.6)*3)) for ratio in ratios[::-1]]
	linestyles[-1] = 'solid'
	alphas = [(ratio*.8)**1.7 for ratio in ratios[::-1]]

	game = SlowDgame(LineTarget())
	fig, ax = plt.subplots()
	colors = ['purple', 'magenta', 'red', 'orange']
	for i, ratio in enumerate(ratios):
	# for ratio in [1/1.2,]:
		print((ratio/1.26)**1.7)
		xy = []
		with open('sim_revision1/sim_barrier_%.2f.csv'%(ratio*100), 'r') as f:
			print('sim_revision1/sim_barrier_%.2f.csv'%(ratio*100))
			lines = f.readlines()[1:]
			for line in lines:
				data = line.split(',')
				xy.append(np.array([float(data[0]), float(data[1])]))
				xy.append(np.array([-float(data[0]), float(data[1])]))
				# xy.append(np.array([-float(data[1]), float(data[0])]))
				# xy.append(np.array([-float(data[1]), -float(data[0])]))
		xy = np.asarray(sorted(xy, key=lambda xys: xys[0]))
		# xy = game.rotate_to_exp_point(xy)
		# line, = self.ax.plot(px[:,0], px[:,1], color=self.colors[pid], label='a=%.1f'%ratio, alpha=(ratio/1.4)**1.7, linestyle=(0, ((ratio-.6)*8, (ratio-.6)*3)))
		ax.plot(xy[:,0], xy[:,1], color='r', label='a=%.1f'%ratio, alpha=alphas[i], linestyle=linestyles[i])
		print(xy)

	# ax.plot([-1.5, 1.5], [-.5, -.5], color='k', label='target')
	# ax.plot([-0.85], [0.2], 'bo')
	# ax.plot([0.85], [0.2], 'bo')
	ax.plot([-.5], [.75], 'rx')
	ax.plot([-.2], [.5], 'ro', )
	ax.plot([-1.5, 1.5], [-.5, -.5], color='k', label='target')
	ax.plot([-0.85], [.2], 'bo')
	ax.plot([0.85], [.2], 'bo')
	ring = Circle((0.85, 0.2), game.r, fc='b', ec='b', alpha=0.6, linewidth=2)
	ax.add_patch(ring)
	ring = Circle((-0.85, 0.2), game.r, fc='b', ec='b', alpha=0.6, linewidth=2)
	ax.add_patch(ring)

	ring = Circle((0.85, 0.2), game.r*1.2, fc='b', ec='b', alpha=0.2)
	ax.add_patch(ring)
	ring = Circle((-0.85, 0.2), game.r*1.2, fc='b', ec='b', alpha=0.2)
	ax.add_patch(ring)
	# plt.xticks([-2.0, -1.5, -1.0, -0.5, 0, 0.5], [2.0, 1.5, 1.0, 0.5, 0, -0.5])
	
	xrs, yrs = [], []
	for t in np.linspace(0, 2*3.1416, 50):
		xrs.append(.85 + game.r*cos(t))
		yrs.append(.2 + game.r*sin(t))
	ax.plot(xrs, yrs, 'b')
	xrs, yrs = [], []
	for t in np.linspace(0, 2*3.1416, 50):
		xrs.append(-.85 + game.r*cos(t))
		yrs.append(.2 + game.r*sin(t))
	ax.plot(xrs, yrs, 'b')	

	ax.grid()
	ax.tick_params(axis='both', which='major', labelsize=15)
	plt.xlabel('y(m)', fontsize=15)
	plt.ylabel('x(m)', fontsize=15)

	plt.gca().legend(prop={'size': 14}, ncol=5, handletextpad=0.1, columnspacing=0.4, loc='lower center', numpoints=1, handlelength=1.2)
	plt.subplots_adjust(bottom=0.13)
	# print('set legend')
	ax.set_aspect('equal')
	ax.set_xlim([-1.2, 1.2])
	ax.set_ylim([-.53, .92])

	plt.show()
Exemple #11
0
    def __init__(self,
                 zDs,
                 zIs,
                 rate=10,
                 param_file='',
                 Ds='',
                 Is='',
                 dstrategy=None,
                 istrategy=None,
                 logger_dir='res1'):

        self.rate = rospy.Rate(rate)
        self.states = dict()

        self.target = LineTarget()
        self.policy_dict = {
            'pt': self.pt_strategy,
            'pp': self.pp_strategy,
            'nn': self.nn_strategy,
            'w': self.w_strategy,
            'h': self.h_strategy,
        }
        self.dstrategy = dstrategy
        self.istrategy = istrategy
        self.last_act = dict()

        self.param_id = param_file.split('.')[0].split('_')[-1]
        self.vdes = dict()
        self.zs = dict()
        self.x0s = dict()
        self.xes = dict()
        self.goal_msgs = dict()
        # self.policy_fns = dict()
        self.fk = 0.0
        self.nv = 20

        self.cap_time = .1
        self.last_cap = False
        self.end = False
        self.time_inrange = 0.
        self.time_end = 0.

        self.vs = dict()
        self.xs = dict()
        self.vnorms = dict()

        self.mocap_sub_callbacks = {
            'D0': self.getD0,
            'D1': self.getD1,
            'I0': self.getI0
        }
        self.mocap_subs = dict()

        self.players = dict()
        self.goal_pubs = dict()
        self.cmdV_pubs = dict()
        self.policy_pubs = dict()
        self.a_pub = rospy.Publisher('/a', Float32, queue_size=1)

        self.takeoff_clients = dict()
        self.auto_clients = dict()
        self.land_clients = dict()
        self.play_clients = dict()
        # self.NNpolicies = dict()

        ofxI = .0 + .0  # .195
        ofxD = .0 + .0
        script_dir = os.path.dirname(__file__)
        with open(os.path.join(script_dir + '/params/', param_file), 'r') as f:
            lines = f.readlines()
            for line in lines:
                data = line.split(',')
                if 'vd' == data[0]:
                    vd = float(data[-1])
                if 'vi' == data[0]:
                    vi = float(data[-1])
                if 'rc' == data[0]:
                    self.r = float(data[-1])
                if 'r_close' == data[0]:
                    self.r_close = float(data[-1]) * self.r
                    print(self.r_close)
                if 'k_close' == data[0]:
                    self.k_close = float(data[-1])
                if 'S' == data[0]:
                    self.S = float(data[-1])
                if 'T' == data[0]:
                    self.T = float(data[-1])
                if 'gmm' == data[0]:
                    self.gmm = float(data[-1])
                if 'D' == data[0]:
                    self.D = float(data[-1])
                if 'delta' == data[0]:
                    self.delta = float(data[-1])
                if 'offy' == data[0]:
                    self.offy = float(data[-1])
                if 'x' in data[0]:
                    role = data[0][1:]
                    if 'I' in role:
                        # +.35 could be good
                        self.x0s[role] = np.array([
                            float(data[1]), float(data[2])
                        ]) + np.array([0, ofxI])
                        self.xes[role] = np.array([
                            float(data[1]), float(data[2])
                        ]) + np.array([0, ofxI])
                        self.xs[role] = np.array([
                            float(data[1]), float(data[2])
                        ]) + np.array([0, ofxI])
                    else:
                        self.x0s[role] = np.array([
                            float(data[1]), float(data[2])
                        ]) + np.array([0, ofxD])
                        self.xes[role] = np.array([
                            float(data[1]), float(data[2])
                        ]) + np.array([0, ofxD])
                        self.xs[role] = np.array([
                            float(data[1]), float(data[2])
                        ]) + np.array([0, ofxD])

        self.a = vd / vi
        # print(self.a)
        if self.a >= 1:
            self.policy_dict['f'] = self.f_strategy_fastD
            self.strategy = self.strategy_fastD
        else:
            self.policy_dict['f'] = self.f_strategy_slowD
            self.strategy = self.strategy_slowD
            self.gmm0 = acos(self.a)

        for i, (D, z) in enumerate(zip(Ds, zDs)):
            if D != '':
                role = 'D' + str(i)
                self.players[role] = D
                # self.NNpolicies[role] = load_model(script_dir+'/Policies/PolicyFn_'+role+'.h5')
                self.zs[role] = float(z)
                self.goal_msgs[role] = PoseStamped()
                # print(D, role)
                self.updateGoal(role, goal=self.x0s[role], init=True)
                self.last_act['p_' + role] = ''

                self.vnorms[role] = []
                self.vs[role] = np.zeros(2)
                self.vdes[role] = vd
                # self.policy_fns[role] = load_model(os.path.join(script_dir, 'Policies/PolicyFn_'+role+'.h5'))
                # self.policy_fns[role] = load_model('PolicyFn_' + role)
                self.states[role] = None

                self.goal_pubs[role] = rospy.Publisher('/' + D + '/goal',
                                                       PoseStamped,
                                                       queue_size=1)
                self.cmdV_pubs[role] = rospy.Publisher('/' + D + '/cmdV',
                                                       Twist,
                                                       queue_size=1)
                self.policy_pubs[role] = rospy.Publisher('/' + D + '/policy',
                                                         String,
                                                         queue_size=1)
                # self.a_pubs[role]	= rospy.Publisher('/'+D+'/a', PoseStamped, queue_size=1)

                self.takeoff_clients[role] = self.service_client(
                    D, '/cftakeoff')
                self.land_clients[role] = self.service_client(D, '/cfland')
                self.play_clients[role] = self.service_client(D, '/cfplay')
                self.auto_clients[role] = self.service_client(D, '/cfauto')
                self.mocap_subs[role] = rospy.Subscriber(
                    '/' + D + '/mocap', Mocap, self.mocap_sub_callbacks[role])

        for i, (I, z) in enumerate(zip(Is, zIs)):
            if I != '':
                role = 'I' + str(i)
                self.players[role] = I
                # self.NNpolicies[role] = load_model('Policies/PolicyFn_'+role+'.h5')
                self.zs[role] = float(z)
                self.goal_msgs[role] = PoseStamped()
                self.updateGoal(role, goal=self.x0s[role], init=True)
                self.last_act['p_' + role] = ''

                self.vnorms[role] = []
                self.vs[role] = np.zeros(2)
                self.vdes[role] = vi
                # self.policy_fns[role] = load_model(os.path.join(script_dir, 'Policies/PolicyFn_'+role+'.h5'))
                # self.policy_fns[role] = load_model('PolicyFn_' + role)
                self.states[role] = None

                self.goal_pubs[role] = rospy.Publisher('/' + I + '/goal',
                                                       PoseStamped,
                                                       queue_size=1)
                self.cmdV_pubs[role] = rospy.Publisher('/' + I + '/cmdV',
                                                       Twist,
                                                       queue_size=1)
                self.policy_pubs[role] = rospy.Publisher('/' + I + '/policy',
                                                         String,
                                                         queue_size=1)

                self.takeoff_clients[role] = self.service_client(
                    I, '/cftakeoff')
                self.land_clients[role] = self.service_client(I, '/cfland')
                self.play_clients[role] = self.service_client(I, '/cfplay')
                self.auto_clients[role] = self.service_client(I, '/cfauto')
                self.mocap_subs[role] = rospy.Subscriber(
                    '/' + I + '/mocap', Mocap, self.mocap_sub_callbacks[role])

        # for role in self.players:
        # 	print(self.goal_msgs[role].pose.position.x, self.goal_msgs[role].pose.position.y)
        rospy.Service('alltakeoff', Empty, self.alltakeoff)
        rospy.Service('allplay', Empty, self.allplay)
        rospy.Service('allland', Empty, self.allland)

        self.inffname = script_dir + '/' + logger_dir + '/info.csv'
        if os.path.exists(self.inffname):
            os.remove(self.inffname)

        with open(self.inffname, 'a') as f:
            f.write('paramid,' + self.param_id + '\n')
            for role, cf in self.players.items():
                f.write(role + ',' + cf + '\n')
                f.write('x' + role + ',%.5f, %.5f\n' %
                        (self.x0s[role][0], self.x0s[role][1]))
            f.write('vd,%.2f' % vd + '\n')
            f.write('vi,%.2f' % vi + '\n')
            f.write('rc,%.2f' % self.r + '\n')
            if vd < vi:
                f.write('r_close,%.2f' % (self.r_close / self.r) + '\n')
                f.write('k_close,%.2f' % self.k_close + '\n')
                f.write('S,%.10f\n' % self.S)
                f.write('T,%.10f\n' % self.T)
                f.write('gmm,%.10f\n' % self.gmm)
                f.write('D,%.10f\n' % self.D)
                f.write('delta,%.10f\n' % self.delta)
                f.write('offy,%.10f\n' % (self.offy + ofxD))
                f.write('offy_dI,%.10f\n' % (ofxI - ofxD))
                f.write('fk,%.10f\n' % self.fk)
            f.write('dstrategy,%s\n' % self.dstrategy)
            f.write('istrategy,%s\n' % self.istrategy)
Exemple #12
0
class RAgame(object):
    def __init__(self,
                 zDs,
                 zIs,
                 rate=10,
                 param_file='',
                 Ds='',
                 Is='',
                 dstrategy=None,
                 istrategy=None,
                 logger_dir='res1'):

        self.rate = rospy.Rate(rate)
        self.states = dict()

        self.target = LineTarget()
        self.policy_dict = {
            'pt': self.pt_strategy,
            'pp': self.pp_strategy,
            'nn': self.nn_strategy,
            'w': self.w_strategy,
            'h': self.h_strategy,
        }
        self.dstrategy = dstrategy
        self.istrategy = istrategy
        self.last_act = dict()

        self.param_id = param_file.split('.')[0].split('_')[-1]
        self.vdes = dict()
        self.zs = dict()
        self.x0s = dict()
        self.xes = dict()
        self.goal_msgs = dict()
        # self.policy_fns = dict()
        self.fk = 0.0
        self.nv = 20

        self.cap_time = .1
        self.last_cap = False
        self.end = False
        self.time_inrange = 0.
        self.time_end = 0.

        self.vs = dict()
        self.xs = dict()
        self.vnorms = dict()

        self.mocap_sub_callbacks = {
            'D0': self.getD0,
            'D1': self.getD1,
            'I0': self.getI0
        }
        self.mocap_subs = dict()

        self.players = dict()
        self.goal_pubs = dict()
        self.cmdV_pubs = dict()
        self.policy_pubs = dict()
        self.a_pub = rospy.Publisher('/a', Float32, queue_size=1)

        self.takeoff_clients = dict()
        self.auto_clients = dict()
        self.land_clients = dict()
        self.play_clients = dict()
        # self.NNpolicies = dict()

        ofxI = .0 + .0  # .195
        ofxD = .0 + .0
        script_dir = os.path.dirname(__file__)
        with open(os.path.join(script_dir + '/params/', param_file), 'r') as f:
            lines = f.readlines()
            for line in lines:
                data = line.split(',')
                if 'vd' == data[0]:
                    vd = float(data[-1])
                if 'vi' == data[0]:
                    vi = float(data[-1])
                if 'rc' == data[0]:
                    self.r = float(data[-1])
                if 'r_close' == data[0]:
                    self.r_close = float(data[-1]) * self.r
                    print(self.r_close)
                if 'k_close' == data[0]:
                    self.k_close = float(data[-1])
                if 'S' == data[0]:
                    self.S = float(data[-1])
                if 'T' == data[0]:
                    self.T = float(data[-1])
                if 'gmm' == data[0]:
                    self.gmm = float(data[-1])
                if 'D' == data[0]:
                    self.D = float(data[-1])
                if 'delta' == data[0]:
                    self.delta = float(data[-1])
                if 'offy' == data[0]:
                    self.offy = float(data[-1])
                if 'x' in data[0]:
                    role = data[0][1:]
                    if 'I' in role:
                        # +.35 could be good
                        self.x0s[role] = np.array([
                            float(data[1]), float(data[2])
                        ]) + np.array([0, ofxI])
                        self.xes[role] = np.array([
                            float(data[1]), float(data[2])
                        ]) + np.array([0, ofxI])
                        self.xs[role] = np.array([
                            float(data[1]), float(data[2])
                        ]) + np.array([0, ofxI])
                    else:
                        self.x0s[role] = np.array([
                            float(data[1]), float(data[2])
                        ]) + np.array([0, ofxD])
                        self.xes[role] = np.array([
                            float(data[1]), float(data[2])
                        ]) + np.array([0, ofxD])
                        self.xs[role] = np.array([
                            float(data[1]), float(data[2])
                        ]) + np.array([0, ofxD])

        self.a = vd / vi
        # print(self.a)
        if self.a >= 1:
            self.policy_dict['f'] = self.f_strategy_fastD
            self.strategy = self.strategy_fastD
        else:
            self.policy_dict['f'] = self.f_strategy_slowD
            self.strategy = self.strategy_slowD
            self.gmm0 = acos(self.a)

        for i, (D, z) in enumerate(zip(Ds, zDs)):
            if D != '':
                role = 'D' + str(i)
                self.players[role] = D
                # self.NNpolicies[role] = load_model(script_dir+'/Policies/PolicyFn_'+role+'.h5')
                self.zs[role] = float(z)
                self.goal_msgs[role] = PoseStamped()
                # print(D, role)
                self.updateGoal(role, goal=self.x0s[role], init=True)
                self.last_act['p_' + role] = ''

                self.vnorms[role] = []
                self.vs[role] = np.zeros(2)
                self.vdes[role] = vd
                # self.policy_fns[role] = load_model(os.path.join(script_dir, 'Policies/PolicyFn_'+role+'.h5'))
                # self.policy_fns[role] = load_model('PolicyFn_' + role)
                self.states[role] = None

                self.goal_pubs[role] = rospy.Publisher('/' + D + '/goal',
                                                       PoseStamped,
                                                       queue_size=1)
                self.cmdV_pubs[role] = rospy.Publisher('/' + D + '/cmdV',
                                                       Twist,
                                                       queue_size=1)
                self.policy_pubs[role] = rospy.Publisher('/' + D + '/policy',
                                                         String,
                                                         queue_size=1)
                # self.a_pubs[role]	= rospy.Publisher('/'+D+'/a', PoseStamped, queue_size=1)

                self.takeoff_clients[role] = self.service_client(
                    D, '/cftakeoff')
                self.land_clients[role] = self.service_client(D, '/cfland')
                self.play_clients[role] = self.service_client(D, '/cfplay')
                self.auto_clients[role] = self.service_client(D, '/cfauto')
                self.mocap_subs[role] = rospy.Subscriber(
                    '/' + D + '/mocap', Mocap, self.mocap_sub_callbacks[role])

        for i, (I, z) in enumerate(zip(Is, zIs)):
            if I != '':
                role = 'I' + str(i)
                self.players[role] = I
                # self.NNpolicies[role] = load_model('Policies/PolicyFn_'+role+'.h5')
                self.zs[role] = float(z)
                self.goal_msgs[role] = PoseStamped()
                self.updateGoal(role, goal=self.x0s[role], init=True)
                self.last_act['p_' + role] = ''

                self.vnorms[role] = []
                self.vs[role] = np.zeros(2)
                self.vdes[role] = vi
                # self.policy_fns[role] = load_model(os.path.join(script_dir, 'Policies/PolicyFn_'+role+'.h5'))
                # self.policy_fns[role] = load_model('PolicyFn_' + role)
                self.states[role] = None

                self.goal_pubs[role] = rospy.Publisher('/' + I + '/goal',
                                                       PoseStamped,
                                                       queue_size=1)
                self.cmdV_pubs[role] = rospy.Publisher('/' + I + '/cmdV',
                                                       Twist,
                                                       queue_size=1)
                self.policy_pubs[role] = rospy.Publisher('/' + I + '/policy',
                                                         String,
                                                         queue_size=1)

                self.takeoff_clients[role] = self.service_client(
                    I, '/cftakeoff')
                self.land_clients[role] = self.service_client(I, '/cfland')
                self.play_clients[role] = self.service_client(I, '/cfplay')
                self.auto_clients[role] = self.service_client(I, '/cfauto')
                self.mocap_subs[role] = rospy.Subscriber(
                    '/' + I + '/mocap', Mocap, self.mocap_sub_callbacks[role])

        # for role in self.players:
        # 	print(self.goal_msgs[role].pose.position.x, self.goal_msgs[role].pose.position.y)
        rospy.Service('alltakeoff', Empty, self.alltakeoff)
        rospy.Service('allplay', Empty, self.allplay)
        rospy.Service('allland', Empty, self.allland)

        self.inffname = script_dir + '/' + logger_dir + '/info.csv'
        if os.path.exists(self.inffname):
            os.remove(self.inffname)

        with open(self.inffname, 'a') as f:
            f.write('paramid,' + self.param_id + '\n')
            for role, cf in self.players.items():
                f.write(role + ',' + cf + '\n')
                f.write('x' + role + ',%.5f, %.5f\n' %
                        (self.x0s[role][0], self.x0s[role][1]))
            f.write('vd,%.2f' % vd + '\n')
            f.write('vi,%.2f' % vi + '\n')
            f.write('rc,%.2f' % self.r + '\n')
            if vd < vi:
                f.write('r_close,%.2f' % (self.r_close / self.r) + '\n')
                f.write('k_close,%.2f' % self.k_close + '\n')
                f.write('S,%.10f\n' % self.S)
                f.write('T,%.10f\n' % self.T)
                f.write('gmm,%.10f\n' % self.gmm)
                f.write('D,%.10f\n' % self.D)
                f.write('delta,%.10f\n' % self.delta)
                f.write('offy,%.10f\n' % (self.offy + ofxD))
                f.write('offy_dI,%.10f\n' % (ofxI - ofxD))
                f.write('fk,%.10f\n' % self.fk)
            f.write('dstrategy,%s\n' % self.dstrategy)
            f.write('istrategy,%s\n' % self.istrategy)

    # def line_target(self, x):
    # 	return x[1]

    def service_client(self, cf, name):
        srv_name = '/' + cf + name
        rospy.wait_for_service(srv_name)
        rospy.loginfo('found' + srv_name + 'service')
        return rospy.ServiceProxy(srv_name, Empty)

    def alltakeoff(self, req):
        for role, takeoff in self.takeoff_clients.items():
            self.states[role] = 'hover'
            takeoff()
        return EmptyResponse()

    def allplay(self, req):
        for role, play in self.play_clients.items():
            self.states[role] = 'play'
            play()
        return EmptyResponse()

    def allland(self, req):
        for role, land in self.land_clients.items():
            self.states[role] = 'land'
            land()
        return EmptyResponse()

    def get_time(self):
        t = rospy.Time.now()
        return t.secs + t.nsecs * 1e-9

    def getD0(self, data):
        self.xs['D0'] = np.array([data.position[0], data.position[1]])
        self.vs['D0'] = np.array([data.velocity[0], data.velocity[1]])
        if len(self.vnorms['D0']) > self.nv:
            self.vnorms['D0'].pop(0)
        self.vnorms['D0'].append(
            sqrt(data.velocity[0]**2 + data.velocity[1]**2))

    def getD1(self, data):
        self.xs['D1'] = np.array([data.position[0], data.position[1]])
        self.vs['D1'] = np.array([data.velocity[0], data.velocity[1]])
        if len(self.vnorms['D1']) > self.nv:
            self.vnorms['D1'].pop(0)
        self.vnorms['D1'].append(
            sqrt(data.velocity[0]**2 + data.velocity[1]**2))

    def getI0(self, data):
        self.xs['I0'] = np.array([data.position[0], data.position[1]])
        self.vs['I0'] = np.array([data.velocity[0], data.velocity[1]])
        if len(self.vnorms['I0']) > self.nv:
            self.vnorms['I0'].pop(0)
        self.vnorms['I0'].append(
            sqrt(data.velocity[0]**2 + data.velocity[1]**2))

    def is_capture(self):
        d0 = np.linalg.norm(self.xs['D0'] - self.xs['I0'])
        d1 = np.linalg.norm(self.xs['D1'] - self.xs['I0'])
        return d0 < self.r or d1 < self.r

    def is_intarget(self, xi):
        print(xi, self.target.level(xi))
        if self.target.level(xi) <= 0:
            return True
        else:
            return False

    def updateGoal(self, role, goal=None, init=False):

        if init:
            self.goal_msgs[role].header.seq = 0
            self.goal_msgs[role].header.frame_id = '/world'
        else:
            self.goal_msgs[role].header.seq += 1

        self.goal_msgs[role].header.stamp = rospy.Time.now()

        if goal is not None:
            self.goal_msgs[role].pose.position.x = goal[0]
            self.goal_msgs[role].pose.position.y = goal[1]
            self.goal_msgs[role].pose.position.z = self.zs[role]
        else:
            self.goal_msgs[role].pose.position.x = self.x0s[role][0]
            self.goal_msgs[role].pose.position.y = self.x0s[role][1]
            self.goal_msgs[role].pose.position.z = self.zs[role]

        # quaternion = transform.transformations.quaternion_from_euler(0, 0, 0)
        self.goal_msgs[role].pose.orientation.x = 0
        self.goal_msgs[role].pose.orientation.y = 0
        self.goal_msgs[role].pose.orientation.z = 0
        self.goal_msgs[role].pose.orientation.w = 1

    def get_a(self):
        if len(self.vnorms['D0']) > 2 and len(self.vnorms['D1']) > 2 and len(
                self.vnorms['I0']) > 2:
            vd0 = np.array(self.vnorms['D0']).mean()
            vd1 = np.array(self.vnorms['D1']).mean()
            vi0 = np.array(self.vnorms['I0']).mean()
            # print((vd1 + vd2)/(2 * vi))
            a = min((vd0 + vd1) / (2 * vi0), 0.99)
        else:
            a = self.a
        self.a_pub.publish(a)
        return a

    def get_vecs(self):
        D1 = np.concatenate((self.xs['D0'], [0]))
        D2 = np.concatenate((self.xs['D1'], [0]))
        I = np.concatenate((self.xs['I0'], [0]))
        D1_I = I - D1
        D2_I = I - D2
        D1_D2 = D2 - D1
        return D1_I, D2_I, D1_D2

    def get_base(self, D1_I, D2_I, D1_D2):
        base_d1 = atan2(D1_I[1], D1_I[0])
        base_d2 = atan2(D2_I[1], D2_I[0])
        base_i = atan2(-D2_I[1], -D2_I[0])
        # print(base_d1*180/pi, base_d2*180/pi, base_i*180/pi)
        return {'D0': base_d1, 'D1': base_d2, 'I0': base_i}

    def get_xyz(self, D1_I, D2_I, D1_D2):
        z = np.linalg.norm(D1_D2) / 2
        x = -np.cross(D1_D2, D1_I)[-1] / (2 * z)
        y = np.dot(D1_D2, D1_I) / (2 * z) - z
        return x, y, z

    def get_theta(self, D1_I, D2_I, D1_D2):
        k1 = atan2(np.cross(D1_D2, D1_I)[-1],
                   np.dot(D1_D2, D1_I))  # angle between D1_D2 to D1_I
        k2 = atan2(np.cross(D1_D2, D2_I)[-1],
                   np.dot(D1_D2, D2_I))  # angle between D1_D2 to D2_I
        tht = k2 - k1
        if k1 < 0:
            tht += 2 * pi
        return tht

    def get_d(self, D1_I, D2_I, D1_D2):
        d1 = max(np.linalg.norm(D1_I), self.r)
        d2 = max(np.linalg.norm(D2_I), self.r)
        return d1, d2

    def get_alpha(self, D1_I, D2_I, D1_D2):
        d1, d2 = self.get_d(D1_I, D2_I, D1_D2)
        a1 = asin(self.r / d1)
        a2 = asin(self.r / d2)
        return d1, d2, a1, a2

    @closeWrapper
    def strategy_slowD(self):
        pass

    @mixWrapper
    def strategy_fastD(self):
        pass

    def dr_intersection(self):

        a = self.get_a()
        a = self.a

        D1_I, D2_I, D1_D2 = self.get_vecs()
        x, y, z = self.get_xyz(D1_I, D2_I, D1_D2)

        A = -a**2 + 1
        B = 2 * a**2 * x
        C = -a**2 * (x**2 + y**2) + z**2 + self.r**2
        a4, a3, a2, a1, a0 = A**2, 2 * A * B, B**2 + 2 * A * C - 4 * self.r**2, 2 * B * C, C**2 - 4 * self.r**2 * z**2
        b, c, d, e = a3 / a4, a2 / a4, a1 / a4, a0 / a4
        p = (8 * c - 3 * b**2) / 8
        q = (b**3 - 4 * b * c + 8 * d) / 8
        r = (-3 * b**4 + 256 * e - 64 * b * d + 16 * b**2 * c) / 256

        cubic = np.roots([8, 8 * p, 2 * p**2 - 8 * r, -q**2])
        for root in cubic:
            if root.imag == 0 and root.real > 0:
                m = root
                break
        # m = cubic[1]
        # print('m=%.5f'%m)
        # for root in cubic:
        # 	print(root)
        # print('\n')

        y1 = cm.sqrt(2 * m) / 2 - cm.sqrt(
            -(2 * p + 2 * m + cm.sqrt(2) * q / cm.sqrt(m))) / 2 - b / 4
        y2 = cm.sqrt(2 * m) / 2 + cm.sqrt(
            -(2 * p + 2 * m + cm.sqrt(2) * q / cm.sqrt(m))) / 2 - b / 4
        # y3 = -cm.sqrt(2*m)/2 - cm.sqrt(-(2*p + 2*m - cm.sqrt(2)*q/cm.sqrt(m)))/2 - b/4
        # y4 = -cm.sqrt(2*m)/2 + cm.sqrt(-(2*p + 2*m - cm.sqrt(2)*q/cm.sqrt(m)))/2 - b/4

        # return np.array([y1.real, 0])
        return np.array([y1.real, 0]), np.array([y2.real, 0])

    def projection_on_target(self, x):
        def dist(xt):
            return sqrt((x[0] - xt[0])**2 + (x[1] - xt[1])**2)

        in_target = NonlinearConstraint(self.target.level, -np.inf, 0)
        sol = minimize(dist, np.array([0, 0]), constraints=(in_target, ))
        return sol.x

    def pt_strategy(self):
        xt = self.projection_on_target(self.xs['I0'])
        P = np.concatenate((xt, [0]))
        I_P = np.concatenate((xt - self.xs['I0'], [0]))
        D0_P = np.concatenate((xt - self.xs['D0'], [0]))
        D1_P = np.concatenate((xt - self.xs['D1'], [0]))

        xaxis = np.array([1, 0, 0])
        psi = atan2(np.cross(xaxis, I_P)[-1], np.dot(xaxis, I_P))
        phi_1 = atan2(np.cross(xaxis, D0_P)[-1], np.dot(xaxis, D0_P))
        phi_2 = atan2(np.cross(xaxis, D1_P)[-1], np.dot(xaxis, D1_P))

        actions = {'D0': phi_1, 'D1': phi_2, 'I0': psi}
        actions['p_' + 'I0'] = 'pt_strategy'
        actions['p_' + 'D0'] = 'pt_strategy'
        actions['p_' + 'D1'] = 'pt_strategy'

        return actions

    def pp_strategy(self):
        xt = self.projection_on_target(self.xs['I0'])
        P = np.concatenate((xt, [0]))
        I_P = np.concatenate((xt - self.xs['I0'], [0]))
        D0_I = np.concatenate((self.xs['I0'] - self.xs['D0'], [0]))
        D1_I = np.concatenate((self.xs['I0'] - self.xs['D1'], [0]))

        xaxis = np.array([1, 0, 0])
        psi = atan2(np.cross(xaxis, I_P)[-1], np.dot(xaxis, I_P))
        phi_1 = atan2(np.cross(xaxis, D0_I)[-1], np.dot(xaxis, D0_I))
        phi_2 = atan2(np.cross(xaxis, D1_I)[-1], np.dot(xaxis, D1_I))

        actions = {'D0': phi_1, 'D1': phi_2, 'I0': psi}
        actions['p_' + 'I0'] = 'pt_strategy'
        actions['p_' + 'D0'] = 'pp_strategy'
        actions['p_' + 'D1'] = 'pp_strategy'

        return actions

    def w_strategy(self):
        D1_I, D2_I, D1_D2 = self.get_vecs()
        base = self.get_base(D1_I, D2_I, D1_D2)
        d1, d2, a1, a2 = self.get_alpha(D1_I, D2_I, D1_D2)
        tht = self.get_theta(D1_I, D2_I, D1_D2)

        phi_1 = -(pi / 2 - a1)
        phi_2 = (pi / 2 - a2)
        delta = (tht - (a1 + a2) - pi + 2 * self.gmm0) / 2
        psi_min = -(tht - (a1 + pi / 2 - self.gmm0))
        psi_max = -(a2 + pi / 2 - self.gmm0)

        I_T = np.concatenate(
            (self.projection_on_target(self.xs['I0']) - self.xs['I0'], [0]))
        angT = atan2(np.cross(-D2_I, I_T)[-1], np.dot(-D2_I, I_T))
        psi = np.clip(angT, psi_min, psi_max)
        # print(angT, psi_min, psi_max, psi)

        phi_1 += base['D0']
        phi_2 += base['D1']
        psi += base['I0']

        acts = {'D0': phi_1, 'D1': phi_2, 'I0': psi}

        for role in self.players:
            acts['p_' + role] = 'w_strategy'

        return acts

    # @Iwin_wrapper
    # # @nullWrapper
    def nn_strategy(self):
        pass

    # 	acts = dict()
    # 	x = np.concatenate((self.xs['D0'], self.xs['I0'], self.xs['D1']))
    # 	for i in [0, 2, 4]:
    # 		x[i] -= self.target.x0
    # 	for i in [1, 3, 5]:
    # 		x[i] -= self.target.y0

    # 	for role, p in self.policy_fns.items():
    # 		acts[role] = p.predict(x[None])[0]
    # 	# print('nn')
    # 	return acts

    def f_strategy_fastD(self):
        dr = DominantRegion(self.r, self.a, self.xs['I0'],
                            (self.xs['D0'], self.xs['D1']))
        # xt = self.target.deepest_point_in_dr(dr, target=self.target)
        xt = self.target.deepest_point_in_dr(dr)

        xi, xds = self.xs['I0'], (self.xs['D0'], self.xs['D1'])

        IT = np.concatenate((xt - xi, np.zeros((1, ))))
        DTs = []
        for xd in xds:
            DT = np.concatenate((xt - xd, np.zeros(1, )))
            DTs.append(DT)
        xaxis = np.array([1, 0, 0])
        psis = [atan2(np.cross(xaxis, IT)[-1], np.dot(xaxis, IT))]
        phis = []
        for DT in DTs:
            phi = atan2(np.cross(xaxis, DT)[-1], np.dot(xaxis, DT))
            phis.append(phi)

        acts = dict()
        for role, p in self.players.items():
            if 'D' in role:
                acts[role] = phis[int(role[-1])]
                acts['p_' + role] = 'f_strategy_fastD'
            elif 'I' in role:
                acts[role] = psis[int(role[-1])]
                acts['p_' + role] = 'f_strategy_fastD'
        return acts

    @Iwin_wrapper
    # @nullWrapper
    def f_strategy_slowD(self):
        D1_I, D2_I, D1_D2 = self.get_vecs()
        base = self.get_base(D1_I, D2_I, D1_D2)
        x, y, z = self.get_xyz(D1_I, D2_I, D1_D2)
        xtI, xtD = self.dr_intersection()
        # xt = self.deepest_in_target(xs)

        Pi = np.concatenate((xtI, [0]))
        Pd = np.concatenate((xtD, [0]))
        P = self.fk * Pd + (1 - self.fk) * Pi

        D1_ = np.array([0, -z, 0])
        D2_ = np.array([0, z, 0])
        I_ = np.array([x, y, 0])

        D1_P = P - D1_
        D2_P = P - D2_
        I_P = Pi - I_
        D1_I_ = I_ - D1_
        D2_I_ = I_ - D2_
        D1_D2_ = D2_ - D1_

        phi_1 = atan2(np.cross(D1_I_, D1_P)[-1], np.dot(D1_I_, D1_P))
        phi_2 = atan2(np.cross(D2_I_, D2_P)[-1], np.dot(D2_I_, D2_P))
        psi = atan2(np.cross(-D2_I_, I_P)[-1], np.dot(-D2_I_, I_P))

        phi_1 += base['D0']
        phi_2 += base['D1']
        psi += base['I0']

        return {'D0': phi_1, 'D1': phi_2, 'I0': psi}

    # def z_strategy(self):
    # 	D1_I, D2_I, D1_D2 = self.get_vecs()
    # 	base = self.get_base(Ddef f1_I, D2_I, D1_D2)
    # 	d1, d2 = self.get_d(D1_I, D2_I, D1_D2)
    # 	tht = self.get_theta(D1_I, D2_I, D1_D2)
    # 	phi_1 = -pi / 2
    # 	phi_2 = pi / 2
    # 	cpsi = d2 * sin(tht)
    # 	spsi = -(d1 - d2 * cos(tht))
    # 	psi = atan2(spsi, cpsi)
    # 	# print('%.2f, %.2f, %.2f'%(phi_1*180/pi, phi_2*180/pi, psi*180/pi))

    # 	phi_1 += base['D0']
    # 	phi_2 += base['D1']
    # 	psi += base['I0']

    # 	# print('%.2f, %.2f, %.2f'%(phi_1*180/pi, phi_2*180/pi, psi*180/pi))

    # 	return {'D0': phi_1, 'D1': phi_2, 'I0': psi, 'policy': 'z_strategy'}

    @Iwin_wrapper
    # @nullWrapper
    def h_strategy(self):
        D1_I, D2_I, D1_D2 = self.get_vecs()
        base = self.get_base(D1_I, D2_I, D1_D2)
        x, y, z = self.get_xyz(D1_I, D2_I, D1_D2)
        a = self.get_a()

        Delta = sqrt(
            np.maximum(
                x**2 - (1 - 1 / self.a**2) * (x**2 + y**2 - (z / self.a)**2),
                0))
        if (x + Delta) / (1 - 1 / self.a**2) - x > 0:
            xP = (x + Delta) / (1 - 1 / self.a**2)
        else:
            xP = -(x + Delta) / (1 - 1 / self.a**2)

        P = np.array([xP, 0, 0])
        D0_ = np.array([0, -z, 0])
        D1_ = np.array([0, z, 0])
        I0_ = np.array([x, y, 0])
        D0_P = P - D0_
        D1_P = P - D1_
        I0_P = P - I0_
        D0_I0_ = I0_ - D0_
        D1_I0_ = I0_ - D1_
        D0_D1_ = D1_ - D0_

        phi_1 = atan2(np.cross(D0_I0_, D0_P)[-1], np.dot(D0_I0_, D0_P))
        phi_2 = atan2(np.cross(D1_I0_, D1_P)[-1], np.dot(D1_I0_, D1_P))
        psi = atan2(np.cross(-D1_I0_, I0_P)[-1], np.dot(-D1_I0_, I0_P))
        # print(phi_1, phi_2, psi)
        phi_1 += base['D0']
        phi_2 += base['D1']
        psi += base['I0']

        return {'D0': phi_1, 'D1': phi_2, 'I0': psi, 'policy': 'h_strategy'}

    def hover(self, role, x):
        self.updateGoal(role, x)
        self.goal_pubs[role].publish(self.goal_msgs[role])

    def game(self, dt):

        if not self.end:
            # acts = self.strategy(self)
            acts = self.strategy()
            for role in self.players:
                vx = self.vdes[role] * cos(acts[role])
                vy = self.vdes[role] * sin(acts[role])
                self.policy_pubs[role].publish(acts['p_' + role])
                cmdV = Twist()
                cmdV.linear.x = vx
                cmdV.linear.y = vy
                self.cmdV_pubs[role].publish(cmdV)
                self.updateGoal(role, self.xs[role])
                self.goal_pubs[role].publish(self.goal_msgs[role])

            if self.is_capture():
                print(self.time_inrange)
                if self.last_cap:
                    self.time_inrange += dt
                else:
                    self.time_inrange = 0
                self.last_cap = True
                # print(self._time_inrange)
            if self.time_inrange > self.cap_time:
                self.end = True
                with open(self.inffname, 'a') as f:
                    f.write('termination,%s\n' % 'captured')
                print('!!!!!!!!!!!!!!!!!!captured, game ends!!!!!!!!!!!!')
                for role in self.players:
                    self.xes[role] = deepcopy(self.xs[role])
                for role in self.players:
                    if 'I' in role:
                        if self.states[role] != 'land':
                            self.states[role] = 'land'
                            self.land_clients[role]()
                    elif 'D' in role:
                        self.states[role] = 'hover'
                        self.auto_clients[role]()
                        # print(role + self.states[role])
            if self.is_intarget(self.xs['I0']):
                self.end = True
                with open(self.inffname, 'a') as f:
                    f.write('termination,%s' % 'entered')
                print('!!!!!!!!!!!!!!!!!!entered, game ends!!!!!!!!!!!!')
                for role in self.players:
                    self.xes[role] = deepcopy(self.xs[role])
                for role in self.players:
                    if 'D' in role:
                        if self.states[role] != 'land':
                            self.states[role] = 'land'
                            self.land_clients[role]()
                    elif 'I' in role:
                        self.states[role] = 'hover'
                        self.auto_clients[role]()

    def iteration(self, event):
        for role in self.players:
            if self.states[role] == 'hover':
                self.hover(role, self.xes[role])
                # print(role +'hover at: ', self.xes[role])
            elif self.states[role] == 'play':
                t = event.current_real - event.last_real
                self.game(t.secs + t.nsecs * 1e-9)