コード例 #1
0
    def _init_robots(self):

        if self.compensation:
            self.controller_def = SwordFightZMQController(mode="def",
                                                          host="flogo4.local")
            self.controller_att = SwordFightZMQController(mode="att",
                                                          host="flogo2.local")
        else:
            self.controller_def = ZMQController(host="flogo4.local")
            self.controller_att = ZMQController(host="flogo2.local")

        self._setSpeedCompliance()

        self.controller_def.get_keys()  # in case there are keys stored
コード例 #2
0
import time

import numpy as np

from poppy_helpers.controller import ZMQController

zmq = ZMQController("flogo3.local")
# zmq.compliant(True)
zmq.compliant(False)
zmq.set_max_speed(100)
# zmq.goto_pos([0, -30, 0, 0, -40, 0]) # no gripper
# zmq.goto_pos([0, -30, 0, 0, -40, 30]) # gripper open
# zmq.goto_pos([0, 20, -30, 0, -30, -60]) # gripper closed
zmq.goto_pos([0, 0, 0, 0, 0, 0])  # rest
# for i in range(6):
#     zmq.set_color(i+1,"green")

# Duckie-tracker
# python pytorch_a2c_ppo_acktr/enjoy.py --env-name ErgoReacher-Tracking-Live-v1 --custom-gym poppy_helpers --model ./trained_models/ppo/ErgoReacher-Headless-Gripper-MobileGoal-v1-190404011143-983040.pt
コード例 #3
0
import os
import time

import numpy as np
from realsense_tracker.camera import Camera
from realsense_tracker.tracker import Tracker, TRACKING_GREEN

from poppy_helpers.config import config_dir
from poppy_helpers.controller import ZMQController

zmq = ZMQController("pokey.local")
zmq.compliant(False)
zmq.set_max_speed(100)
zmq.goto_pos([0, 0, 0])
print(zmq.get_posvel())
コード例 #4
0
import numpy as np
from realsense_tracker.camera import Camera
from realsense_tracker.tracker import Tracker, TRACKING_GREEN

from poppy_helpers.config import config_dir
from poppy_helpers.controller import ZMQController

ROBOT = True

cam = Camera(color=True)

tracker = Tracker(TRACKING_GREEN["nas_lower"], TRACKING_GREEN["nas_upper"])

if ROBOT:
    zmq = ZMQController("flogo3.local")
    zmq.compliant(False)
    zmq.set_max_speed(100)

    # rest
    zmq.goto_pos([0, 0, 0, 0, 0, 0])
    time.sleep(2)

# max forward   posx: 0.224 posy: 0.102,    [1, -1, 0, 0]
# max backward  posx: -0.148 posy: 0.016,   [-1, -1, 0, 0]
# max up        posx: 0.013	posy: 0.25,     [.1, -1, -.1, 0]

calibration = np.zeros((3, 2), dtype=np.uint16)

# full forward
コード例 #5
0
    def __init__(self):
        self.goals_done = 0

        self.metadata = {'render.modes': ['human']}

        # observation = 3 joints + 3 velocities + 2 puck position + 2 coordinates for target
        self.observation_space = spaces.Box(low=-1,
                                            high=1,
                                            shape=(3 + 3 + 2 + 2, ),
                                            dtype=np.float32)  #

        # action = 3 joint angles
        self.action_space = spaces.Box(low=-1,
                                       high=1,
                                       shape=(3, ),
                                       dtype=np.float32)  #

        self.cam = Camera(color=True)
        self.tracker_puck = Tracker(TRACKING_RED["pusher_lower"],
                                    TRACKING_RED["pusher_upper"])
        self.tracker_pusher = Tracker(TRACKING_GREEN["pusher_lower"],
                                      TRACKING_GREEN["pusher_upper"])

        self.calib_sim = np.load(
            os.path.join(MODEL_PATH, "calibration-pusher-sim.npz"))['arr_0']
        self.calib_real = np.load(
            os.path.join(config_dir(),
                         "calibration-pusher-real.npz"))["arr_0"].astype(
                             np.int16)

        self.x_rr = self.calib_real[0, 0]
        self.x_rl = self.calib_real[2, 0]
        self.x_sr = self.calib_sim[0, 0]
        self.x_sl = self.calib_sim[2, 0]

        self.y_rr = self.calib_real[0, 1]
        self.y_rl = self.calib_real[2, 1]
        self.y_sr = self.calib_sim[0, 1]
        self.y_sl = self.calib_sim[2, 1]

        self.x_rm = self.calib_real[1, 0]
        self.y_rm = self.calib_real[1, 1]
        self.x_sm = self.calib_sim[1, 0]
        self.y_sm = self.calib_sim[1, 1]

        self.x_rc = (self.x_rl + self.x_rr) / 2
        self.y_rc = (self.y_rl + self.y_rr) / 2

        # [0.13473208 0.00849917]
        # [0.00820219 0.1347505]
        # [-0.13471367  0.00878573]

        # [56 266]
        # [226 441]
        # [398 251]

        self.last_step_time = time.time()
        self.pts_tip = deque(maxlen=32)
        self.pts_puck = deque(maxlen=32)

        self.rest_pos = np.array([-.5, 1, .5])

        # self.last_frame = np.ones((480, 640, 3), dtype=np.uint8)

        # DEBUGGING CODE

        # frame = Camera.to_numpy(self.cam.get_color())[:, :, ::-1]
        # frame2 = np.ascontiguousarray(frame, dtype=np.uint8)
        # print (frame.shape)
        # for _ in range(500):
        #     self._sample_goal()
        #
        #     pix = self._sim2pixel(self.goal)
        #     print(self.goal, pix)
        #     cv2.circle(frame2, (int(pix[0]), int(pix[1])), 5, (255,0,0), 4)
        #
        # cv2.circle(frame2, (int(50), int(50)), 10, (0, 0, 0), 4)
        # cv2.circle(frame2, (int(50), int(5)), 10, (0, 255, 0), 4)
        # cv2.circle(frame2, (int(5), int(50)), 10, (0, 255, 255), 4)
        #
        # cv2.imshow("Frame2", frame2)
        # cv2.waitKey(10000)
        #
        # quit()

        self.controller = ZMQController(host="pokey.local")
        self._setSpeedCompliance()

        super().__init__()
コード例 #6
0
class ErgoPusherLiveEnv(gym.Env):
    def __init__(self):
        self.goals_done = 0

        self.metadata = {'render.modes': ['human']}

        # observation = 3 joints + 3 velocities + 2 puck position + 2 coordinates for target
        self.observation_space = spaces.Box(low=-1,
                                            high=1,
                                            shape=(3 + 3 + 2 + 2, ),
                                            dtype=np.float32)  #

        # action = 3 joint angles
        self.action_space = spaces.Box(low=-1,
                                       high=1,
                                       shape=(3, ),
                                       dtype=np.float32)  #

        self.cam = Camera(color=True)
        self.tracker_puck = Tracker(TRACKING_RED["pusher_lower"],
                                    TRACKING_RED["pusher_upper"])
        self.tracker_pusher = Tracker(TRACKING_GREEN["pusher_lower"],
                                      TRACKING_GREEN["pusher_upper"])

        self.calib_sim = np.load(
            os.path.join(MODEL_PATH, "calibration-pusher-sim.npz"))['arr_0']
        self.calib_real = np.load(
            os.path.join(config_dir(),
                         "calibration-pusher-real.npz"))["arr_0"].astype(
                             np.int16)

        self.x_rr = self.calib_real[0, 0]
        self.x_rl = self.calib_real[2, 0]
        self.x_sr = self.calib_sim[0, 0]
        self.x_sl = self.calib_sim[2, 0]

        self.y_rr = self.calib_real[0, 1]
        self.y_rl = self.calib_real[2, 1]
        self.y_sr = self.calib_sim[0, 1]
        self.y_sl = self.calib_sim[2, 1]

        self.x_rm = self.calib_real[1, 0]
        self.y_rm = self.calib_real[1, 1]
        self.x_sm = self.calib_sim[1, 0]
        self.y_sm = self.calib_sim[1, 1]

        self.x_rc = (self.x_rl + self.x_rr) / 2
        self.y_rc = (self.y_rl + self.y_rr) / 2

        # [0.13473208 0.00849917]
        # [0.00820219 0.1347505]
        # [-0.13471367  0.00878573]

        # [56 266]
        # [226 441]
        # [398 251]

        self.last_step_time = time.time()
        self.pts_tip = deque(maxlen=32)
        self.pts_puck = deque(maxlen=32)

        self.rest_pos = np.array([-.5, 1, .5])

        # self.last_frame = np.ones((480, 640, 3), dtype=np.uint8)

        # DEBUGGING CODE

        # frame = Camera.to_numpy(self.cam.get_color())[:, :, ::-1]
        # frame2 = np.ascontiguousarray(frame, dtype=np.uint8)
        # print (frame.shape)
        # for _ in range(500):
        #     self._sample_goal()
        #
        #     pix = self._sim2pixel(self.goal)
        #     print(self.goal, pix)
        #     cv2.circle(frame2, (int(pix[0]), int(pix[1])), 5, (255,0,0), 4)
        #
        # cv2.circle(frame2, (int(50), int(50)), 10, (0, 0, 0), 4)
        # cv2.circle(frame2, (int(50), int(5)), 10, (0, 255, 0), 4)
        # cv2.circle(frame2, (int(5), int(50)), 10, (0, 255, 255), 4)
        #
        # cv2.imshow("Frame2", frame2)
        # cv2.waitKey(10000)
        #
        # quit()

        self.controller = ZMQController(host="pokey.local")
        self._setSpeedCompliance()

        super().__init__()

    def sim2real(self, act):
        act = np.clip(act, -1, 1)
        return ((np.array(act) + self.rest_pos) * -90).tolist()

    def _setSpeedCompliance(self):
        self.controller.compliant(False)
        self.controller.set_max_speed(100)  # default: 100

    def setSpeed(self, speed):
        assert speed > 0 and speed < 1000
        self.controller.set_max_speed(speed)
        self.last_speed = speed

    def seed(self, seed=None):
        return [np.random.seed(seed)]

    def step(self, action):
        action = self.sim2real(action)
        self.controller.goto_pos(action)

        reward, done, distance, frame, center_puck = self._get_reward()
        cv2.imshow("Live Feed", frame)
        cv2.waitKey(1)

        obs = self._get_obs(center_puck)
        return obs, reward, done, {"distance": distance, "img": frame}

    def _sim2pixel(self, sim_coords):
        x_t = sim_coords[0]
        x = -x_t * (self.x_rl - self.x_rr) / (self.x_sr -
                                              self.x_sl) + self.x_rc

        y_t = sim_coords[1]
        y = (y_t - self.y_sr) * (self.y_rm - self.y_rc) / (
            self.y_sm - self.y_sr) + self.y_rc

        return [int(x), int(y)]

    def _pixel2sim(self, pix_coords):
        x_g = pix_coords[0]
        x = (self.x_rc - x_g) * (self.x_sr - self.x_sl) / (self.x_rl -
                                                           self.x_rr)

        y_g = pix_coords[1]
        y = (y_g - self.y_rc) * (self.y_sm - self.y_sr) / (
            self.y_rm - self.y_rc) + self.y_sr

        return [x, y]

    def _img_and_track(self):
        while True:
            frame = Camera.to_numpy(
                self.cam.get_color())[:, :, ::-1]  # RGB to BGR for cv2
            hsv = self.tracker_pusher.blur_img(frame)
            # mask_tip = self.tracker_pusher.prep_image(hsv)
            #
            # # center of mass, radius of enclosing circle, x/y of enclosing circle
            # center_tip, radius_tip, x_tip, y_tip = self.tracker_pusher.track(mask_tip)

            mask_puck = self.tracker_puck.prep_image(hsv)
            center_puck, radius_puck, x_puck, y_puck = self.tracker_puck.track(
                mask_puck)

            # grab more frames until the green blob is big enough / visible
            if center_puck is not None and radius_puck > DETECTION_RADIUS:
                break

        frame2 = np.ascontiguousarray(frame, dtype=np.uint8)

        goal = self._sim2pixel(self.goal)

        self._render_tracking(frame2,
                              center_puck,
                              radius_puck,
                              x_puck,
                              y_puck,
                              pts=self.pts_puck)

        cv2.circle(frame2, tuple(goal), 10, (255, 0, 255), 4)

        return frame2, self._pixel2sim(center_puck)

    def reset(self):
        self.setSpeed(100)

        self._sample_goal_and_puck()

        qpos = np.random.uniform(low=-0.1, high=0.1, size=3)

        self.controller.goto_pos(self.sim2real(qpos))

        while True:
            frame, center = self._img_and_track()

            puck = self._sim2pixel(self.puck)

            cv2.circle(frame, tuple(puck), 10, (255, 0, 0), 4)

            add_text(frame, "MOVE PUCK TO BLUE CIRCLE AND PRESS KEY",
                     (0, 0, 0), .5)

            cv2.imshow("Live Feed", frame)
            key = cv2.waitKey(1) & 0xFF

            if key != 255:
                break

        return self._get_obs()

    def _get_obs(self, center_puck=None):
        pv = self.controller.get_posvel()

        goal = self.normalize_goal()

        if center_puck is None:
            puck = self.normalize_puck(self.puck)
        else:
            puck = self.normalize_puck(center_puck)

        self.observation = np.hstack((self._normalize(pv), puck, goal))
        return self.observation

    def _normalize(self, pos):
        pos = np.array(pos).astype('float32') * -1  # joints are inverted
        pos[:3] = ((pos[:3] + 90) / 180) * 2 - 1  # positions
        pos[3:] = ((pos[3:] + 300) / 600) * 2 - 1  # velocities
        return pos

    def _render_tracking(self, frame, center, radius, x, y, pts):
        # circle center
        cv2.circle(frame, (int(x), int(y)), int(radius), (0, 255, 255), 2)

        # center of mass
        cv2.circle(frame, center, 5, (0, 0, 255), -1)

        # update the points queue
        pts.appendleft(center)

        # loop over the set of tracked points
        for i in range(1, len(pts)):
            # if either of the tracked points are None, ignore
            # them
            if pts[i - 1] is None or pts[i] is None:
                continue

            # otherwise, compute the thickness of the line and
            # draw the connecting lines
            thickness = int(np.sqrt(32 / float(i + 1)) * 2.5)
            cv2.line(frame, pts[i - 1], pts[i], (0, 0, 255), thickness)

        return frame

    def _sample_goal_and_puck(self):
        self.goal = [
            np.random.uniform(PUSHER_GOAL_X[0], PUSHER_GOAL_X[1]),
            np.random.uniform(PUSHER_GOAL_Y[0], PUSHER_GOAL_Y[1])
        ]
        self.puck = [
            np.random.uniform(PUSHER_PUCK_X[0], PUSHER_PUCK_X[1]),
            np.random.uniform(PUSHER_PUCK_Y[0], PUSHER_PUCK_Y[1])
        ]

    def normalize_goal(self):
        x = (self.goal[0] - PUSHER_GOAL_X[0]) / (PUSHER_GOAL_X[1] -
                                                 PUSHER_GOAL_X[0])
        y = (self.goal[1] - PUSHER_GOAL_Y[0]) / (PUSHER_GOAL_Y[1] -
                                                 PUSHER_GOAL_Y[0])
        return np.array([x, y])

    def normalize_puck(self, puck):
        x = (puck[0] - PUSHER_PUCK_X_NORM[0]) / (PUSHER_PUCK_X_NORM[1] -
                                                 PUSHER_PUCK_X_NORM[0])
        y = (puck[1] - PUSHER_PUCK_Y_NORM[0]) / (PUSHER_PUCK_Y_NORM[1] -
                                                 PUSHER_PUCK_Y_NORM[0])
        return np.array([x, y])

    def _get_reward(self):
        done = False

        frame, center_puck = self._img_and_track()

        reward = np.linalg.norm(np.array(self.goal) - np.array(center_puck))
        distance = reward.copy()

        reward *= -1  # the reward is the inverse distance
        if distance < MIN_DIST:  # this is a bit arbitrary, but works well
            done = True
            reward = 1

        return reward, done, distance, frame.copy(), center_puck
コード例 #7
0
 def _init_robot(self):
     self.controller = ZMQController(host="flogo3.local")
     self._setSpeedCompliance()
コード例 #8
0
class ErgoReacherLiveEnv(gym.Env):
    def __init__(self, multi=False, multi_no=3, tracking=False):
        self.multigoal = multi
        self.n_goals = multi_no
        self.tracking = tracking
        self.rand = Randomizer()

        self.goal = None

        # self.step_in_episode = 0

        self.metadata = {
            'render.modes': ['human', 'rgb_array']
        }

        self.rhis = RandomPointInHalfSphere(0.0, 0.0369, 0.0437,
                                            radius=0.2022, height=0.2610,
                                            min_dist=0.1, halfsphere=True)

        # observation = 4 joints + 4 velocities + 2 coordinates for target
        joints_no = 4
        if self.tracking:
            joints_no = 3

        self.observation_space = spaces.Box(low=-1, high=1, shape=(joints_no + joints_no + 2,), dtype=np.float32)

        # action = 4 joint angles
        self.action_space = spaces.Box(low=-1, high=1, shape=(joints_no,), dtype=np.float32)

        self.diffs = [JOINT_LIMITS[i][1] - JOINT_LIMITS[i][0] for i in range(6)]

        self.cam = Camera(color=True)
        self.tracker = Tracker(TRACKING_GREEN["maxime_lower"], TRACKING_GREEN["maxime_upper"])
        # self.tracker = Tracker(TRACKING_GREEN["duct_lower"], TRACKING_GREEN["duct_upper"])
        if self.tracking:
            self.tracker_goal = Tracker(TRACKING_YELLOW["duckie_lower"], TRACKING_YELLOW["duckie_upper"])

        self.calibration = np.load(os.path.join(config_dir(), "calib-ergoreacher-adr.npz"))["calibration"].astype(
            np.int16)

        self.last_step_time = time.time()
        self.pts_tip = deque(maxlen=32)
        self.pts_goal = deque(maxlen=32)

        self.last_frame = np.ones((480, 640, 3), dtype=np.uint8)

        self.pause_counter = 0

        self.last_speed = 100

        self.goals_done = 0

        # self.goal_states=[]

        # for _ in range(10000):
        #     goal = self.rhis.sampleSimplePoint()
        #     self.goal_states.append(self._goal2pixel(goal))

        self._init_robot()

        super().__init__()

    def _init_robot(self):
        self.controller = ZMQController(host="flogo3.local")
        self._setSpeedCompliance()

    def _setSpeedCompliance(self):
        self.controller.compliant(False)
        self.controller.set_max_speed(500)  # default: 100

    def setSpeed(self, speed):
        assert speed > 0 and speed < 1000
        self.controller.set_max_speed(speed)
        self.last_speed = speed

    def seed(self, seed=None):
        np.random.seed(seed)

    def reset(self):
        self.setSpeed(100)  # do resetting at a normal speed

        if self.multigoal:
            # sample N goals, calculate total reward as distance between them. Add distances to list. Subtract list elements on rew calculation
            self.goal_distances = []
            self.goal_positions = []
            for goal_idx in range(self.n_goals):
                point = self.rhis.sampleSimplePoint()
                self.goal_positions.append(point)
            for goal_idx in range(self.n_goals - 1):
                dist = np.linalg.norm(self.goal_positions[goal_idx] -
                                      self.goal_positions[goal_idx + 1])
                self.goal_distances.append(dist)

        self.gripper_closed = False
        self.gripper_closed_frames = 0
        self.closest_distance = np.inf
        self.ready_to_close = False
        self.tracking_frames = 0

        if not self.tracking:
            self.goal = self.rhis.sampleSimplePoint()

        # goals = []

        # for _ in range(100000):
        #     goals.append(self.rhis.sampleSimplePoint())
        #
        # goals = np.array(goals)
        # print ("x min/max",goals[:,1].min(),goals[:,1].max())
        # print ("y min/max",goals[:,2].min(),goals[:,2].max())

        qpos = np.random.uniform(low=-0.2, high=0.2, size=6)
        qpos[[0, 3]] = 0

        self.pts_tip.clear()

        add_text(self.last_frame, "=== RESETTING ===")

        if (self.pause_counter > ITERATIONS_MAX):
            self.pause_counter = 0
            self.controller.compliant(True)
            input("\n\n=== MAINTENANCE: PLEASE CHECK THE ROBOT AND THEN PRESS ENTER TO CONTINUE ===")
            self.controller.compliant(False)
            time.sleep(.5)  # wait briefly to make sure all joints are non-compliant / powered

        # this counts as rest position
        self.controller.goto_normalized(qpos)

        cv2.imshow("Frame", self.last_frame)
        cv2.waitKey(1000)

        if self.multigoal:

            while True:
                frame = Camera.to_numpy(self.cam.get_color())[:, :, ::-1]  # RGB to BGR for cv2
                hsv = self.tracker.blur_img(frame)
                mask_tip = self.tracker.prep_image(hsv)

                # center of mass, radius of enclosing circle, x/y of enclosing circle
                center_tip, radius_tip, x_tip, y_tip = self.tracker.track(mask_tip)

                # grab more frames until the green blob is big enough / visible
                if center_tip is not None and radius_tip > DETECTION_RADIUS:
                    break

            pos_tip = self._pixel2goal(center_tip)

            reward = np.linalg.norm(np.array(self.goal[1:]) - np.array(pos_tip))
            distance = reward.copy()

            self.goal_distances.append(distance)

        self.setSpeed(self.last_speed)

        self.last_step_time = time.time()

        return self._get_obs()

    def _get_obs(self):
        if self.goal is None:
            return np.zeros(8)

        pv = self.controller.get_posvel()
        if not self.tracking:
            self.observation = self._normalize(pv)[[1, 2, 4, 5, 7, 8, 10, 11]]  # leave out joint0/3
        else:
            self.observation = self._normalize(pv)[[1, 2, 4, 7, 8, 10]]  # leave out joint0/3/5
        self.observation = np.hstack((self.observation, self.rhis.normalize(self.goal)[1:]))
        return self.observation

    def _normalize(self, pos):
        pos = np.array(pos).astype('float32')
        pos[:6] = ((pos[:6] + 90) / 180) * 2 - 1  # positions
        pos[6:] = ((pos[6:] + 300) / 600) * 2 - 1  # velocities
        return pos

    def _pixel2goal(self, camcenter):
        pos = np.array(camcenter).astype(np.int16)
        x_n = (pos[0] - self.calibration[0, 0]) / (self.calibration[1, 0] - self.calibration[0, 0])
        x_d = .224 - (x_n * (.224 + .148))
        y_n = (pos[1] - self.calibration[2, 1]) / (self.calibration[1, 1] - self.calibration[2, 1])
        y_d = .25 - (y_n * (.23 - .046))  # .056 has been modified from the .016 below
        return np.array([x_d, y_d])

    def _goal2pixel(self, goal):
        # print (self.goal[1:], self.calibration)

        # x_n = (float(goal[1]) + .0979) / (.2390 + .0979)
        # x_d = x_n * (self.calibration[0, 0] - self.calibration[1, 0]) + self.calibration[1, 0]
        # y_n = (float(goal[2]) - .0437) / (.2458 - .0437)
        # y_d = self.calibration[0, 1] - (y_n * (self.calibration[0, 1] - self.calibration[2, 1]))

        x_n = (float(goal[1]) + .148) / (.224 + .148)
        x_d = x_n * (self.calibration[0, 0] - self.calibration[1, 0]) + self.calibration[1, 0]
        y_n = (float(goal[2]) - .016) / (.25 - .016)
        y_d = self.calibration[0, 1] - (y_n * (self.calibration[0, 1] - self.calibration[2, 1]))
        # print ([x_d, y_d])
        return np.array([int(round(x_d)), int(round(y_d))])

    def _render_img(self, frame, center, radius, x, y, pts, color_a=(0, 255, 255), color_b=(0, 0, 255)):
        g = self._goal2pixel(self.goal)
        cv2.circle(frame, (g[0], g[1]), int(5), (255, 0, 255), 3)

        # circle center
        cv2.circle(frame, (int(x), int(y)), int(radius), color_a, 2)

        # center of mass
        cv2.circle(frame, center, 5, color_b, -1)

        # update the points queue
        pts.appendleft(center)

        # loop over the set of tracked points
        for i in range(1, len(pts)):
            # if either of the tracked points are None, ignore
            # them
            if pts[i - 1] is None or pts[i] is None:
                continue

            # otherwise, compute the thickness of the line and
            # draw the connecting lines
            thickness = int(np.sqrt(32 / float(i + 1)) * 2.5)
            cv2.line(frame, pts[i - 1], pts[i], color_b, thickness)

        self.last_frame = frame.copy()

        return frame

    def _get_reward(self):
        done = False
        center_goal = None

        while True:
            frame = Camera.to_numpy(self.cam.get_color())[:, :, ::-1]  # RGB to BGR for cv2
            hsv = self.tracker.blur_img(frame)
            mask_tip = self.tracker.prep_image(hsv)

            # center of mass, radius of enclosing circle, x/y of enclosing circle
            center_tip, radius_tip, x_tip, y_tip = self.tracker.track(mask_tip)

            if self.tracking:
                mask_goal = self.tracker_goal.prep_image(hsv)
                center_goal, radius_goal, x_goal, y_goal = self.tracker_goal.track(mask_goal)

            # grab more frames until the green blob is big enough / visible
            if center_tip is not None and radius_tip > DETECTION_RADIUS:
                break

        frame2 = np.ascontiguousarray(frame, dtype=np.uint8)

        if self.tracking and center_goal is not None:
            self.goal = np.zeros(3, dtype=np.float32)
            self.goal[1:] = self._pixel2goal(center_goal)

            frame2 = self._render_img(frame2, center_goal, radius_goal, x_goal, y_goal, pts=self.pts_goal)

        if self.goal is not None:
            frame2 = self._render_img(frame2, center_tip, radius_tip, x_tip, y_tip, pts=self.pts_tip,
                                      color_a=(255, 255, 0), color_b=(255, 0, 0))

        if self.tracking and center_goal is None:
            self.goal = None
            return 0, False, np.inf, frame2.copy()

        pos_tip = self._pixel2goal(center_tip)

        reward = np.linalg.norm(np.array(self.goal[1:]) - np.array(pos_tip))
        distance = reward.copy()
        reward *= -1  # the reward is the inverse distance

        if not self.multigoal:
            if reward > MIN_DIST:  # this is a bit arbitrary, but works well
                done = True
                reward = 1

        else:
            reward *= -1
            # self.goal = self.rhis.sampleSimplePoint()

            dirty = False  # in case we _just_ hit the goal
            if -reward > MIN_DIST:
                self.goals_done += 1
                if self.goals_done == self.n_goals:
                    done = True
                else:
                    # TODO: MOVE BALL
                    dirty = True

            if done or self.goals_done == self.n_goals:
                reward = 1
            else:
                # reward is distance to current target + sum of all other distances divided by total distance
                if dirty:
                    # take it off before the reward calc
                    self.goals_done -= 1

                reward = 1 + (-(reward + sum(
                    self.goal_distances[:-(self.goals_done + 1)])) /
                              sum(self.goal_distances))
                if dirty:
                    # add it back after the reward cald
                    self.goals_done += 1

        return reward, done, distance, frame2.copy()

    def step(self, action):
        action_ = np.zeros(6, np.float32)
        if not self.tracking:
            action_[[1, 2, 4, 5]] = action
        else:
            action_[[1, 2, 4]] = action
            action_[5] = 50 / 90

            if self.gripper_closed:
                self.gripper_closed_frames += 1
                action_[5] = -10 / 90

                if self.gripper_closed_frames >= GRIPPER_CLOSED_MAX_FRAMES:
                    self.gripper_closed = False
                    self.gripper_closed_frames = 0

        action = np.clip(action_, -1, 1)

        if self.goal is not None:
            self.controller.goto_normalized(action)

        reward, done, distance, frame = self._get_reward()
        cv2.imshow("Frame", frame)
        cv2.waitKey(1)

        if self.tracking:
            done = False

        if not self.gripper_closed:
            if distance < self.closest_distance:
                self.closest_distance = distance
                self.ready_to_close += 1
            else:
                if self.ready_to_close > CLOSING_FRAMES:
                    self.ready_to_close = 0
                    self.closest_distance = np.inf
                    self.gripper_closed = True  # only takes effect on next step
                else:
                    self.tracking_frames += 1
                    if self.tracking_frames > 50:
                        self.closest_distance = np.inf

        dt = (time.time() - self.last_step_time) * 1000
        if dt < MAX_REFRESHRATE:
            time.sleep((MAX_REFRESHRATE - dt) / 1000)

        self.last_step_time = time.time()

        self.pause_counter += 1

        return self._get_obs(), reward, done, {"distance": distance, "img": frame}

    def render(self, mode='human', close=False):
        pass
コード例 #9
0
import time
import os
import numpy as np
from poppy_helpers.controller import ZMQController
from arguments import get_args

args = get_args()

zmq = ZMQController('flogo3.local')
zmq.compliant(False)
zmq.set_max_speed(100)

file_path = os.getcwd() + f'/data/freq{args.freq}/{args.approach}/'

rest_interval = 10 * 100
freq = args.freq
steps_until_resample = 100 / freq

actions_trajectories = np.load(file_path + 'action_trajectories.npz')
actions = actions_trajectories["actions"]
real_trajectories = np.zeros((actions.shape[0], 12))

for epi in range(actions.shape[0]):
    start = time.time()

    if epi % rest_interval == 0:
        print(f"Episodes completed: {epi}")
        zmq.rest()
    action = actions[epi, :]
    zmq.goto_normalized(action)
コード例 #10
0
import time

import numpy as np

from poppy_helpers.controller import ZMQController

zmq = ZMQController("flogo3.local")
zmq.compliant(False)  # make robot stiff
zmq.set_max_speed(100)
zmq.goto_pos([0, 0, 0, 0, 0, 0])

for i in range(6):
    action_val = 30
    if i == 5:
        action_val = 20
    action = [0] * (i) + [action_val] + [0] * (6 - i - 1)
    print(action)
    zmq.goto_pos(action)
    time.sleep(1)
    action = (np.array(action) * -1).tolist()
    print(action)
    zmq.goto_pos(action)

    print("current pos", zmq.get_pos())

    time.sleep(1)

zmq.goto_pos([0, 0, 0, 0, 0, 0])
time.sleep(.5)
コード例 #11
0
import time

import numpy as np

from poppy_helpers.controller import ZMQController

time.sleep(5)

zmq = ZMQController("pokey.local")
zmq.compliant(False)
# zmq.compliant(True)
zmq.set_max_speed(100)
zmq.goto_pos([45, -90, -45])
time.sleep(2)
zmq.set_max_speed(100)
time.sleep(.5)
zmq.goto_pos([-45, -45, 0])
コード例 #12
0
import time
import numpy as np
from poppy_helpers.controller import ZMQController

zmq = ZMQController('flogo3.local')
zmq.compliant(False)
zmq.set_max_speed(100)

action = np.array([0, 1, 0, 0, 0, 0])
zmq.goto_normalized(action)
zmq.rest()
コード例 #13
0
class ErgoFightLiveEnv(gym.Env):
    def __init__(self,
                 no_move=False,
                 scaling=1,
                 shield=True,
                 sim=False,
                 compensation=True):
        self.no_move = no_move
        self.scaling = scaling
        self.shield = shield
        self.sim = sim
        self.compensation = compensation

        self.rand = Randomizer()
        self.last_step_time = None

        self.step_in_episode = 0
        self.step_in_fight = 0

        self.metadata = {'render.modes': ['human', 'rgb_array']}

        # 6 own joint pos, 6 own joint vel, 6 enemy joint pos, 6 enemy joint vel
        self.observation_space = spaces.Box(low=-1,
                                            high=1,
                                            shape=(6 + 6 + 6 + 6, ),
                                            dtype=np.float32)

        self.action_space = spaces.Box(low=-1,
                                       high=1,
                                       shape=(6, ),
                                       dtype=np.float32)

        self.diffs = [
            JOINT_LIMITS[i][1] - JOINT_LIMITS[i][0] for i in range(6)
        ]

        self._init_robots()

    def _init_robots(self):

        if self.compensation:
            self.controller_def = SwordFightZMQController(mode="def",
                                                          host="flogo4.local")
            self.controller_att = SwordFightZMQController(mode="att",
                                                          host="flogo2.local")
        else:
            self.controller_def = ZMQController(host="flogo4.local")
            self.controller_att = ZMQController(host="flogo2.local")

        self._setSpeedCompliance()

        self.controller_def.get_keys()  # in case there are keys stored

    def _setSpeedCompliance(self):
        self.controller_def.compliant(False)
        self.controller_att.compliant(False)
        self.controller_att.set_max_speed(100)
        self.controller_def.set_max_speed(100)

    def _seed(self, seed=None):
        np.random.seed(seed)

    def _restPos(self):
        self.done = False

        self._setSpeedCompliance()

        self.controller_def.safe_rest()
        self.controller_att.safe_rest()

        time.sleep(.6)  # FIXME: run loop, check joints

        self.randomize(robot=1, scaling=self.scaling)

        time.sleep(1)  # FIXME: instead run a loop here and check when
        # the joints are close to the given configuration
        self.controller_def.get_keys()  # clear queue

    def randomize(self, robot=1, scaling=1.0):
        new_pos = self.rand.random_sf(scaling)
        robot_ctrl = self.controller_att
        if robot == 1:
            robot_ctrl = self.controller_def

        robot_ctrl.goto_pos(new_pos)

    def _reset(self):
        self.step_in_episode = 0
        self._restPos()
        self._self_observe()
        self.last_step_time = time.time()
        return self.observation

    def _get_reward(self):

        collisions = self.controller_def.get_keys()

        reward = 0
        if "s" in collisions:
            reward = 1
            self._restPos()

        return reward

    def _self_observe(self):
        joint_vel_att = self.controller_att.get_posvel()
        joint_vel_def = self.controller_def.get_posvel()
        # self.observation = ((self._normalize( ### OLD AND WRONG IMPL
        #     np.hstack((joint_vel_att, joint_vel_def)).astype('float32')
        # )), )
        self.observation = (
            (  # this should be the technically correct solution
                np.hstack((self._normalize(joint_vel_att),
                           self._normalize(joint_vel_def)))), )

    def _normalize(self, pos):
        pos = np.array(pos).astype('float32')
        out = []
        for i in range(6):
            shifted = (pos[i] -
                       JOINT_LIMITS[i][0]) / self.diffs[i]  # now it's in [0,1]
            norm = shifted * 2 - 1
            if i == 0 and self.sim:
                norm *= -1  # between sim and real the first joint it inverted
            out.append(norm)
        if len(pos) > 6:
            shifted = (pos[6:] + JOINT_LIMITS_SPEED) / (JOINT_LIMITS_SPEED * 2)
            norm = shifted * 2 - 1
            if self.sim:
                norm[
                    0] *= -1  # between sim and real the first joint it inverted (probably both pos * vel)
            out += list(norm)
        return out

    def _denormalize(self, actions):
        out = []
        if self.sim:
            actions[0] *= -1

        for i in range(6):
            shifted = (actions[i] + 1) / 2  # now it's within [0,1]
            denorm = shifted * self.diffs[i] + JOINT_LIMITS[i][0]
            out.append(denorm)
        return out

    def prep_actions(self, actions):
        actions = np.clip(actions, -1,
                          1)  # first make sure actions are normalized
        actions = self._denormalize(
            actions)  # then scale them to the actual joint angles
        return actions

    def _step(self, actions):
        self.step_in_episode += 1
        actions = self.prep_actions(actions)

        self.controller_att.goto_pos(actions)

        if not self.no_move:
            if self.step_in_episode % MOVE_EVERY_N_STEPS == 0:
                # print ("step {}, randomizing".format(self.step_in_episode))
                self.randomize(1, scaling=self.scaling)

        # observe again
        self._self_observe()
        reward = self._get_reward()

        dt = (time.time() - self.last_step_time) * 1000
        if dt < MAX_REFRESHRATE:
            time.sleep((MAX_REFRESHRATE - dt) / 1000)

        self.last_step_time = time.time()
        return self.observation, reward, self.done, {}

    def _test_second_robot(self, actions):
        actions = self.prep_actions(actions)
        self.controller_def.goto_pos(actions)

    def _render(self, mode='human', close=False):
        # This intentionally does nothing and is only here for wrapper functions.
        # if you want graphical output, use the environments
        # "ErgoBallThrowAirtime-Graphical-Normalized-v0"
        # or
        # "ErgoBallThrowAirtime-Graphical-v0"
        # ... not the ones with "...-Headless-..."
        pass