def setUpClass(cls): """ Launches a simulation and spawns the Pepper virtual robot """ cls.pepper_virtual = PepperVirtual() cls.pepper_virtual.loadRobot( [0, 0, 0], [0, 0, 0, 1])
def setUpClass(cls): """ Launches a simulation and spawns the Pepper virtual robot """ CameraTest.robot = PepperVirtual() CameraTest.robot.loadRobot([0, 0, 0], [0, 0, 0, 1])
def create_dataset(number_images=8000, image_size=(60, 80), dir_dataset="dataset", dir_bkgd="dataset/background_dataset", keep_green=False): """ Create a dataset of first-person-view images of the Pepper robot moving its right arm. A simulated Pepper created thanks to qibullet moves its arm in front of a green wall by randomly changing its shoulder roll in [-1,0], shoulder pitch in [-1,1], elbow roll in [0, 1], and elbow yaw in [-1, 1]. This green background is then replaced by images from another dataset. Parameters: number_images - number of images to generate image_size - (height, width) of the generated images dir_dataset - directory where to save the dataset dir_bkgd - directory where to find the background dataset keep_green - flag to also save the original images with the green background """ # todo: define the joints to be explored as inputs # check background directory if not os.path.exists(dir_bkgd): print("Error: incorrect path for the background dataset - {} doesn't exists.".format(dir_bkgd)) return # list the background images bkgd_images_list = glob.glob(dir_bkgd + "/*.png") if len(bkgd_images_list) == 0: print("Error: the directory {} doesn't contain any png image.".format(dir_bkgd)) return # check destination directory if os.path.exists(dir_dataset): ans = input("> The folder {} already exists; do you want to overwrite its content? [y,n]: ".format(dir_dataset)) if ans is not "y": print("exiting the program") return # create directories if necessary dir_combined = dir_dataset + "/combined" if not os.path.exists(dir_combined): os.makedirs(dir_combined) if keep_green: dir_green = dir_dataset + "/green" if not os.path.exists(dir_green): os.makedirs(dir_green) # check the desired image_size camera_resolution = (240, 320) if any([a > b for a, b in zip(image_size, camera_resolution)]): ans = input("Warning: the desired image size is larger than Pepper's camera resolution (240, 320). Continue? [y, n]: ") if ans is not "y": print("exiting the program") return # initialize the simulator pb.connect(pb.GUI) # configure OpenGL settings pb.configureDebugVisualizer(pb.COV_ENABLE_RENDERING, 1) pb.configureDebugVisualizer(pb.COV_ENABLE_RGB_BUFFER_PREVIEW, 0) pb.configureDebugVisualizer(pb.COV_ENABLE_DEPTH_BUFFER_PREVIEW, 0) pb.configureDebugVisualizer(pb.COV_ENABLE_SEGMENTATION_MARK_PREVIEW, 0) # command the server to run in realtime pb.setRealTimeSimulation(1) # set gravity pb.setGravity(0, 0, -9.81) # use pybullet own data files pb.setAdditionalSearchPath(pybullet_data.getDataPath()) pb.loadMJCF("mjcf/ground_plane.xml") # create the Pepper robot pepper = PepperVirtual() pepper.loadRobot([0, 0, 0], [0, 0, 0, 1]) pepper_id = 1 # define the initial posture joint_parameters = list() for name, joint in pepper.joint_dict.items(): if "Finger" not in name and "Thumb" not in name: if name == "HeadPitch": pepper.setAngles(name, -0.253, 1.0) elif name == "HeadYaw": pepper.setAngles(name, -0.7, 1.0) else: pepper.setAngles(name, 0, 1.0) joint_parameters.append((len(joint_parameters), name)) # subscribe to the camera pepper.subscribeCamera(PepperVirtual.ID_CAMERA_BOTTOM) # turn the Pepper around to avoid shadows pb.resetBasePositionAndOrientation(pepper_id, posObj=[0, 0, 0], ornObj=pb.getQuaternionFromEuler((0, 0, 1.25 * np.pi))) # add a green wall in front of Pepper wall = pb.createVisualShape(pb.GEOM_BOX, halfExtents=[0.1, 3, 1], rgbaColor=[0, 1, 0, 1]) pb.createMultiBody(baseVisualShapeIndex=wall, basePosition=[-0.65, 0, 1]) # allocate memory joints_configs = np.full((number_images, 4), np.nan) # generate the images print("generating the data...") for t in range(number_images): if t % max(1, number_images//100) == 0: print("\r{:3.0f}%".format((t+1)/number_images * 100), end="") # draw random joints configurations RShoulRoll_val = float(np.random.rand(1) - 1) RElbRoll_val = float(np.random.rand(1)) RElbYaw_val = float(2 * np.random.rand(1) - 1) RShoulPitch_val = float(2 * np.random.rand(1) - 1) # set the posture pepper.setAngles("RShoulderRoll", RShoulRoll_val, 1) pepper.setAngles("RElbowRoll", RElbRoll_val, 1) pepper.setAngles("RElbowYaw", RElbYaw_val, 1) pepper.setAngles("RShoulderPitch", RShoulPitch_val, 1) # wait for the movement to be finished cv2.waitKey(800) # get the camera input and display it img = pepper.getCameraFrame() cv2.imshow("bottom camera", img) cv2.waitKey(1) # save the joint_configuration joints_configs[t, :] = [RShoulRoll_val, RElbRoll_val, RElbYaw_val, RShoulPitch_val] # downscale the image image_green = cv2.resize(img, dsize=image_size[::-1], interpolation=cv2.INTER_NEAREST) # save the green image lossless if keep_green: filename = dir_green + "/img_green_{:05}.png".format(t) cv2.imwrite(filename, image_green, [cv2.IMWRITE_PNG_COMPRESSION, 0]) # draw a random background background = cv2.imread(np.random.choice(bkgd_images_list)) # rescale the background if necessary if not background.shape == image_size: background = cv2.resize(background, dsize=image_size[::-1], interpolation=cv2.INTER_NEAREST) # fill the green background with the background image to_fill = (image_green[:, :, 0] == 0) & (image_green[:, :, 1] == 141) & (image_green[:, :, 2] == 0) image_green[to_fill, :] = background[to_fill, :] # save the composite image lossless filename = dir_combined + "/img_{:05}.png".format(t) cv2.imwrite(filename, image_green, [cv2.IMWRITE_PNG_COMPRESSION, 0]) # display the composite image cv2.imshow('mit', image_green) cv2.waitKey(1) # stop the simulation for camera in Camera._getInstances(): camera._resetActiveCamera() for controller in BaseController._getInstances(): controller._terminateController() pb.disconnect() # save the joint configuration data with open(dir_combined + "/positions.txt", 'w') as file: json.dump(joints_configs.tolist(), file) if keep_green: with open(dir_green + "/positions.txt", 'w') as file: json.dump(joints_configs.tolist(), file) print("data generation finished")
def deplacement(X, Y, Theta): PepperVirtual.getPosition() PepperVirtual.moveTo(X, Y, Theta, 0)
from qibullet import PepperVirtual if __name__ == "__main__": physicsClient = p.connect(p.GUI) p.configureDebugVisualizer(p.COV_ENABLE_RENDERING, 1) p.configureDebugVisualizer(p.COV_ENABLE_RGB_BUFFER_PREVIEW, 0) p.configureDebugVisualizer(p.COV_ENABLE_DEPTH_BUFFER_PREVIEW, 0) p.configureDebugVisualizer(p.COV_ENABLE_SEGMENTATION_MARK_PREVIEW, 0) p.setRealTimeSimulation(1) p.setGravity(0, 0, -9.81) p.setAdditionalSearchPath(pybullet_data.getDataPath()) p.loadMJCF("mjcf/ground_plane.xml") pepper = PepperVirtual() pepper.loadRobot([0, 0, 0], [0, 0, 0, 1]) joint_parameters = list() for name, joint in pepper.joint_dict.items(): if "Finger" not in name and "Thumb" not in name: joint_parameters.append((p.addUserDebugParameter(name, -4, 4, 0), name)) pepper.subscribeCamera(PepperVirtual.ID_CAMERA_BOTTOM) while True: img = pepper.getCameraFrame() cv2.imshow("bottom camera", img) cv2.waitKey(1)