def start_dance(self): """ Start a robotic dance """ dance.dance(self.motion_service)
def test_dance(self): SERIAL_PORT = "/dev/ttyUSB0" r = create.Create(SERIAL_PORT) dance.dance(r, 2) r.shutdown()
length, data = audio.inp.read() signal = numpy.fromstring(data, dtype='int16') fourier = numpy.fft.fft(signal) / length # apply fft and normalize fourier = fourier[range(length / 4)] line.set_ydata(abs(fourier.real)) return line, if __name__ == '__main__': while True: length, data = audio.inp.read() a = numpy.fromstring(data, dtype='int16') if not recent: if dance(length, a): timeout = 0 recent = True else: if timeout < audio.RATE * .4: timeout += audio.CHUNK else: recent = False #import matplotlib.pyplot as plt #import matplotlib.animation as animation #length, data = audio.inp.read() #signal = numpy.fromstring(data, dtype='int16') #fig, ax = plt.subplots()
elif op == "crossRoad": send_op(op, B_ip) elif op == "Monitor": det = Monitor.faceRecognition() if det == False: send_op("alter", B_ip) elif op == "chat": utterance = raw_input("utterance: ") token = gen_token() while utterance != "exit": data = "chat " + utterance + " token" send_op(data, C_ip) utterance = raw_input("utterance: ") elif op == "mood": cur_mood = mood.mood_judge() if mood_judge() == "Sad": RobotApi.ubtVoiceTTS(0,"主人别伤心辣,我给你跳支舞吧") send_op(op, B_ip) send_op(op, C_ip) dance() else: print("你看起来一点儿也不悲伤呀~") else: print("invalid options!") print("options: morning, crossRoad, Monitor, chat, mood") op = str(raw_input("please enter op: ")) #-----------Disconnect--------- RobotApi.ubtRobotDisconnect("SDK","1",gIPAddr) RobotApi.ubtRobotDeinitialize()