diff --git a/ros/angel_system_nodes/angel_system_nodes/audio/asr.py b/ros/angel_system_nodes/angel_system_nodes/audio/asr.py index de09efb7..d172a97f 100644 --- a/ros/angel_system_nodes/angel_system_nodes/audio/asr.py +++ b/ros/angel_system_nodes/angel_system_nodes/audio/asr.py @@ -75,7 +75,7 @@ class ASR(dialogue.AbstractDialogueNode): self._feedback_topic = ( self.get_parameter(FEEDBACK_TOPIC).get_parameter_value().string_value ) - + self.log.info( f"Audio topic: " f"({type(self._audio_topic).__name__}) " @@ -226,11 +226,11 @@ class ASR(dialogue.AbstractDialogueNode): published_msg.utterance_text = text colored_utterance = colored(published_msg.utterance_text, "light_blue") self.log.info("Publishing message: " + f'"{colored_utterance}"') - + if ( - "angela" in text.lower() - or "angel" in text.lower() - or "angela," in text.lower() + "angela" in text.lower() + or "angel" in text.lower() + or "angela," in text.lower() or "angel," in text.lower() ): self.log.info("Publish thinking feedback") @@ -246,6 +246,7 @@ class ASR(dialogue.AbstractDialogueNode): publish_msg.response = "thinking" self._feedback_publisher.publish(publish_msg) + main = make_default_main(ASR) diff --git a/ros/angel_system_nodes/angel_system_nodes/audio/question_answerer.py b/ros/angel_system_nodes/angel_system_nodes/audio/question_answerer.py index 5a329d10..584f6f9d 100644 --- a/ros/angel_system_nodes/angel_system_nodes/audio/question_answerer.py +++ b/ros/angel_system_nodes/angel_system_nodes/audio/question_answerer.py @@ -33,6 +33,7 @@ CHAT_HISTORY_LENGTH = "chat_history_length" IMAGE_TOPIC = "image_topic" TASK_STATE_TOPIC = "task_state_topic" + class QuestionAnswerer(dialogue.AbstractDialogueNode): def __init__(self): super().__init__() @@ -152,7 +153,7 @@ class QuestionAnswerer(dialogue.AbstractDialogueNode): jpg_container = io.BytesIO() img_rgb.save(jpg_container, format="JPEG") self.image_msg = base64.b64encode(jpg_container.getvalue()).decode("utf-8") - + def question_answer_callback(self, msg): self.log.debug(f"Received message:\n\n{msg.utterance_text}") if not self._apply_filter(msg): @@ -169,7 +170,7 @@ class QuestionAnswerer(dialogue.AbstractDialogueNode): optional_fields = self._get_optional_fields_string( self.current_step, self.completed_steps ) - + response = self.get_response(msg, optional_fields) self.publish_generated_response(msg, response) @@ -207,9 +208,9 @@ class QuestionAnswerer(dialogue.AbstractDialogueNode): "model": model, "messages": [ { - "role": "user", + "role": "user", "content": [ - { + { "type": "text", "text": "Use the image to answer the question." + prompt, @@ -217,7 +218,7 @@ class QuestionAnswerer(dialogue.AbstractDialogueNode): { "type": "image_url", "image_url": { - "url": "data:image/jpeg;base64,"+self.image_msg + "url": "data:image/jpeg;base64," + self.image_msg }, }, ], @@ -225,8 +226,7 @@ class QuestionAnswerer(dialogue.AbstractDialogueNode): ], "temperature": 0.0, "max_tokens": 128, - } - + } req = requests.post( "https://api.openai.com/v1/chat/completions", @@ -306,9 +306,9 @@ class QuestionAnswerer(dialogue.AbstractDialogueNode): msg if it can be included. """ if ( - "angela" in msg.utterance_text.lower() - or "angel" in msg.utterance_text.lower() - or "angela," in msg.utterance_text.lower() + "angela" in msg.utterance_text.lower() + or "angel" in msg.utterance_text.lower() + or "angela," in msg.utterance_text.lower() or "angel," in msg.utterance_text.lower() ): return msg diff --git a/ros/angel_system_nodes/angel_system_nodes/task_monitoring/global_step_predictor.py b/ros/angel_system_nodes/angel_system_nodes/task_monitoring/global_step_predictor.py index 3808fcb0..9bcb6bd5 100644 --- a/ros/angel_system_nodes/angel_system_nodes/task_monitoring/global_step_predictor.py +++ b/ros/angel_system_nodes/angel_system_nodes/task_monitoring/global_step_predictor.py @@ -51,7 +51,6 @@ PARAM_GT_VIDEO_ID = "gt_video_id" PARAM_GT_OUTPUT_DIR = "gt_output_dir" # output directory override. - VALID_STEP_MODES = {"broad", "granular"}