diff options
author | Théo de la Hogue | 2024-06-24 18:06:54 +0200 |
---|---|---|
committer | Théo de la Hogue | 2024-06-24 18:06:54 +0200 |
commit | a9f5d765182f8851b0dab6768aafb52809a58503 (patch) | |
tree | a63ea6746a4c8d075c88efdbd48159d0d7975ea5 /src | |
parent | 3c3890e8ec44b32c0df112b5c1ba6ada97fc4c13 (diff) | |
download | argaze-a9f5d765182f8851b0dab6768aafb52809a58503.zip argaze-a9f5d765182f8851b0dab6768aafb52809a58503.tar.gz argaze-a9f5d765182f8851b0dab6768aafb52809a58503.tar.bz2 argaze-a9f5d765182f8851b0dab6768aafb52809a58503.tar.xz |
Adding key interaction to draw help and start recording.
Diffstat (limited to 'src')
-rw-r--r-- | src/argaze/__main__.py | 96 |
1 files changed, 75 insertions, 21 deletions
diff --git a/src/argaze/__main__.py b/src/argaze/__main__.py index df3d338..2647198 100644 --- a/src/argaze/__main__.py +++ b/src/argaze/__main__.py @@ -52,22 +52,51 @@ def load_context(args): logging.info('%s pipe opened', args.pipe_path) - def display(name, image, factor): + def display(name, image, factor = 0.75, draw_help = False): """Adapt image to display dimension.""" + height, width, _ = image.shape + + if draw_help: + + cv2.rectangle(image, (int(width/4), int(height/3)), (int(width*3/4), int(height*2/3)), (127, 127, 127), -1) + + info_stack = 1 + cv2.putText(image, f'Help', (int(width/4)+20, int(height/3)+(info_stack*40)), cv2.FONT_HERSHEY_SIMPLEX, 1, (255, 255, 255), 1, cv2.LINE_AA) + + if issubclass(type(context), LiveProcessingContext): + + info_stack += 1 + cv2.putText(image, f'Press Enter to start calibration', (int(width/4)+20, int(height/3)+(info_stack*40)), cv2.FONT_HERSHEY_SIMPLEX, 1, (255, 255, 255), 1, cv2.LINE_AA) + + info_stack += 1 + cv2.putText(image, f'Press r to start/stop recording', (int(width/4)+20, int(height/3)+(info_stack*40)), cv2.FONT_HERSHEY_SIMPLEX, 1, (255, 255, 255), 1, cv2.LINE_AA) + + if issubclass(type(context), PostProcessingContext): + + info_stack += 1 + cv2.putText(image, f'Press Space bar to pause/resume processing', (int(width/3)+20, int(height/3)+(info_stack*40)), cv2.FONT_HERSHEY_SIMPLEX, 1, (255, 255, 255), 1, cv2.LINE_AA) + + info_stack += 1 + cv2.putText(image, f'Press f to pause/resume visualisation', (int(width/4)+20, int(height/3)+(info_stack*40)), cv2.FONT_HERSHEY_SIMPLEX, 1, (255, 255, 255), 1, cv2.LINE_AA) + + info_stack += 1 + cv2.putText(image, f'Press h to hide/show this help panel', (int(width/4)+20, int(height/3)+(info_stack*40)), cv2.FONT_HERSHEY_SIMPLEX, 1, (255, 255, 255), 1, cv2.LINE_AA) + + info_stack += 1 + cv2.putText(image, f'Press Escape to quit', (int(width/4)+20, int(height/3)+(info_stack*40)), cv2.FONT_HERSHEY_SIMPLEX, 1, (255, 255, 255), 1, cv2.LINE_AA) + if args.display is not None: display_size = tuple(args.display) - height, width, _ = image.shape + image_ratio = width/height new_image_size = (int(display_size[1] * factor * image_ratio), int(display_size[1] * factor)) - cv2.imshow(name, cv2.resize(image, dsize=new_image_size, interpolation=cv2.INTER_LINEAR)) - - else: + image = cv2.resize(image, dsize=new_image_size, interpolation=cv2.INTER_LINEAR) - cv2.imshow(name, image) + cv2.imshow(name, image) # Load context from JSON file with load(args.context_file) as context: @@ -93,8 +122,9 @@ def load_context(args): # Assess processing time start_time = time.time() - # Draw context pipeline by default + # Draw parameters draw_pipeline = True + draw_help = True # Waiting for 'ctrl+C' interruption with contextlib.suppress(KeyboardInterrupt), os.fdopen(pipe_file) if args.pipe_path is not None else contextlib.nullcontext() as pipe: @@ -125,7 +155,7 @@ def load_context(args): try: # Display context - display(context.name, context.image(draw_pipeline=draw_pipeline), 0.75) + display(context.name, context.image(draw_pipeline=draw_pipeline), 0.75, draw_help=draw_help) except SharedObjectBusy: @@ -149,34 +179,58 @@ def load_context(args): key_pressed = cv2.waitKey(40) #print("key_pressed", key_pressed) - # Esc: close window - if key_pressed == 27: + # Enter: start calibration + if key_pressed == 13: - raise KeyboardInterrupt() + if issubclass(type(context), LiveProcessingContext): + + context.calibrate() # Space bar: pause/resume pipeline processing if key_pressed == 32: - if context.is_paused(): + if issubclass(type(context), PostProcessingContext): + + if context.is_paused(): + + context.resume() + + else: + + context.pause() + + # r: start/stop recording + if key_pressed == 114: + + if issubclass(type(context), LiveProcessingContext): - context.resume() + # FIXME: the following commands only work with TobiiGlassesPro2.LiveStream context. + recording_status = context.get_recording_status() - else: + if recording_status == 'recording': - context.pause() + context.stop_recording() + + else: + + context.create_recording() + context.start_recording() # f: disable/enable pipeline drawing if key_pressed == 102: draw_pipeline = not draw_pipeline - # Enter: start calibration - if key_pressed == 13: + # h: disable/enable help drawing + if key_pressed == 104: - if issubclass(type(context), LiveProcessingContext): + draw_help = not draw_help - context.calibrate() + # Esc: close window + if key_pressed == 27: + raise KeyboardInterrupt() + # Window mode off else: @@ -185,12 +239,12 @@ def load_context(args): prefix = f'Progression' suffix = f'| {int(context.progression*context.duration * 1e-3)}s in {int(time.time()-start_time)}s' - look_time, look_freq = context.process_gaze_position_performance() + look_time, look_freq = context.pipeline.execution_info('look') suffix += f' | Look {look_time:.2f}ms at {look_freq}Hz' if issubclass(type(context.pipeline), ArCamera): - watch_time, watch_freq = context.process_camera_image_performance() + watch_time, watch_freq = context.pipeline.execution_info('watch') suffix += f' | Watch {int(watch_time)}ms at {watch_freq}Hz' # Clear old longer print |