aboutsummaryrefslogtreecommitdiff
path: root/src/argaze/utils/demo_aruco_markers_run.py
blob: ce81da41f972ecdd02f74667fd21371de1acb7fc (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
#!/usr/bin/env python

"""Augmented Reality pipeline demo script."""

__author__ = "Théo de la Hogue"
__credits__ = []
__copyright__ = "Copyright 2023, Ecole Nationale de l'Aviation Civile (ENAC)"
__license__ = "BSD"

import argparse
import contextlib
import os
import time

from argaze import ArFeatures, GazeFeatures
from argaze.ArUcoMarkers import ArUcoCamera
from argaze.utils import UtilsFeatures

import cv2
import numpy

def main():
    """
    Load ArUcoCamera from .json file, detect ArUco markers into camera device images and project it.
    """

    current_directory = os.path.dirname(os.path.abspath(__file__))

    # Manage arguments
    parser = argparse.ArgumentParser(description=main.__doc__.split('-')[0])
    parser.add_argument('aruco_camera', metavar='ARUCO_CAMERA', type=str, help='ArUco camera filepath')
    parser.add_argument('-s', '--source', metavar='SOURCE', type=str, default='0', help='video capture source (a number to select camera device or a filepath to load a movie)')
    args = parser.parse_args()

    # Load ArUcoCamera
    aruco_camera = ArUcoCamera.ArUcoCamera.from_json(args.aruco_camera)

    # Create a window to display ArUcoCamera
    cv2.namedWindow(aruco_camera.name, cv2.WINDOW_AUTOSIZE)

    # Init timestamp
    start_time = time.time()

    # Prepare gaze analysis assessment
    call_chrono = UtilsFeatures.TimeProbe()
    call_chrono.start()

    gaze_positions_frequency = 0
    gaze_analysis_time = 0

    # Fake gaze position with mouse pointer
    def on_mouse_event(event, x, y, flags, param):

        nonlocal gaze_positions_frequency
        nonlocal gaze_analysis_time

        # Assess gaze analysis
        lap_time, nb_laps, elapsed_time = call_chrono.lap()

        if elapsed_time > 1e3:

            gaze_positions_frequency = nb_laps
            call_chrono.restart()

        gaze_analysis_time = 0

        # Edit millisecond timestamp
        timestamp = int((time.time() - start_time) * 1e3)

        # Project gaze position into camera
        for frame, look_data in aruco_camera.look(timestamp, GazeFeatures.GazePosition((x, y))):

            # Unpack look data
            if look_data:

                gaze_position, gaze_movement, scan_step_analysis, layer_analysis, execution_times, exception = look_data

                # Assess gaze analysis
                gaze_analysis_time += execution_times['total']

    # Attach mouse callback to window
    cv2.setMouseCallback(aruco_camera.name, on_mouse_event)

    # Prepare video fps assessment
    video_fps = 0
    video_chrono = UtilsFeatures.TimeProbe()
    video_chrono.start()

    # Prepare visualisation time assessment
    visualisation_time = 0

    # Enable camera video capture into separate thread
    video_capture = cv2.VideoCapture(int(args.source) if args.source.isdecimal() else args.source)

    # Waiting for 'ctrl+C' interruption
    with contextlib.suppress(KeyboardInterrupt):

        # Capture images
        while video_capture.isOpened():

            # Assess capture time
            capture_start = time.time()

            # Read video image
            success, video_image = video_capture.read()

            # Assess capture time
            capture_time = int((time.time() - capture_start) * 1e3)

            if success:

                # Assess video fps
                lap_time, nb_laps, elapsed_time = video_chrono.lap()

                if elapsed_time > 1e3:

                    video_fps = nb_laps
                    video_chrono.restart()

                # Detect and project AR features
                detection_time, projection_time, exceptions = aruco_camera.watch(video_image)

                # Assess visualisation time
                visualisation_start = time.time()

                # Get ArUcoCamera frame image
                aruco_camera_image = aruco_camera.image()

                # Write time info
                cv2.rectangle(aruco_camera_image, (0, 0), (aruco_camera.size[0], 100), (63, 63, 63), -1)
                cv2.putText(aruco_camera_image, f'{video_fps} FPS | Capture {capture_time}ms | Detection {int(detection_time)}ms | Projection {int(projection_time)}ms | Visualisation {visualisation_time}ms', (20, 40), cv2.FONT_HERSHEY_SIMPLEX, 1, (255, 255, 255), 1, cv2.LINE_AA)
                cv2.putText(aruco_camera_image, f'{gaze_positions_frequency} gaze positions/s | Gaze analysis {gaze_analysis_time:.2f}ms', (20, 80), cv2.FONT_HERSHEY_SIMPLEX, 1, (255, 255, 255), 1, cv2.LINE_AA)

                # Handle exceptions
                for i, (scene_name, e) in enumerate(exceptions.items()):

                     # Write errors
                    cv2.rectangle(aruco_camera_image, (0, (i+1)*100), (aruco_camera.size[0], (i+2)*80), (127, 127, 127), -1)
                    cv2.putText(aruco_camera_image, f'{scene_name} error: {e}', (20, (i+1)*140), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 255, 255), 1, cv2.LINE_AA)

                # Write hint
                cv2.putText(aruco_camera_image, 'Mouve mouse pointer over gray rectangle area', (20, aruco_camera.size[1]-40), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 255, 255), 1, cv2.LINE_AA)

                # Display ArUcoCamera frame image
                cv2.imshow(aruco_camera.name, aruco_camera_image)

                # Draw and display each scene frames
                for scene_frame in aruco_camera.scene_frames:

                    # Display scene frame
                    cv2.imshow(f'{scene_frame.parent.name}:{scene_frame.name}', scene_frame.image())

            else:

                # Assess visualisation time
                visualisation_start = time.time()

            # Stop by pressing 'Esc' key
            # NOTE: on MacOS, cv2.waitKey(1) waits ~40ms
            if cv2.waitKey(1) == 27:

                # Close camera video capture
                video_capture.release()

            # Assess visualisation time
            visualisation_time = int((time.time() - visualisation_start) * 1e3)

    # Stop image display
    cv2.destroyAllWindows()

if __name__ == '__main__':

    main()