1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
|
#!/usr/bin/env python
"""Augmented Reality pipeline demo script."""
__author__ = "Théo de la Hogue"
__credits__ = []
__copyright__ = "Copyright 2023, Ecole Nationale de l'Aviation Civile (ENAC)"
__license__ = "BSD"
import argparse
import contextlib
import os
import time
from argaze import ArFeatures, GazeFeatures
from argaze.ArUcoMarkers import ArUcoCamera
from argaze.utils import UtilsFeatures
import cv2
import numpy
def main():
"""
Load ArUcoCamera from .json file, detect ArUco markers into camera device images and project it.
"""
current_directory = os.path.dirname(os.path.abspath(__file__))
# Manage arguments
parser = argparse.ArgumentParser(description=main.__doc__.split('-')[0])
parser.add_argument('aruco_camera', metavar='ARUCO_CAMERA', type=str, help='ArUco camera filepath')
parser.add_argument('-s', '--source', metavar='SOURCE', type=str, default='0', help='video capture source (a number to select camera device or a filepath to load a movie)')
args = parser.parse_args()
# Load ArUcoCamera
aruco_camera = ArUcoCamera.ArUcoCamera.from_json(args.aruco_camera)
# Create a window to display ArUcoCamera
cv2.namedWindow(aruco_camera.name, cv2.WINDOW_AUTOSIZE)
# Init timestamp
start_time = time.time()
# Prepare gaze analysis assessment
call_chrono = UtilsFeatures.TimeProbe()
call_chrono.start()
gaze_positions_frequency = 0
gaze_analysis_time = 0
# Fake gaze position with mouse pointer
def on_mouse_event(event, x, y, flags, param):
nonlocal gaze_positions_frequency
nonlocal gaze_analysis_time
# Assess gaze analysis
lap_time, nb_laps, elapsed_time = call_chrono.lap()
if elapsed_time > 1e3:
gaze_positions_frequency = nb_laps
call_chrono.restart()
gaze_analysis_time = 0
# Edit millisecond timestamp
timestamp = int((time.time() - start_time) * 1e3)
# Project gaze position into camera
for frame, look_data in aruco_camera.look(timestamp, GazeFeatures.GazePosition((x, y))):
# Unpack look data
if look_data:
gaze_position, gaze_movement, scan_step_analysis, layer_analysis, execution_times, exception = look_data
# Assess gaze analysis
gaze_analysis_time += execution_times['total']
# Attach mouse callback to window
cv2.setMouseCallback(aruco_camera.name, on_mouse_event)
# Prepare video fps assessment
video_fps = 0
video_chrono = UtilsFeatures.TimeProbe()
video_chrono.start()
# Prepare visualisation time assessment
visualisation_time = 0
# Enable camera video capture into separate thread
video_capture = cv2.VideoCapture(int(args.source) if args.source.isdecimal() else args.source)
# Waiting for 'ctrl+C' interruption
with contextlib.suppress(KeyboardInterrupt):
# Capture images
while video_capture.isOpened():
# Assess capture time
capture_start = time.time()
# Read video image
success, video_image = video_capture.read()
# Assess capture time
capture_time = int((time.time() - capture_start) * 1e3)
if success:
# Assess video fps
lap_time, nb_laps, elapsed_time = video_chrono.lap()
if elapsed_time > 1e3:
video_fps = nb_laps
video_chrono.restart()
# Detect and project AR features
detection_time, projection_time, exceptions = aruco_camera.watch(video_image)
# Assess visualisation time
visualisation_start = time.time()
# Get ArUcoCamera frame image
aruco_camera_image = aruco_camera.image()
# Write time info
cv2.rectangle(aruco_camera_image, (0, 0), (aruco_camera.size[0], 100), (63, 63, 63), -1)
cv2.putText(aruco_camera_image, f'{video_fps} FPS | Capture {capture_time}ms | Detection {int(detection_time)}ms | Projection {int(projection_time)}ms | Visualisation {visualisation_time}ms', (20, 40), cv2.FONT_HERSHEY_SIMPLEX, 1, (255, 255, 255), 1, cv2.LINE_AA)
cv2.putText(aruco_camera_image, f'{gaze_positions_frequency} gaze positions/s | Gaze analysis {gaze_analysis_time:.2f}ms', (20, 80), cv2.FONT_HERSHEY_SIMPLEX, 1, (255, 255, 255), 1, cv2.LINE_AA)
# Handle exceptions
for i, (scene_name, e) in enumerate(exceptions.items()):
# Write errors
cv2.rectangle(aruco_camera_image, (0, (i+1)*100), (aruco_camera.size[0], (i+2)*80), (127, 127, 127), -1)
cv2.putText(aruco_camera_image, f'{scene_name} error: {e}', (20, (i+1)*140), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 255, 255), 1, cv2.LINE_AA)
# Write hint
cv2.putText(aruco_camera_image, 'Mouve mouse pointer over gray rectangle area', (20, aruco_camera.size[1]-40), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 255, 255), 1, cv2.LINE_AA)
# Display ArUcoCamera frame image
cv2.imshow(aruco_camera.name, aruco_camera_image)
# Draw and display each scene frames
for scene_frame in aruco_camera.scene_frames:
# Display scene frame
cv2.imshow(f'{scene_frame.parent.name}:{scene_frame.name}', scene_frame.image())
else:
# Assess visualisation time
visualisation_start = time.time()
# Stop by pressing 'Esc' key
# NOTE: on MacOS, cv2.waitKey(1) waits ~40ms
if cv2.waitKey(1) == 27:
# Close camera video capture
video_capture.release()
# Assess visualisation time
visualisation_time = int((time.time() - visualisation_start) * 1e3)
# Stop image display
cv2.destroyAllWindows()
if __name__ == '__main__':
main()
|