aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorThéo de la Hogue2023-07-04 14:03:51 +0200
committerThéo de la Hogue2023-07-04 14:03:51 +0200
commitf29d0ce4a84b8f13299fbcfadf09702e2f7a3152 (patch)
tree2737143e242ebe2b2a23c6055f6c4bde73cb9b14
parentebc7a6274a4788bd32802447f98dd39df78e9dfd (diff)
downloadargaze-f29d0ce4a84b8f13299fbcfadf09702e2f7a3152.zip
argaze-f29d0ce4a84b8f13299fbcfadf09702e2f7a3152.tar.gz
argaze-f29d0ce4a84b8f13299fbcfadf09702e2f7a3152.tar.bz2
argaze-f29d0ce4a84b8f13299fbcfadf09702e2f7a3152.tar.xz
Refactoring look and draw methods into ArFeatures.
-rw-r--r--src/argaze/ArFeatures.py166
1 files changed, 67 insertions, 99 deletions
diff --git a/src/argaze/ArFeatures.py b/src/argaze/ArFeatures.py
index 62ce4d8..f983a69 100644
--- a/src/argaze/ArFeatures.py
+++ b/src/argaze/ArFeatures.py
@@ -374,7 +374,7 @@ class ArEnvironment():
# For each aoi scene projection
for scene_name, scene in self.scenes.items():
- scene.look(timestamp, gaze_position)
+ yield scene_name, scene.look(timestamp, gaze_position)
def to_json(self, json_filepath):
"""Save environment to .json file."""
@@ -446,6 +446,12 @@ class ArFrame():
# Init current gaze position
self.__gaze_position = GazeFeatures.UnvalidGazePosition()
+ # Init current gaze movement
+ self.__gaze_movement = GazeFeatures.UnvalidGazeMovement()
+
+ # Init current look at aoi
+ self.__look_at = self.name
+
# Init heatmap if required
if self.heatmap:
@@ -478,78 +484,25 @@ class ArFrame():
)
@property
- def current_gaze_position(self):
- """Get current gaze position on frame."""
-
- # Wait for frame to be unlocked
- while self.__looking_lock.locked():
- pass
-
- return self.__gaze_position
-
- @property
- def current_gaze_movement(self):
- """Get current gaze movement on frame."""
-
- # Wait for frame to be unlocked
- while self.__looking_lock.locked():
- pass
-
- # Check current frame fixation
- current_fixation = self.gaze_movement_identifier.current_fixation
-
- if current_fixation.valid:
-
- return current_fixation
-
- # Check current frame saccade
- current_saccade = self.gaze_movement_identifier.current_saccade
-
- if current_saccade.valid:
-
- return current_saccade
-
- return GazeFeatures.UnvalidGazeMovement()
-
- @property
- def current_scan_path_analysis(self) -> dict:
- """Get current scan path analysis."""
-
- # Edit dictionary with all analysis
- scan_path_analysis = {}
-
- if scan_path != None:
-
- # Wait for frame to be unlocked
- while self.__looking_lock.locked():
- pass
-
- # For each scan path analyzer
- for scan_path_analyzer_type, scan_path_analyzer in self.scan_path_analyzers.items():
-
- scan_path_analysis[scan_path_analyzer_type] = scan_path_analyzer.analysis
-
- return scan_path_analysis
-
- @property
- def current_aoi_scan_path_analysis(self) -> dict:
- """Get current aoi scan path analysis."""
+ def image(self):
+ """
+ Get background image + heatmap image
+ """
- # Edit dictionary with all analysis
- aoi_scan_path_analysis = {}
+ # Lock frame exploitation
+ self.__looking_lock.acquire()
- if aoi_scan_path != None:
+ image = self.background.copy()
- # Wait for frame to be unlocked
- while self.__looking_lock.locked():
- pass
+ # Draw heatmap
+ if self.heatmap:
- # For each aoi scan path analyzer
- for aoi_scan_path_analyzer_type, aoi_scan_path_analyzer in self.aoi_scan_path_analyzers.items():
+ image = cv2.addWeighted(self.heatmap.image, 0.5, image, 1., 0)
- aoi_scan_path_analysis[aoi_scan_path_analyzer_type] = aoi_scan_path_analyzer.analysis
+ # Unlock frame exploitation
+ self.__looking_lock.release()
- return aoi_scan_path_analysis
+ return image
def look(self, timestamp: int|float, inner_gaze_position: GazeFeatures.GazePosition) -> Tuple[GazeFeatures.GazeMovement, str, dict, dict]:
"""
@@ -565,12 +518,10 @@ class ArFrame():
# Lock frame exploitation
self.__looking_lock.acquire()
- # Update internal gaze position
+ # Update current gaze position
self.__gaze_position = inner_gaze_position
- # Prepare looking data
- gaze_movement = GazeFeatures.UnvalidGazeMovement()
- look_at = self.name
+ # Init scan path analysis report
scan_step_analysis = {}
aoi_scan_step_analysis = {}
@@ -578,31 +529,35 @@ class ArFrame():
if self.gaze_movement_identifier:
# Identify gaze movement
- gaze_movement = self.gaze_movement_identifier.identify(timestamp, self.__gaze_position)
+ new_gaze_movement = self.gaze_movement_identifier.identify(timestamp, self.__gaze_position)
- if GazeFeatures.is_fixation(gaze_movement):
+ if GazeFeatures.is_fixation(new_gaze_movement):
+
+ # Update current gaze movement
+ self.__gaze_movement = new_gaze_movement
# Does the fixation match an AOI?
for name, aoi in self.aoi_2d_scene.items():
- _, _, circle_ratio = aoi.circle_intersection(gaze_movement.focus, gaze_movement.deviation_max)
+ _, _, circle_ratio = aoi.circle_intersection(self.__gaze_movement.focus, self.__gaze_movement.deviation_max)
if circle_ratio > 0.25:
if name != self.name:
- look_at = name
+ # Update current lookt at
+ self.__look_at = name
break
# Append fixation to scan path
if self.scan_path != None:
- self.scan_path.append_fixation(timestamp, gaze_movement)
+ self.scan_path.append_fixation(timestamp, self.__gaze_movement)
# Append fixation to aoi scan path
if self.aoi_scan_path != None:
- aoi_scan_step = self.aoi_scan_path.append_fixation(timestamp, gaze_movement, look_at)
+ aoi_scan_step = self.aoi_scan_path.append_fixation(timestamp, self.__gaze_movement, self.__look_at)
# Analyze aoi scan path
if aoi_scan_step and len(self.aoi_scan_path) > 1:
@@ -613,12 +568,15 @@ class ArFrame():
aoi_scan_step_analysis[aoi_scan_path_analyzer_type] = aoi_scan_path_analyzer.analysis
- elif GazeFeatures.is_saccade(gaze_movement):
+ elif GazeFeatures.is_saccade(new_gaze_movement):
+
+ # Update current gaze movement
+ self.__gaze_movement = new_gaze_movement
# Append saccade to scan path
if self.scan_path != None:
- scan_step = self.scan_path.append_saccade(timestamp, gaze_movement)
+ scan_step = self.scan_path.append_saccade(timestamp, self.__gaze_movement)
# Analyze aoi scan path
if scan_step and len(self.scan_path) > 1:
@@ -632,7 +590,7 @@ class ArFrame():
# Append saccade to aoi scan path
if self.aoi_scan_path != None:
- self.aoi_scan_path.append_saccade(timestamp, gaze_movement)
+ self.aoi_scan_path.append_saccade(timestamp, self.__gaze_movement)
# Update heatmap
if self.heatmap:
@@ -643,29 +601,39 @@ class ArFrame():
self.__looking_lock.release()
# Return looking data
- return gaze_movement, look_at, scan_step_analysis, aoi_scan_step_analysis
+ return self.__gaze_movement, self.__look_at, scan_step_analysis, aoi_scan_step_analysis
- @property
- def image(self,):
+ def draw(self, image:numpy.array):
"""
- Get frame image
+ Draw frame into image.
+
+ Parameters:
+ image: where to draw
"""
- image = self.background.copy()
-
- self.aoi_2d_scene.draw(image, color=(255, 255, 255))
- self.current_gaze_position.draw(image, color=(255, 255, 255))
+ # Lock frame exploitation
+ self.__looking_lock.acquire()
- self.current_gaze_movement.draw(image, color=(0, 255, 255))
- self.current_gaze_movement.draw_positions(image)
+ # Draw aoi
+ self.aoi_2d_scene.draw(image, color=(0, 0, 0))
- # Check frame fixation
- if GazeFeatures.is_fixation(self.current_gaze_movement):
+ # Draw gaze position
+ self.__gaze_position.draw(image, color=(255, 255, 255))
- # Draw looked AOI
- self.aoi_2d_scene.draw_circlecast(image, self.current_gaze_movement.focus, self.current_gaze_movement.deviation_max, base_color=(0, 0, 0), matching_color=(255, 255, 255))
+ # Draw gaze movement
+ if gaze_movement_identifier:
- return image
+ self.__gaze_movement.draw(image, color=(0, 255, 255))
+ self.__gaze_movement.draw_positions(image)
+
+ # Check fixation case
+ if GazeFeatures.is_fixation(self.__gaze_movement):
+
+ # Draw looked AOI
+ self.aoi_2d_scene.draw_circlecast(image, self.__gaze_movement.focus, self.__gaze_movement.deviation_max, base_color=(0, 0, 0), matching_color=(255, 255, 255))
+
+ # Unlock frame exploitation
+ self.__looking_lock.release()
@dataclass
class ArScene():
@@ -901,7 +869,7 @@ class ArScene():
self.__camera_frame_lock.acquire()
# Project gaze position in camera frame
- yield frame.look(timestamp, inner_gaze_position * frame.size)
+ yield None, self.camera_frame.look(timestamp, gaze_position)
# Project gaze position into each aoi frames if possible
for aoi_name, frame in self.aoi_frames.items():
@@ -919,7 +887,7 @@ class ArScene():
# QUESTION: How to project gaze precision?
inner_gaze_position = GazeFeatures.GazePosition((inner_x, inner_y))
- yield frame.look(timestamp, inner_gaze_position * frame.size)
+ yield aoi_name, frame.look(timestamp, inner_gaze_position * frame.size)
# Ignore missing aoi frame projection
except KeyError:
@@ -931,13 +899,13 @@ class ArScene():
def draw(self, image: numpy.array):
"""
- Draw camera frame
+ Draw camera frame into image.
Parameters:
image: where to draw
"""
- self.camera_frame.aoi_2d_scene.draw(image)
+ self.camera_frame.draw(image)
def draw_axis(self, image: numpy.array):
"""