From 5d7772e823fe66fefe9455848bce0389f8b8ab56 Mon Sep 17 00:00:00 2001 From: Théo de la Hogue Date: Wed, 27 Mar 2024 09:26:03 +0100 Subject: Defining new ArContext class. --- docs/user_guide/utils/demonstrations_scripts.md | 26 +++++++++++++++++-------- docs/user_guide/utils/ready-made_scripts.md | 8 ++++++++ 2 files changed, 26 insertions(+), 8 deletions(-) (limited to 'docs') diff --git a/docs/user_guide/utils/demonstrations_scripts.md b/docs/user_guide/utils/demonstrations_scripts.md index a911173..c3a5c9b 100644 --- a/docs/user_guide/utils/demonstrations_scripts.md +++ b/docs/user_guide/utils/demonstrations_scripts.md @@ -14,7 +14,7 @@ Collection of command-line scripts for demonstration purpose. Load ArFrame with a single ArLayer from **demo_gaze_analysis_setup.json** file then, simulate gaze position using mouse pointer to illustrate gaze features. ```shell -python ./src/argaze/utils/demo_gaze_analysis_run.py ./src/argaze/utils/demo_data/demo_gaze_analysis_setup.json +python ./src/argaze/utils/demo_gaze_analysis_run.py ./src/argaze/utils/demo/gaze_analysis_pipeline.json ``` ## ArUco markers pipeline demonstration @@ -22,7 +22,7 @@ python ./src/argaze/utils/demo_gaze_analysis_run.py ./src/argaze/utils/demo_data Load ArUcoCamera from **demo_aruco_markers_setup.json** file then, detect ArUco markers into a demo video source and estimate camera pose. ```shell -python ./src/argaze/utils/demo_aruco_markers_run.py ./src/argaze/utils/demo_data/demo_aruco_markers_setup.json -s ./src/argaze/utils/demo_data/demo.mov +python ./src/argaze/utils/demo_aruco_markers_run.py ./src/argaze/utils/demo/aruco_markers_pipeline.json -s ./src/argaze/utils/demo_data/demo.mov ``` !!! note @@ -34,18 +34,28 @@ Load ArUcoCamera from a configuration file then, stream and process gaze positio ### With Tobii Pro Glasses 2 device -To use a Tobii Pro Glasses 2 device, you need to edit **provider_setup.json** file as below with your own parameters values: +To use a Tobii Pro Glasses 2 device, you need to edit **eyetracker_setup.json** file as below with your own parameters values: -```shell +```json { - "TobiiProGlasses2" : { - "address": "10.34.0.17", + "argaze.utils.eyetrackers.TobiiProGlasses2.LiveStream" : { + "address": "10.34.0.12", "project": "MyProject", - "participant": "NewParticipant" + "participant": "NewParticipant", + "configuration": { + "sys_ec_preset": "Indoor", + "sys_sc_width": 1920, + "sys_sc_height": 1080, + "sys_sc_fps": 25, + "sys_sc_preset": "Auto", + "sys_et_freq": 50, + "sys_mems_freq": 100 + }, + "pipeline": "demo_aruco_markers_setup.json" } } ``` ```shell -python ./src/argaze/utils/worn_device_stream.py ./src/argaze/utils/demo_data/demo_aruco_markers_setup.json +python ./src/argaze/utils/pipeline_run.py ./src/argaze/utils/demo/eyetracker_setup.json ``` \ No newline at end of file diff --git a/docs/user_guide/utils/ready-made_scripts.md b/docs/user_guide/utils/ready-made_scripts.md index 9b8d455..262a0ef 100644 --- a/docs/user_guide/utils/ready-made_scripts.md +++ b/docs/user_guide/utils/ready-made_scripts.md @@ -9,6 +9,14 @@ Collection of command-line scripts to provide useful features. !!! note *Use -h option to get command arguments documentation.* +## Eyetracker pipeline handler + +Load and execute eyetracker pipeline. + +```shell +python ./src/argaze/utils/pipeline_run.py CONFIGURATION +``` + ## ArUco markers group exporter Detect DICTIONARY and SIZE ArUco markers inside a MOVIE frame then, export detected ArUco markers group as .obj file into an OUTPUT folder. -- cgit v1.1