SMART Lab, Purdue University
Teleoperation systems have advanced in recent years, offering many benefits to remote operators. However, they often overlook the affective loads that human operators experience and are related to the system performance directly and indirectly. Researchers attempt to understand the affective loads using stimuli from psychology, but these do not adequately generate affective loads in teleoperation scenarios. To address this gap, a new stimulus tool has been introduced, called SMART-TeleLoad. It is an open-source-based graphic user interface (GUI) program that can modulate various control variables relating to the stimulus as needed for research and practical applications. It also supports a lab-streaming layer (LSL) to connect with other systems like physiological sensors and other GUI programs, independent of operating systems. Furthermore, we validated the performance of the proposed stimulus tool by analyzing subjective questionnaires with aspects of the cognitive and emotional loads through an extensive user study involving 30 participants.
There is a demonstration video available at https://youtu.be/qaYlWQdBHfI, or click below image.
This repository is composed of two main folders:
doc
: paper and figures for this gitsrc
: main stimulus GUI programs and resources as below folder tree:
-
src/darknet_files
: including the essential cfg file and split weights zip files necessary for object detection algorithms. -
src/pyqt_ui_files
: including six UI design files used in the SMART-TeleLoad program -
src/resource
: having three folders including supplementary files, such as images, effect sounds, and pre-recorded video files. -
src/subjective_results
: saving directory for the human subject's answers and mission scores from the subjective questionnaire -
src/control_room_gui_node.py
: -
src/lsl_outlet_reader.py
: an example code to read the LSL stream data, which is an optional code for user to check the LSL outlet stream data. -
src/lsl_stream_setup.py
: defining the LSL out streams which is connected to
SMART-TeleLoad was inititaly developed in Python 3.8 under a Linux environment (e.g., Ubuntu 20.04), so it may not work on any other operating systems. There is an instructional video available at https://youtu.be/qaYlWQdBHfI that demonstrates how to run codes and navigate the SMART-TeleLoad.
Before starting this program, your machine should have enabled cuDNN on OpenCV and below python libaries to run the main GUI program:
- cv2
- pandas
- playsound
- pylsl
- PyQt5
After installing the libaries, please
cd src
#to go to thesrc
folderpython control_room_gui_node.py
#run main program
There are four outlet streams to transmit the subject's information
(1) teleload_mouse_pos
: It is to track the mouse cursor positions (x and y positions).
(2) teleload_mouse_btn
: It is to recognize mouse button clicks (e.g., pressed or released) within the window of the GUI.
(3) teleload_task_accuracy
: It is to measure the task accuracy based on the number of success and failure clicks.
(4) teleload_exp_status
: It is to track the experiment status, such as the experiment start, preparation, main task, and end.
The human subject's answers on the subjective questionnaires (e.g, SAM, NATA-TLX) and mission scores are saved to src/subjective_results
as a CSV file named this format:
[participant number]_cam_[Camera number]_speed_[Moving speed].csv
.
This material is based upon work supported by the National Science Foundation under Grant No. IIS-1846221. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation.
Please use the following citation:
Wonse Jo, Go-Eum Cha, Dan Foti, Byung-Cheol Min,. ''SMART-TeleLoad: A New Graphic User Interface to Generate Affective Loads for Teleoperation''. SoftwareX, Volume 26, 2024, 101757, ISSN 2352-7110, https://doi.org/10.1016/j.softx.2024.101757.
@article{jo_smart-teleload_2024,
title = {{SMART}-{TeleLoad}: {A} new graphic user interface to generate affective loads for teleoperation},
volume = {26},
issn = {2352-7110},
url = {https://www.sciencedirect.com/science/article/pii/S2352711024001286},
doi = {https://doi.org/10.1016/j.softx.2024.101757},
journal = {SoftwareX},
author = {Jo, Wonse and Cha, Go-Eum and Foti, Dan and Min, Byung-Cheol},
year = {2024},
keywords = {Affective loads, Human–robot interaction (HRI), Stimulus tool},
pages = {101757},
}