Unlock Fear

Unlock the Secret of Acrophobia

Overview

“Unlock the secret of fear”

Acrophobia, the fear of heights, significantly affects daily activities, career choices, and overall quality of life. People who suffer from acrophobia know this fear is excessive and unreasonable, yet 6.4% of total population suffer from this issue, according to DSM-5.

We investigate the feasibility of passive sensing data for fear measurement and discern the most influential physiological and behavioral indicators. This technology may fill the current gap for VR therapy, which shows promise in exposure therapy but requires ongoing monitoring and adjustments.  

Role

Co-first author (of HCI research paper)

Duration

2022 July – 2023 Sept. (1 year 2 months) 

Director

Yuntao Wang, Tsinghua PI HCI Lab

Category

Affective Computing, VR, Virtual Therapy

Status

Currently under review, IMWUT 24′

Acrophobia, or fear of heights, is a prevalent mental health condition that can profoundly impact daily activities, career choices, and overall quality of life. While virtual reality (VR) offers promising exposure therapy interventions, many rely on self-reported fear levels, demanding consistent monitoring and adjustments. This paper explores the potential of physiological and behavioral data to gauge real-time fear levels among acrophobia sufferers. We investigate the feasibility of passive sensing data for fear measurement and discern the most influential physiological and behavioral indicators. Through a VR study involving 25 participants, we collected physiological signals, gaze patterns, and subjective fear ratings. Our methodology attained an RMSE of 1.91 and MAE of 1.7 on a 10-point fear intensity scale, and we pinpointed significant factors affecting acrophobia severity measurements. Ultimately, our findings offer insights into more tailored and effective therapeutic strategies for acrophobia, enhancing the quality of life and holistic well-being of affected individuals.

Present methods show a good performance in the classification prediction task of acrophobia in a coarse-grained manner (2-4 classes). This implies the potential to further develop a fine-grained fear prediction model, which is desired by therapists and wasn’t mentioned in current studies.

 Meanwhile, current prediction models mainly focus on internal physiological signals (EEG, EDA, HRV, etc.), which may be supplented and improved by adding in eye data, including pupil dilation and eye openess data.

Figure.1 shows the process of the data collection. During the data collection, we utilised two types of hardware devices: BIOPAC set and Vive Pro Eye. For BIOPAC set, we installed one BIOPAC station and attached two terminals to the participants (1 on the wrist, 1 on the chest). The terminal on the chest was used for collecting ECG data, while the other was used for recording EDA and PPG signals.  Additionally, participants were asked to wear the Vive Pro Eye headset. Throughout the data collection, participants were instructed to wear VR headsets and keep the BIOPAC terminal attached to their bodies.

Preparation   Before data collection, we synced wristbands and VR sets using a laptop. Participants donned BIOPAC devices, starting with a one-minute calm baseline. We recorded daily fear of heights and performance feedback. Participants experienced a realistic spatial perception in virtual reality with an elevator cabin and outside platform simulation.

Height Simulation   Participants, after VR setup and eye calibration, experienced a height simulation. In a virtual elevator, they moved to random heights, exiting onto a platform and rating fear on a 1-10 scale. Elevator doors closed for 15 seconds, transitioning to another height (0m to 100m). Chosen to balance data, these heights align with heightened physiological and emotional responses near ground level and saturation around 100 meters. The experimenter observed fear-related behaviors, like trembling or altered stride length, during height exposure outside the elevator. This comprehensive approach aimed to capture nuanced reactions to simulated heights.

Session Repetition   After returning to the elevator, it moved for 15 seconds before stopping at a new height. Participants repeated the platform-walking and fear-rating procedure. Three sessions, each on a different virtual site, comprised 10 heights each. Rests between sessions were allowed.

Data Overview   Post-data collection, participants removed the device, and VR position recalibration occurred. Each participant produced 30 sets of data, encompassing 30-second records of 4 signal channels and behavioral traits. Entries, paired with self-reported fear scores (1-10), totaled 450 instances, forming a comprehensive dataset for analysis. 

We recruited 46 participants from a local university using diverse social media advertisements. They completed the Heights Interpretation Questionnaire (HIQ) to assess height-related distress and avoidance. Those with HIQ scores ≤45 were excluded. Eleven participants dropped out, resulting in a final sample of 13 (5 females, 8 males, average age = 22.1, STD = 2.25).

Physiological and Activity Data   In the experiment, participants wore the BioNomadix Wireless Transmitter and Vive Pro Eye. The BioNomadix has electrodermal activity (EDA), photoplethysmography (PPG), and Electrocardiograph (ECG) sensors. EDA captures skin conductivity changes, PPG measures blood volume pulse, and Vive Pro Eye records pupil size and eye openness at 90Hz.

 

Ground Truth   Self-report Fear Level Data : This study employed self-report evaluations to gauge participants’ subjective fear responses to simulated high-altitude environments. Participants rated their fear level on a scale of 1 to 10 after each exposure, with 10 indicating the highest level of acrophobia.

Data Cleaning   During data cleaning, anomalies in Biopac sensors (EDA, ECG, PPG) and Vive Pro Eye pupil-related data are addressed. Flat responses over 5 seconds, mainly due to poor electrode-skin contact, led to excluding data from four participants in building the regression model. Nonetheless, their data contribute to discussing modality impacts. Data Pre-processing   Our data pre-processing targets noise and inter-subject variability to enhance our cross-user prediction model. For ECG signals, the ecga_clean function from the Neuralkit2 library is applied, incorporating a 0.5 Hz high-pass Butterworth filter (order = 5) and powerline filtering. In handling EDA signals, noticeable artifacts from body movements necessitate a 1 Hz low-pass Butterworth filter (order = 5) to eliminate abrupt negative peaks. Subsequently, z-score normalization is applied to scale EDA data, and the cvxEDA method decomposes signals into phasic and tonic components. This decomposition aids in capturing stable arousal levels (phasic) and momentary emotional responses (tonic).

Feature Extraction   Features are derived from original data, using existing signal processing algorithm.

Traditional fear of height prediction employs classification or clustering, but in Virtual Reality Exposure Therapy (VRET), accurately measuring fear intensity is crucial. A regression-based approach, predicting continuous values, offers precise estimates, enabling personalized and timely feedback. This fine-grained fear prediction enhances VRET effectiveness in addressing acrophobia.

Ground Truth    Self-reported fear of height scores (1-10) were adopted as ground truth, using a Verbal Rating Scale (VRS) for recording. VRS, deemed more suitable for its time efficiency and intuitiveness, was preferred over the previously used SAM in our task.

Regressors    Our research develops regression models for precise fear prediction in height-related scenarios. Considering data from three modalities with numerous features, including correlated and potentially insignificant ones, demands a robust model for high-dimensional, non-linear situations. We employ three well-established regression models: Random Forest, Support Vector Regression (SVR), and LightGBM Regression. These models, especially ensemble ones like RF and LightGBM, enhance accuracy and reliability, crucial for predicting fear levels associated with heights.

Validation    To optimize our models, we employ cross-validation for both cross-user and individual-based models. Using leave-one-out cross-validation on our small dataset (390 points, 30 per individual), we average metrics (MAE, RMSE) over subjects. Grid search in the sklearn library refines cross-user model parameters, influencing individual models.

Baseline and Metrics   To assess our model, we compare it to three baseline methods. Mean averages fear scores within the train set, Random generates random scores, and Linear builds a basic regression model. All baselines undergo the same validation as prediction models for unbiased comparison. Metrics used include root mean squared error (RMSE) and mean absolute error (MAE), standard in regression tasks.

Overall Prediction Result    Our cross-user model excels in predicting fear of heights, outperforming baseline methods. RF and LightGBM models achieve superior results, surpassing the mean baseline by approximately 0.5 in both MAE and RMSE metrics. While generally effective for most participants, fluctuations occur due to individual variability. Linear Regression, despite common use, lacks advantages, possibly due to feature redundancy and non-linearity. In contrast, ensemble models like RF and LightGBM prove more robust and accurate in this scenario. Individual Prediction Result    We analyze the individual-based model’s performance on single-subject data, using the Random Forest model for computational convenience. Random Forest, excelling in cross-user MAE, outperforms linear regression and mean baseline in RMSE for each user. Prediction stability for individuals is notable, with RMSE below 1.5. Participant HIQ scores influence predictions, with better performance in higher fear groups (RMSE 1.01) compared to lower fear groups (1.17), possibly indicating heightened physiological responses in individuals with higher fear scores Impact of sensor combinations    In a comparative study exploring optimal sensor combinations with limited resources, pairs of EDA, ECG, and Eye Tracking sensors are evaluated. The EDA and ECG combination outperforms others, achieving an RMSE of 1.636 and an MAE of 1.271, comparable to the original three-sensor model in cross-user scenarios. While the full model demonstrates more stable output with a lower error standard deviation (0.319), the EDA+ECG model excels in prediction ability. Conversely, Eye Tracking and ECG yield the lowest performance, indicating that the EDA+ECG combination is the most effective sensor combination for fear prediction in this context.

Nolibox (confidential)

#Algorithm #AI Design #Knowledge Map

NeRF Virtual Avatar Reconsturction

Utilizing NeRF technology, I achieved the reconstruction of virtual avatars from 2D photos to 3D models. A 3D model was rebuilt from 24 high-definition photographs.

Learn more about this program

SpaceFlow

An astronaut-specific tea kettle leveraging space’s zero-gravity. It’s engineered to avoid spills and enables convenient swapping of tea bags.

Tools: Rhino, Realflow

Floral Echos

An artistic phone sound amplifier with a special chamber design for audio enhancement, doubling as a vase for fresh flowers, offering both visual and auditory delight.

Tools: Rhino, Keyshot

Bugoides

A biomimetic animal based on a four-bar linkage, replicating an insect’s gait, composed entirely of one type of part.

Tools: Solidworks, Mechanical Manufacturing

→ Learn more about this program

Kinetic Art

A dynamic art installation, emulating Alexandar’s style, depicting a Chinese verse:

“You stand on the bridge, beholding the view / While from the tower, the viewers behold you / The bright moon adorns your window’s frame / And in others’ dreams, you do the same.”

i: Companion

A personal companion app that establishes a unique personality, founded on the ‘I as my own partner’ principle, by constructing a virtual self from user’s linguistic inputs.

Tool:  Swift

Moonster Music

‘Moonster’ has identified users’ tendencies to either venture beyond or stick to their comfort zones in music.

This platform allows you to explore what songs others are enjoying and adapt your monster persona to mingle in various music-listening groups.

Tools: Figma

→ Visit my video to read more

Space Article Builder

An exploration into a new reading medium. I’ve envisioned a virtual space where users can create virtual buildings, embedding articles, videos, and other information within, offering a non-linear reading experience.

Tool: Minecraft

→ Visit my video to read more

New Wave

‘New Wave’ is a memorable art performance event. It took five months of preparation and was presented to over 1,000 spectators. Yuchen served as the stage director, and the event featured a ‘vintage radio’ artistic concept, linking all performances in a time-traveling radio story.

ASL Sign Search (not released)

#UX #CV #Assistive Technology

Connect with me

yuchenyao_thu@163.com

→ LinkedIn

→ Instagram

Tabc T

TabcT (Thermal Activated Bacteria Cancer Therapy) is a temperature-controlled bacterial therapy developed by Tsinghua 2023, targeting breast cancer, and it offers the advantages of being 5C: cheap, convenient, controlled, continuous, and comprehensive. This is a collaborative project, Yuchen partook in this program as a design advisor and software developer.

→ Visit our wiki to see more

Fufu Companion

“Record and soothe depression with a soft robot”

Fufu’s diary is program aimed at helping the depression patients, especially in the situation thatCOVID-19 pandemic is separating people physically, making it harder to get offline treatment.

Fufu contains both Online And Offline (OAO) product, all the way to an integrated service system. Here we present fufu’s diary, in a very story-telling way.

→ See more details

MuSee

“Reimagine the unheard miracles in an alternate expression”

Imagine a world where no mellifluous birdsong, soft-spoken words, or majestic melodies exist. How poignant to miss these sounds! Yet, 448 million globally endure this quietude due to auditory impairment.

In 2021, a group of Tsinghua University students aspired to bridge this void. They sought to ‘play’ a symphony in lights, enabling those with hearing loss to partake in the symphony’s wonder, in a one-to-one sensory dance.

→ See more details

Unlock Fear

Current Status: IMWUT 24′ under review

“Unlock the secret of fear”

Acrophobia, the fear of heights, significantly affects overall quality of life. People who suffer from acrophobia know this fear is excessive and unreasonable, yet 6.4% of total population suffer from this issue.

We investigated the feasibility of passive sensing data for fear measurement and discern the most influential physiological and behavioral indicators. 

→ See more details

Resonant Echoes​

“Bridging Deaf Children to Quality Education Recovery”

Resonant Echoes  is an innovative augmented reality interface that helps teachers understand the intentions of language-impaired kids, facilitating better rehabilitation and social engagement for deaf students.

Deaf children often lack sufficient language skills during rehabilitation, which hinders their socialization due to difficulties in communication. The ‘Resonant Echoes’ helps therapists overcome understanding difficulties with an XR display and embedded AI, focusing on a role-play game scenario.

→ See more details

G-Voila

Current Status: IMWUT 24′ under review

“Can AI understand us better if eye data is shared?”

As developers of general artificial intelligence increasingly shift their focus towards incorporating more modalities of information such as images and sounds, we can’t help but wonder: If we provide such rich eye movement information to artificial intelligence, could it enable them to better understand us?

→ See more details

Kosmos-2 Interface (not released)

#UX #LLM #Interface

Time of City

An artistic installation that embodies a variety of light and shadow patterns within its Sisyphean cycle.

Tools: Raspberry Pi, Smart Car

→ Visit my video to read more