Multimodal Affective Computing

2021–2027

Affective Computing Signal Processing Virtual Reality Multimodal Learning
Multimodal Affective Computing

Overview

Collaborating with Shaftesbury (now Wellwave Inc.), we will develop a machine learning model for mutlimodal assessment of stress. The multimodal sensors can be physiological (e.g. EEG, heart-rate) and behavioural (e.g. facial expressions). The target is to use the assessed stress for Shaftesbury’s Positive Distraction Entertainment System which adapts game content dynamically to reduce stress in children before a complex medical procedure, which can reduce complexity and recovery time.

Funding

  • NSERC Collaborative Research and Development Grant (CRD) - $450K
  • NSERC Alliance + Mitacs Accelerate Grant - $488K
  • New Frontiers in Research Fund - Exploration - $250K