Skip to content

The repository contains our implementation for the work to be presented at Ubicomp 2022

Notifications You must be signed in to change notification settings

asalekin-ubiquitouslab/Modality-wise-Multple-Instance-Learning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 

Repository files navigation

Modality-wise-Multple-Instance-Learning

This reposistory contains the code for the paper Psychophysiological Arousal in Young Children Who Stutter: An Interpretable AI Approach, IMWUT 2022

Modality Invariant-MIL (MI-MIL) Approach

The MI-MIL approach takes the modality-specific bag representations (𝐵𝑚 ={𝑥1𝑚,𝑥2𝑚,...𝑥𝑘𝑚}, 𝑘 = 19,𝑚 = EDA, HR, RSP-amp, RSP-rate) of a 20s physiological sensing data as input. As shown in figure below, MI-MIL has four components: (1) modality specific embedding block, (2) modality specific self-attention pooling block, (3) modality fusion Block, and (4) classifier Block. While the first two blocks are applied to each modality 𝑚 independently, the latter two combine the cross-modality information to generate inference.

MI-MIL

ACM Reference Format:

Harshit Sharma, Yi Xiao, Victoria Tumanova, and Asif Salekin. 2022. Psychophysiological Arousal in Young Children Who Stutter: An Interpretable AI Approach. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 6, 3, Article 137 (September 2022), 32 pages. https://doi.org/10.1145/3550326

Contact

[email protected]

About

The repository contains our implementation for the work to be presented at Ubicomp 2022

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published