Skip to content

Multimoal Emotion Recognition for senior people

License

Notifications You must be signed in to change notification settings

Neuro-AI-Lab/MER-Senior

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

25 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

1. MER based on GCL for Senior

This is the implementation module for multimodal emotion recognition (MER) based on graph contrastive learning (GCL).

The database consists of text and audio which is acquired when uttering scripts evoking certain emotions.

There are 7 target emotions: 1) joy, 2) neutral, 3) anxiety, 4) embarrassment, 5) hurt, 6) sadness, and 7) anger

image

2. Dependencies

  • torch
  • pandas
  • numpy
  • sklearn
  • pyyaml
  • typing
  • matplotlib
  • datetime

Install all dependencies as

pip install -r requirements.txt

You should consider deep learning setups such as CUDA and PyTorch versions available in your local environments.

3. Usage

Train and evaluate the model by executing as

python train.py --dataset IITP-SMED-STT --cuda_id 0

Available --dataset arguments must be one of [IITP-SMED-ORIGIN, IITP-SMED-STT, IITP-SMED-AUDIO, IITP-SMED-ORIGIN-TEXT, IITP-SMED-STT-TEXT]

You can choose a single GPU, and cuda_id is the order of available GPU devices.

IITP-SMED and IITP-SMED-STT are our empirical datasets constructed by taking funds from IITP in South Korea.

See details of AIHUB-SER datasets online available link.

About

Multimoal Emotion Recognition for senior people

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%