Introduction

Elderly Multimodal Interpersonal Conversation Dataset (E-MIC Dataset) is a dataset built to recognize the timing of turn-taking in human-human social interaction. We collected interpersonal conversational interaction data, including video and audio, from 100 elderly people.

Fig 1_dataset
[Figure 1] Samples of E-MIC dataset
  • The subjects are 100 elderly people over the age of 65 (66 women and 34 men, 65~ 85 years old).
  • We collected RGB image data with two cameras and audio data with two microphones.
  • One session involves two elderly people sitting face to face and interacting with daily conversations, each lasting 15 to 20 minutes.
  • The E-MIC dataset includes various conversations (e.g., travel, hobby, health) of the elderly.
  • The E-MIC dataset contains conversation interaction data of the elderly from a first-person perspective.
  • The E-MIC dataset contains approximately 33 hours of video and three hours of audio data.

Annotations

We annotated 19 types of information (e.g., take, release, wait, hold, gaze, backchannel, lip state) related to four types of turn-taking.

take hold release wait
interrrupt speech_start pause backchannel
gaze_front speech_end gaze_front affirm
lip_open gesture_hand lip_close other_state
negation   lip_habit other_speak

The E-MIC dataset used the annotation tool ELAN (https://tla.mpi.nl/tools/tla-tools/elan/download/)) to generate annotations about images and audios. Annotation information can be edited through the .eaf file included in this dataset and exported (export) in various forms of file (XML, CSV, etc.)

Fig 2_annotation
[Figure 2] Sample of annotation of the collected video and audio clip

Collected Data

Collected Data File Format Size
RGB videos MPEG4 89.84GB
Audio data WAV 20.58GB
Annotation data EAF, CSV, XML 25.24MB

Setup Environment

We collected interpersonal conversational interaction data, including video and audio, from 100 elderly people. Figure 3 shows the data collection environment. Participants wear microphones, sit facing each other, and have conversations.

Fig 3_setup
[Figure 3] The data collection setup environment: (left) participants wear microphones, sit facing each other, and have conversations

Download

Please follow the link below, and join as a member to get to the download page:

Contact

Please email bettle@etri.re.kr if you have any questions or comments.

Acknowledgment

  • The protocol and consent of data collection were approved by the Institutional Review Board(IRB) at Suwon Science College, our joint research institute.
  • This work was supported by the ICT R&D program of MSIP/IITP. [2017-0-00162, Development of Human-care Robot Technology for Aging Society].