VocADom@DOMUS

(in construction)

Context of the acquisitions

This corpus was recorded by the LIG laboratory (Laboratoire d’Informatique de Grenoble, UMR 5217 CNRS/UGA) thanks to the  VocADom project founded by the French National Research Agency (Agence Nationale de la Recherche/ANR-16-CE33-0006). The authors would like to thank the participants who accepted to perform the experiments.

This corpus is composed of audio and home automation data acquired in a real smart home with French speakers. This campaign was conducted within the VocADom project aiming at designing a new smart home system based on audio technology. The developed system provides assistance via natural man-machine interaction (voice and tactile command) and security reassurance by detecting distress situations so that the person can manage, from anywhere in the house, her environment at any time in the most natural way possible.

Home automation Smart Home of the LIG Laboratory

DOMUS smart apartment is part of the experimentation platform of the LIG laboratory and is dedicated for research projects.  According to the different research projects, experimentations are conducted with users performing scenarios of daily housework and leisure. Multimodal corpus are produced, synchronized and analyzed in order to evaluate and validate the concerned concept or system.

Appartment

DOMUS is a 35m2 apartment used to collect this dataset has the following layout (kitchen, bathroom, bedroom, living room):

Sensors

DOMUS is fully functional and equipped with sensors, such as energy and water consumption, level of hygrometry, temperature, and effectors able to control lighting, shutters, multimedia diffusion, distributed in the kitchen, the bedroom, the office and the bathroom. An observation instrumentation, with cameras, microphones and activity tracking systems, allows to control and supervise experimentations from a control room connected to Amiqual4Home. The flat also is equipped with 16 microphones (4 arrays of 4 microphones each) set into the ceiling that can be recorded in real-time thanks to a dedicated software able to record simultaneously the audio channels.

Description of the dataset

Grammar of the voice command

set/check an actuator: key initiateCommand object
(e. g., KEYWORD turn off the light)
(e. g., KEYWORD is the light on?)
emergency call: key emergencyCommand
(e. g., KEYWORD help)

Scenarios

  1. Phase 1:
        • (2 persons) keywords, elicitation of vocal commands

        • (1 person) free discovery of the vocal command system

        • (2 persons) demonstration of the system to a naive participant (trainee)

      • (3 persons) interview with the experimenter
  2. Phase 2:
        • (4 persons) 2 technicians are coming in to maintain the system during an interview between the experimenter and the participant

      • (2 persons) technicians are going out and a risk situation is notified

Participants’ recordings (duration in the format hour:minute:second; AG=aged people, VI=visually impaired)

Participant Age group (years) Gender/Type Duration Chosen keyword
S00 not available not available not available not available
S01 50-60 M/VI 01:25:52 minouche
S02 50-60 M/VI 01:06:40 minouche
S03 50-60 M/VI 01:08:56 allo cirrus
S04 60-70 M/AG 01:06:43 ulysse
S05 60-70 M/AG 01:25:15 allo cirrus
S06 70-75 F/AG 01:10:39 ulysse
S07 60-70 F/AG 01:07:11 minouche
S08 <30 F/VI 00:59:08 vocadom
S09 50-60 M/VI 01:26:04 ulysse
S10 70-75 F/AG 01:16:56 minouche
S11 50-60 M/VI 01:28:25 ulysse
S12 50-60 F/VI 01:09:41 ulysse
All (mean) 50-60 5F/7M 14:51:30 4 keywords

Data structure

record/ S<number>/ mic_array/ (microphone array recordings)
S<number>.txt (non identifiant information)
openhab_log/ (log of the home automation network)
video/ (video records for annotation purpose)

Annotations

Data type Format
Audio Transcriber
Localisation and activities Elan
NLU Home made