%0 Generic %A Raman, Chirag %A Vargas Quiros, Jose %A Tan, Stephanie %A Islam, Ashraful %A Gedik, Ekin %A Hung, Hayley %D 2022 %T Raw Data for ConfLab: A Data Collection Concept, Dataset, and Benchmark for Machine Analysis of Free-Standing Social Interactions in the Wild %U https://data.4tu.nl/articles/dataset/Raw-Data_for_ConfLab_A_Rich_Multimodal_Multisensor_Dataset_of_Free-Standing_Social_Interactions_In-the-Wild/20017748/2 %R 10.4121/20017748.v2 %K data-raw %K conflab %K cameras %K wearables %K multimodal %K ConfLab %X

This file contains raw data for cameras and wearables of the ConfLab dataset. 


./cameras 

contains the overhead video recordings for 9 cameras (cam2-10) in MP4 files. 

    These cameras cover the whole interaction floor, with camera 2 capturing the

    bottom of the scene layout, and camera 10 capturing top of the scene layout.

    Note that cam5 ran out of battery before the other cameras and thus the recordings 

    are cut short. However, cam4 and 6 contain significant overlap with cam 5, to 

    reconstruct any information needed.


    Note that the annotations are made and provided in 2 minute segments. 

    The annotated portions of the video include the last 3min38sec of x2xxx.MP4 

    video files, and the first 12 min of x3xxx.MP4 files for cameras (2,4,6,8,10), 

    with "x" being the placeholder character in the mp4 file names. If one wishes 

    to separate the video into 2 min segments as we did, the "video-splitting.sh"

    script is provided. 


./camera-calibration contains the camera instrinsic files obtained from 

    https://github.com/idiap/multicamera-calibration. Camera extrinsic parameters can 

    be calculated using the existing intrinsic parameters and the instructions in the 

    multicamera-calibration repo. The coordinates in the image are provided by the 

    crosses marked on the floor, which are visible in the video recordings. 

    The crosses are 1m apart (=100cm).  


./wearables

subdirectory includes the IMU, proximity and audio data from each 

    participant at the Conflab event (48 in total). In the directory numbered 

    by participant ID, the following data are included:

        1. raw audio file

        2. proximity (bluetooth) pings (RSSI) file (raw and csv) and a visualization 

        3. Tri-axial accelerometer data (raw and csv) and a visualization 

        4. Tri-axial gyroscope data (raw and csv) and a visualization 

        5. Tri-axial magnetometer data (raw and csv) and a visualization

        6. Game rotation vector (raw and csv), recorded in quaternions. 


    All files are timestamped.

    The sampling frequencies are:

        - audio: 1250 Hz

        - rest: around 50Hz. However, the sample rate is not fixed

        and instead the timestamps should be used. 


    For rotation, the game rotation vector's output frequency is limited by the 

    actual sampling frequency of the magnetometer. For more information, please refer to 

    https://invensense.tdk.com/wp-content/uploads/2016/06/DS-000189-ICM-20948-v1.3.pdf


    Audio files in this folder are in raw binary form. The following can be used to convert

    them to WAV files (1250Hz):


        ffmpeg -f s16le -ar 1250 -ac 1 -i /path/to/audio/file


Synchronization of cameras and werables data

    Raw videos contain timecode information which matches the timestamps of the data in

    the "wearables" folder. The starting timecode of a video can be read as:

        ffprobe -hide_banner -show_streams -i /path/to/video


./audio

./sync: contains wav files per each subject

./sync_files: auxiliary csv files used to sync the audio. Can be used to improve the synchronization. 

The code used for syncing the audio can be found here:   

https://github.com/TUDelft-SPC-Lab/conflab/tree/master/preprocessing/audio 

%I 4TU.ResearchData