Skip to content

This dataset contains the verbal, head, and eye behaviour of ten pairs of participants performing a collaborative explorative data analysis task in VR; for more information about this, read the section Data. Furthermore, this dataset also contains the segmented and semantically annotated visual context of the explorative data analysis.

Notifications You must be signed in to change notification settings

Collaborative-Immersive-Visual-Toolkit/Speech-Gaze-Head-Datatset

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 

Repository files navigation

Collaborative Speech Gaze and Head Behaviour Datatset

This dataset contains the verbal, head, and eye behaviour of ten pairs of participants performing a collaborative explorative data analysis task in VR; for more information about this, read the section Data. Furthermore, this dataset also contains the segmented and semantically annotated visual context of the explorative data analysis; for more information about this, read the section Visualizations. This dataset has been released with the paper Speech-Augmented Cone-of-Vision for Exploratory Data Analysis at CHI 2023.

@article{Bovo:2023,
    author  = "Riccardo Bovo, Daniele Giunchi, Ludwig Sidenmark, Hans Gellersen, Enrico Costanza, Thomas Heinis", 
    year    = 2023,
    journal = "TBD",
    doi = "10.1145/3544548.3581283"  }

Folder structure

The dataset contains 2 folders: Data, Visualizations

Data

The Data folders contains 10 foilders each representing one of the experimental session. Each experimental session folder is named with the dayNumber_hourNumber ( 365 days a hear 24 hours a day). Each experimental session contains 3 experimental trials folder, each folder name is composed by the experimentalCondition_visualizationNumber

The experimenta lConditions are 1 = "CoV", 2 = "CoV+Speech", 3 = "eye-cursor".
The visualization Number are 1 = "Movie Dataset", 2 = "Gender bias in Movie Dataset", 3 = "Car insurance risk Dataset".

Within each experimental trials there is a data.csv file, the section Columns describe the data contained with the CSV file. For more information about the information contained in the data.csv file read the section Columns.

Visualizations

The Visualization folder contains a texture folder which contains the 24 images (8 screen for each of the 3 dataset) furtheremore it contains keywords_coordinates.json containing the segmented and semantically annotated information of each datatset and

data.csv

Each data.csv file contains the 218 columns.
A part from time_seconds all other colums have a either U1 U2 prefix indicating the user.
For example LeftEyePosX,Y,Z means there will be 6 different columns: U1LeftEyePosX, U1LeftEyePosY, U1LeftEyePosZ, U2LeftEyePosX, U2LeftEyePosY, U2LeftEyePosZ

time_seconds : time in seconds since the start of the trial
PosX,Y,Z : position in word coordinate of the player object
HeadPosX,Y,Z : position in word coordinate of the players head
HeadForward,X,Y,Z : forward direction of the players head (i.e. direction of the nose)
HeadUpX,Y,Z : upward direction of the players head
LeftEyePosX,Y,Z : position in word coordinate of the players left eye
RightEyePosX,Y,Z : position in word coordinate of the players right eye
LeftEyeVecX,Y,Z : forward direction of the players left eye
RightEyeVecX,Y,Z : forward direction of the players right eye
GazeX,Y,Z : word position of the intersection between the eyes forward vector (left and right) and the data
GazeU,V : normalized texture positon of the intersection between the eyes forward vector (left and right) and the data (this makes it easier to underst on which element the user is whatchig)
ControllerRPosX,Y,Z : word position of the Right controller
ControllerREAngX,Y,Z : euler angles of the Right controller
ControllerLPosX,Y,Z : word position of the Left controller
ControllerLEAngX,Y,Z : euler angles of the Left controller
PointerLeftX,Y,Z : word position of the intersection between the Left controller direction (use euler angles to calculate) and the data
PointerRightX,Y,Z : word position of the intersection between the Right controller direction (use euler angles to calculate) and the data
PointerLeftU,V : normalized texture positon of the intersection between the Right controller direction and the data (this makes it easier to underst on which element the user is pointing at)
PointerRightU,V : normalized texture positon of the intersection between the Left controller direction and the data (this makes it easier to underst on which element the user is pointing at)
Condition : integer value indicating the experimental condition ( 0 "undefined", 1 "CoV", 2 "CoV+Speech", 3 "eye-cursor" )
Visualization : integer value indicating the dataset ( 0 "undefined", 1 "Movie Dataset", 2 "Gender bias in Movie Dataset", 3 "Car insurance risk Dataset" )
InsightRecording : binary valu indicating if the user is recording an insight or not ( 0 "particpant not recording" , 1 "particpant recording" )
VisualizationHalf : integer value indicating which half of the visualization is beeing displayed ( 0 "first" , 1 "second" )
HeadGazeX,Y,Z : word position of the intersection between the head forward vector and the data
HeadGazeU,V : normalized texture positon of the intersection between the head forward vector and the data (this makes it easier to underst on which element the user head is pointing at)
HeadCone00_x to HeadCone19_x : U coordinate position of the 20 points of the projected CoV (for the CoV + Speech such coordinate are dynamic)
HeadCone00_y to HeadCone19_y : V coordinate position of the 20 points of the projected CoV (for the CoV + Speech such coordinate are dynamic)
transcription : this is the data extracted from the audio files with the whisper model (audio files not included due to anonimoization)

Visualizations

The visualization section contains information about the vidsualized dataset.

Visualized Dataset

The visualizations conists of 3 datasets:

  • Movie Dataset
  • Gender Bias in Movie Dataset
  • Insurance risk in cars Datatset

The texture folder which contains the 24 images (8 screen for each of the 3 dataset), the Movie Dataset textures are named :

  • Oscar.png
  • Scatterplot1.png
  • Scatterplot2.png
  • BoxAndWhiskers.png
  • BoxAndWhiskers2.png
  • Histograms.png
  • StackBarChart.png
  • Instructions.png

The Gender Bias in Movie Dataset have the same names but with the suffix _Gender.png The Insurance risk in cars Datatset have the same names but with the suffix _Third.png

Layout

Each Dataset pages is visualized in a round layout (see image below)

U V Coordinates

As each page is rendered on a web browser displayed in the VR environment with a size of 980px by 551px the total UV texture dimension is 7344px by 551px.

Elements bounding boxs

The elements bounding boxes are contained in the keywords_coordinates.json file and they can be loaded in to UV texture space with the python file LoadRoi.py

About

This dataset contains the verbal, head, and eye behaviour of ten pairs of participants performing a collaborative explorative data analysis task in VR; for more information about this, read the section Data. Furthermore, this dataset also contains the segmented and semantically annotated visual context of the explorative data analysis.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages