+ - 0:00:00
Notes for current slide
Notes for next slide

These slides are made, making use of the Xaringan package in R. Great Stuff!

Modelling eye-tracking data in comparative judgement


EARLI SIG27 - 2020

Sven De Maeyer; Marije Lesterhuis; Marijn Gijsen

University of Antwerp

2020-11-25

1 / 26

Introduction

2 / 26

These slides are made, making use of the Xaringan package in R. Great Stuff!

Judging 'argumentative texts'

alt text

3 / 26

Two types of processes?

http://www.youtube.com/watch?v=z2f_Ue45KWM

4 / 26

So...

Gaze duration data might result from different cognitive processes:

shorter visits in both texts (reading for building a first 'mental model' / scanning)

longer visits in one of both texts (reading for text comprehension)

5 / 26

But...

How to statistically model the resulting duration data?

... without categorisation (setting tresholds to distinguish scanning from text comprehension)

... and avoiding aggregation

Goal of this research:

build and test statistical models that acknowledge the data generating cognitive processes and that do not make use of 'trimming', 'setting tresholds' or 'aggregation'

6 / 26

Methodology

7 / 26

Procedure

  • 26 high school teachers (Dutch) voluntary participated

  • each did 10 comparisons of 2 argumentative texts from 10th graders

  • 3 batches of comparisons with random allocation of judges to one of the batches

  • all batches similar composition of comparisons regarding the characteristics of the pairs; the pairs, however not the same

  • Tobii TX300 dark pupil eye-tracker with a 23-inch TFT monitor (max. resolution of 1920 x 1080 pixels)

  • data sampled binocularly at 300 Hz

8 / 26

AOI's

alt text

9 / 26

The data

10 / 26

Statistical model 1

Simple mixed effects model

= 'Gaze Event duration' of AOI visits are nested within the combination of juges and comparisons

yi(jk)=β0+μ0j+ν0k+ϵi(jk)

with:

  • yi(jk) = gaze event duration of a visit to an AOI (text);
  • β0 = the intercept (overall average duration);
  • μ0j = unique effect of judge j;
  • ν0k = unique effect of comparison k;
  • ϵi(jk) = residual;
11 / 26

Statistical model 2

first mixed effects finite mixture model

Model1 + assuming two data generating processes + gaze event durations from 'text comprehension' differ for judges and comparisons

yi(jk)=θ×(β1+ϵ1i(jk))+(1θ)×(β2+μ2j+ν2k+ϵ2i(jk))

with:

  • β1 = intercept 1, overall average duration process 1 'scanning';
  • β2 = intercept 2, overall average duration process 2 'text comprehension reading';
  • μ2j = unique effect of judge j on 'text comprehension reading';
  • ν2k = unique effect of comparison k on 'text comprehension reading';
  • ϵ1i(jk) & ϵ2i(jk) = residuals;
  • θ = mixing proportion (weight)
12 / 26

Statistical model 3

Model 1 + assuming two data generating processes + gaze event durations from 'scanning' differ for judges and comparisons

yi(jk)=θ×(β1+μ1j+ν1k+ϵ1i(jk))+(1θ)×(β2+ϵ2i(jk))

with:

  • β1 = intercept 1, overall average duration process 1 'scanning';
  • β2 = intercept 2, overall average duration process 2 'text comprehension reading';
  • μ1j = unique effect of judge j on 'scanning';
  • ν1k = unique effect of comparison k on 'scanning';
  • ϵ1i(jk) & ϵ2i(jk) = residuals;
  • θ = mixing proportion (weight)
13 / 26

Statistical model 4

Model 1 + two data generating processes + gaze event durations from both processes differ for judges and comparisons

yi(jk)=θ×(β1+μ1j+ν1k+ϵ1i(jk))+(1θ)×(β2+μ2j+ν2k+ϵ2i(jk))

With:

  • β1 = intercept 1, overall average duration process 1 'scanning';
  • β2 = intercept 2, overall average duration process 2 'text comprehension reading';
  • μ1j & μ2j = unique effect of judge j on 'scanning' & 'text comprehension reading';
  • ν1k & ν2k = unique effect of comparison k on 'scanning' & 'text comprehension reading';
  • ϵ1i(jk) & ϵ2i(jk) = residuals;
  • θ = mixing proportion (weight)
14 / 26

Analysis (Bayesian estimation)

  • brms (wrapper around Stan) to estimate the models making use of MCMC in R

  • flat (uninformed) priors

  • 4 chains of 15000 iterations (with 1000 burn-in it.)

  • compare the models with 'leave-one-out cross-validation' approach

  • summarize best model (interpret the posterior distribution)

15 / 26

Results

16 / 26

Comparison of the models

Model comparison expressed as expected log predictive density with standard errors between brackets
`Δelpd^` `elpd^`
Model 4 0.0 (0.0) -5548.8 (29.5)
Model 2 -56.4 (10.3) -5604.8 (29.3)
Model 3 -121.4 (20.1) -5669.9 (28.4)
Model 1 -239.7 (19.6) -5788.1 (30.5)

Model 4 fits best

  • 2 data generating processes

  • differences between judges AND comparisons

17 / 26

Posterior distribution of fixed effects


average duration of visits coming from ...

  • 'scanning' +/- .44 secs

  • 'text comprehension' +/- 8.10 secs

18 / 26

Posterior distribution of theta & residual variances


  • more visits from 'scanning' than 'text comprehension'

  • residual variance larger for 'text comprehension'

19 / 26

Posterior distribution of random effects


  • judges stronger impact on both types of durations

  • biggest differences for durations form 'text comprehension'

20 / 26

Diff. between judges in reading for text comprehension


This plot shows how judges differ in durations coming form 'text comprehension'

21 / 26

Diff. between comparisons in reading for text comprehension


This plot shows how comparisons result in different durations for 'text comprehension'

clearly more alike than judges

22 / 26

Conclusion & Discussion

23 / 26

Conclusion

  • model that acknowledges gaze event durations come from different cognitive processes is most likely

  • judges and comparisons result in different gaze event durations

  • mixed effects finite mixture models are promising for this kind of data

    • no need to decide which fixation duration point to scanning and which to text comprehension

    • opens up many possibilities for follow-up research questions and analyses

24 / 26

Discussion

  • what about more than 2 processes?

  • more informed priors

  • a need for triangulation to understand the two types of processes

  • what if representations are not texts; how task specific are these models?

  • other applications of (mixed effects) finite mixture models when modelling process data?

25 / 26

Questions?

Do not hesitate to contact us!

sven.demaeyer@uantwerpen.be

The material is shared on OSF [https://osf.io/evrsf/]

26 / 26

Introduction

2 / 26

These slides are made, making use of the Xaringan package in R. Great Stuff!

Paused

Help

Keyboard shortcuts

, , Pg Up, k Go to previous slide
, , Pg Dn, Space, j Go to next slide
Home Go to first slide
End Go to last slide
Number + Return Go to specific slide
b / m / f Toggle blackout / mirrored / fullscreen mode
c Clone slideshow
p Toggle presenter mode
t Restart the presentation timer
?, h Toggle this help
Esc Back to slideshow