Upload
others
View
0
Download
0
Embed Size (px)
Citation preview
Supporting Networked Collaborative Learning !Using Eye-Tracking Technology to Augment Joint Visual Attention"
!!!
Bertrand Schneider, Roy Pea
Stanford University, School of Education
METHODS Participants: 42 college-level students from a community college (average age 23.0, SD = 8.3; 28 females, 14 males) Experimental Conditions: “visible-gaze” condition with the gaze-awareness tool (N = 24); “no-gaze” condition (N = 20) without the gaze-awareness tool. Measures: • Learning test (3 sub-dimensions: concepts, memory, transfer) • Quality of collaboration (Meier, Spada and Rummel, 2007) • Eye-tracking data (fixations, saccades, pupil size)
INTRODUCTION Joint attention is fundamental to any kind of social coordination: young infants communicate their emotions by being in a state of synchrony with their caregivers, which in turn helps them achieve visual coordination when learning to speak (Stern, 1977; Brooks & Meltzoff, 2008). Professors teach by highlighting subtle nuances between students’ and experts’ conceptual understanding of a domain (Roth, 2001). Groups of students rely on the coordination between its members to reach the solution of a problem (Barron, 2003).
PURPOSE The goal of our work is to develop new ways of supporting the establishment of joint attention. We use eye-tracking technologies to share users’ gaze in a collaborative learning situation. More specifically, our first attempt involves dyads studying contrasting cases (Schwartz & Bransford, 1998) in a remote collaboration. We introduce a new kind of awareness tool that provides participants with the position of their partner’s gaze on the screen, allowing for real-time mutual gaze perception. RESULTS
1. Dyads learned more and had a higher quality of collaboration when using the gaze-awareness tool (F(1,40) = 7.81, p < 0.01). Interaction effect with the students’ status (leader/follower): F(1,38) = 5.29, p < 0.05
2. The number of moments of joint attention is sig. correlated with the learning gain (r = 0.39, p < 0.05)
DISCUSSION Conclusions: • Our study shows that the role of joint attention is preponderant in collaborative learning situations. • This process can be enhanced by technological tools (i.e. eye-trackers)
Future Work: • replicate those results in other settings (e.g. co-located dyad)
References Barron, B. (2003). When smart groups fail. Journal of the Learning Sciences, 12(3), 307-359.
Bransford, J., & Schwartz, D. (1999). Rethinking Transfer: A Simple Proposal with Multiple Implications. Review of Research in Education, 24.
Brooks, R., & Meltzoff, A. N. (2008). Infant gaze following and pointing predict accelerated vocabulary growth through two years of age: a longitudinal, growth curve modeling study. Journal of Child Language, 35(01), 207–220.
Meier, A., Spada, H., & Rummel, N. (2007). A rating scheme for assessing the quality of computer- supported collaboration processes. International Journal of Computer- Supported Collaborative Learning, 2, 63–86.
Roth, W. M. (2001). Gestures: Their role in teaching and learning. Review of Educational Research, 71, 365–392.
Stem, D. (1977). The first relationship. Cambridge, MA: Harvard University Press. Trilling, B., & Fadel, C. (2009). 21st Century Skills: Learning for Life in Our Times. John Wiley & Sons.
Acknowledgment: Supported by a grant from the National Science Foundation (SMA-0835854) to the LIFE Science of Learning Center.
SLD-2: Joint Attention Theory Gate 2
Figure 2. Results of the learning test (categorizing each subject as leader or follower).
Figure 1. Dyads worked on these five contrasting cases. The right side shows possible answers. Using a jigsaw collaborative learning method, for any given dyad, the answer for case #1 was given to subject-1 and the answer for case #6 was given to subject-2. Participants had to solve three remaining cases (#2, top middle and #4, #5 bottom left and right of the screen).
EXPERIMENTAL TASK Dyads remotely worked on a set of contrasting cases to learn how the human brain processes visual information (12 min). They then read a textbook chapter on the same topic (12 min.
0"
0.1"
0.2"
0.3"
0.4"
0.5"
0.6"
0.7"
0.8"
Lesions" Terminology" Transfer" Total"
Learning(Gain(for(Followers(/(Leaders(in(dyads(
dyad<gaze:follower"
dyad<gaze:leader"
dyad<nogaze:follower"
dyad<nogaz:leader"
visible−gaze
no−gaze
0.0 0.2 0.4 0.6Percentage
cond
ition
s
Joint Attention
Figure 3. Percentage of moments of joint attention (JA) across our two experimental conditions. Dyads in the “visible-gaze” condition had sig. more moments of JA (F(1,30) = 22.45, p < 0.001)
Conditions
Joint Attention
Collaboration
Cog. Load
Learning
c’
a
a
a
b
b
b
Figure 4. mediation model for the learning gain. Only the number of moments of joint attention was a significant mediator (CI: [0.03; 0.19]).