Automatic Detection of Gaze and Body Orientation in Elementary School Classrooms

Araya, Roberto and Sossa-Rivera, Jorge (2021) Automatic Detection of Gaze and Body Orientation in Elementary School Classrooms. Frontiers in Robotics and AI, 8. ISSN 2296-9144

[thumbnail of pubmed-zip/versions/2/package-entries/frobt-08-729832-r1/frobt-08-729832.pdf] Text
pubmed-zip/versions/2/package-entries/frobt-08-729832-r1/frobt-08-729832.pdf - Published Version

Download (1MB)

Abstract

Detecting the direction of the gaze and orientation of the body of both teacher and students is essential to estimate who is paying attention to whom. It also provides vital clues for understanding their unconscious, non-verbal behavior. These are called “honest signals” since they are unconscious subtle patterns in our interaction with other people that help reveal the focus of our attention. Inside the classroom, they provide important clues about teaching practices and students' responses to different conscious and unconscious teaching strategies. Scanning this non-verbal behavior in the classroom can provide important feedback to the teacher in order for them to improve their teaching practices. This type of analysis usually requires sophisticated eye-tracking equipment, motion sensors, or multiple cameras. However, for this to be a useful tool in the teacher's daily practice, an alternative must be found using only a smartphone. A smartphone is the only instrument that a teacher always has at their disposal and is nowadays considered truly ubiquitous. Our study looks at data from a group of first-grade classrooms. We show how video recordings on a teacher's smartphone can be used in order to estimate the direction of the teacher and students’ gaze, as well as their body orientation. Using the output from the OpenPose software, we run Machine Learning (ML) algorithms to train an estimator to recognize the direction of the students’ gaze and body orientation. We found that the level of accuracy achieved is comparable to that of human observers watching frames from the videos. The mean square errors (RMSE) of the predicted pitch and yaw angles for head and body directions are on average 11% lower than the RMSE between human annotators. However, our solution is much faster, avoids the tedium of doing it manually, and makes it possible to design solutions that give the teacher feedback as soon as they finish the class.

Item Type: Article
Subjects: EP Archives > Mathematical Science
Depositing User: Managing Editor
Date Deposited: 28 Jun 2023 04:14
Last Modified: 16 Sep 2023 05:13
URI: http://research.send4journal.com/id/eprint/2447

Actions (login required)

View Item
View Item