Quarterly (Winter, Spring, Summer, Fall)
170 pp. per issue
8 1/2 x 11
E-ISSN
2470-2986

Open Mind

2021, Vol. 5, Pages 20-29
(doi: 10.1162/opmi_a_00039)
© 2021 Massachusetts Institute of Technology. Published under a Creative Commons Attribution 4.0 International (CC BY 4.0) license.
SAYCam: A Large, Longitudinal Audiovisual Dataset Recorded From the Infant’s Perspective
Article PDF (423.11 KB)
Abstract

We introduce a new resource: the SAYCam corpus. Infants aged 6–32 months wore a head-mounted camera for approximately 2 hr per week, over the course of approximately two-and-a-half years. The result is a large, naturalistic, longitudinal dataset of infant- and child-perspective videos. Over 200,000 words of naturalistic speech have already been transcribed. Similarly, the dataset is searchable using a number of criteria (e.g., age of participant, location, setting, objects present). The resulting dataset will be of broad use to psychologists, linguists, and computer scientists.