Look who's talking: A comparison of automated and human-generated speaker tags in naturalistic day-long recordings.

dc.contributor.author

Bulgarelli, Federica

dc.contributor.author

Bergelson, Elika

dc.date.accessioned

2020-01-01T18:47:34Z

dc.date.available

2020-01-01T18:47:34Z

dc.date.issued

2019-07-24

dc.date.updated

2020-01-01T18:47:33Z

dc.description.abstract

The LENA system has revolutionized research on language acquisition, providing both a wearable device to collect day-long recordings of children's environments, and a set of automated outputs that process, identify, and classify speech using proprietary algorithms. This output includes information about input sources (e.g., adult male, electronics). While this system has been tested across a variety of settings, here we delve deeper into validating the accuracy and reliability of LENA's automated diarization, i.e., tags of who is talking. Specifically, we compare LENA's output with a gold standard set of manually generated talker tags from a dataset of 88 day-long recordings, taken from 44 infants at 6 and 7 months, which includes 57,983 utterances. We compare accuracy across a range of classifications from the original Lena Technical Report, alongside a set of analyses examining classification accuracy by utterance type (e.g., declarative, singing). Consistent with previous validations, we find overall high agreement between the human and LENA-generated speaker tags for adult speech in particular, with poorer performance identifying child, overlap, noise, and electronic speech (accuracy range across all measures: 0-92%). We discuss several clear benefits of using this automated system alongside potential caveats based on the error patterns we observe, concluding with implications for research using LENA-generated speaker tags.

dc.identifier

10.3758/s13428-019-01265-7

dc.identifier.issn

1554-351X

dc.identifier.issn

1554-3528

dc.identifier.uri

https://hdl.handle.net/10161/19711

dc.language

eng

dc.publisher

Springer Science and Business Media LLC

dc.relation.ispartof

Behavior research methods

dc.relation.isversionof

10.3758/s13428-019-01265-7

dc.subject

LENA system

dc.subject

LENA system reliability

dc.subject

Talker variability

dc.title

Look who's talking: A comparison of automated and human-generated speaker tags in naturalistic day-long recordings.

dc.type

Journal article

duke.contributor.orcid

Bergelson, Elika|0000-0003-2742-4797

pubs.organisational-group

Trinity College of Arts & Sciences

pubs.organisational-group

Duke

pubs.organisational-group

Psychology and Neuroscience

pubs.organisational-group

Linguistics

pubs.organisational-group

Duke Institute for Brain Sciences

pubs.organisational-group

University Institutes and Centers

pubs.organisational-group

Institutes and Provost's Academic Units

pubs.organisational-group

Surgery, Head and Neck Surgery and Communication Sciences

pubs.organisational-group

Surgery

pubs.organisational-group

Clinical Science Departments

pubs.organisational-group

School of Medicine

pubs.publication-status

Published

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Bulgarelli-Bergelson2019_Article_LookWhoSTalkingAComparisonOfAu.pdf
Size:
453.28 KB
Format:
Adobe Portable Document Format
Description:
Published version