‘Um, so like, is this how I speak?’: design implications for automated visual feedback systems on speech

Soonho Kwon, Hyojin Lee, Soojeong Park, Yerin Heo, Keeheon Lee, Younah Kang

Research output: Contribution to journalArticlepeer-review

Abstract

At a time when diverse forms of automated visual feedback systems on speech are introduced to the Human–Computer Interaction (HCI) community, our research investigates the possible design spaces and design implications to consider when creating them. We utilised a Wizard of Oz prototype that recognises fillers during speech and sends out real-time feedback in the form of a flickering screen and a post-report with graphic charts. We let 27 participants use this prototype and interviewed them about their general experience of using such systems. Based on thematic analysis, we propose eight design implications to consider when designing an automated visual feedback system on speech. Additionally, we discuss the possible future for automated visual feedback systems on speech where they collaborate with fields such as personal informatics (quantified self), self-determination theory (SDT) and motivation, and other socio-ethnic areas.

Original languageEnglish
Pages (from-to)3091-3110
Number of pages20
JournalBehaviour and Information Technology
Volume43
Issue number13
DOIs
Publication statusPublished - 2024

Bibliographical note

Publisher Copyright:
© 2023 Informa UK Limited, trading as Taylor & Francis Group.

All Science Journal Classification (ASJC) codes

  • Developmental and Educational Psychology
  • Arts and Humanities (miscellaneous)
  • General Social Sciences
  • Human-Computer Interaction

Fingerprint

Dive into the research topics of '‘Um, so like, is this how I speak?’: design implications for automated visual feedback systems on speech'. Together they form a unique fingerprint.

Cite this