Comparison of native and nonnative speakers' perspective in animated text visualization tool

Nur Syafikah Binti Samsudin, Kazunori Mano

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Media conversion technologies such as speech recognition and speech synthesis have gained a lot of attention in recent years. They are applied in various human communication tools through smart phones and personal computers such as in language learning system. However, the learner still have loss in naturalness pronunciation problem because of effects on mother tongue interference. Although the learner can realize that his/her speech is different from the trainer's, the learner still cannot detect or check the precise wrong part in their utterances speech. We indicate these differences by visualizing a learner's wrong and correct pronunciation with speech-to-animated text visualization tool. In this study, we focused on the media conversion process between speech prosodic information and animated text information by using AHP method as the mapping method. The pairwise comparison between both speech and text information were conducted and evaluated by Japanese native and nonnative speakers. Here, we investigated the comparison between native and nonnative speakers' perspective and determined the ideal matching elements between attributes of speech and text information.

LanguageEnglish
Title of host publicationTENCON 2015 - 2015 IEEE Region 10 Conference
PublisherInstitute of Electrical and Electronics Engineers Inc.
Volume2016-January
ISBN (Electronic)9781479986415
DOIs
StatePublished - 2016 Jan 5
Event35th IEEE Region 10 Conference, TENCON 2015 - Macau, Macao
Duration: 2015 Nov 12015 Nov 4

Other

Other35th IEEE Region 10 Conference, TENCON 2015
CountryMacao
CityMacau
Period15/11/115/11/4

Fingerprint

Visualization
Speech synthesis
Speech recognition
Personal computers
Learning systems
Communication

Keywords

  • Animated text
  • Nonlinguistic information
  • Paralinguistic information
  • Speech visualization

ASJC Scopus subject areas

  • Electrical and Electronic Engineering
  • Computer Science Applications

Cite this

Binti Samsudin, N. S., & Mano, K. (2016). Comparison of native and nonnative speakers' perspective in animated text visualization tool. In TENCON 2015 - 2015 IEEE Region 10 Conference (Vol. 2016-January). [7372934] Institute of Electrical and Electronics Engineers Inc.. DOI: 10.1109/TENCON.2015.7372934

Comparison of native and nonnative speakers' perspective in animated text visualization tool. / Binti Samsudin, Nur Syafikah; Mano, Kazunori.

TENCON 2015 - 2015 IEEE Region 10 Conference. Vol. 2016-January Institute of Electrical and Electronics Engineers Inc., 2016. 7372934.

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Binti Samsudin, NS & Mano, K 2016, Comparison of native and nonnative speakers' perspective in animated text visualization tool. in TENCON 2015 - 2015 IEEE Region 10 Conference. vol. 2016-January, 7372934, Institute of Electrical and Electronics Engineers Inc., 35th IEEE Region 10 Conference, TENCON 2015, Macau, Macao, 15/11/1. DOI: 10.1109/TENCON.2015.7372934
Binti Samsudin NS, Mano K. Comparison of native and nonnative speakers' perspective in animated text visualization tool. In TENCON 2015 - 2015 IEEE Region 10 Conference. Vol. 2016-January. Institute of Electrical and Electronics Engineers Inc.2016. 7372934. Available from, DOI: 10.1109/TENCON.2015.7372934
Binti Samsudin, Nur Syafikah ; Mano, Kazunori. / Comparison of native and nonnative speakers' perspective in animated text visualization tool. TENCON 2015 - 2015 IEEE Region 10 Conference. Vol. 2016-January Institute of Electrical and Electronics Engineers Inc., 2016.
@inproceedings{ff958346274d49b09bf56156e671e4f5,
title = "Comparison of native and nonnative speakers\{textquoteleft} perspective in animated text visualization tool",
abstract = "Media conversion technologies such as speech recognition and speech synthesis have gained a lot of attention in recent years. They are applied in various human communication tools through smart phones and personal computers such as in language learning system. However, the learner still have loss in naturalness pronunciation problem because of effects on mother tongue interference. Although the learner can realize that his/her speech is different from the trainer\{textquoteleft}s, the learner still cannot detect or check the precise wrong part in their utterances speech. We indicate these differences by visualizing a learner\{textquoteleft}s wrong and correct pronunciation with speech-to-animated text visualization tool. In this study, we focused on the media conversion process between speech prosodic information and animated text information by using AHP method as the mapping method. The pairwise comparison between both speech and text information were conducted and evaluated by Japanese native and nonnative speakers. Here, we investigated the comparison between native and nonnative speakers\{textquoteleft} perspective and determined the ideal matching elements between attributes of speech and text information.",
keywords = "Animated text, Nonlinguistic information, Paralinguistic information, Speech visualization",
author = "{Binti Samsudin}, {Nur Syafikah} and Kazunori Mano",
year = "2016",
month = "1",
day = "5",
doi = "10.1109/TENCON.2015.7372934",
language = "English",
volume = "2016-January",
booktitle = "TENCON 2015 - 2015 IEEE Region 10 Conference",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
address = "United States",

}

TY - GEN

T1 - Comparison of native and nonnative speakers' perspective in animated text visualization tool

AU - Binti Samsudin,Nur Syafikah

AU - Mano,Kazunori

PY - 2016/1/5

Y1 - 2016/1/5

N2 - Media conversion technologies such as speech recognition and speech synthesis have gained a lot of attention in recent years. They are applied in various human communication tools through smart phones and personal computers such as in language learning system. However, the learner still have loss in naturalness pronunciation problem because of effects on mother tongue interference. Although the learner can realize that his/her speech is different from the trainer's, the learner still cannot detect or check the precise wrong part in their utterances speech. We indicate these differences by visualizing a learner's wrong and correct pronunciation with speech-to-animated text visualization tool. In this study, we focused on the media conversion process between speech prosodic information and animated text information by using AHP method as the mapping method. The pairwise comparison between both speech and text information were conducted and evaluated by Japanese native and nonnative speakers. Here, we investigated the comparison between native and nonnative speakers' perspective and determined the ideal matching elements between attributes of speech and text information.

AB - Media conversion technologies such as speech recognition and speech synthesis have gained a lot of attention in recent years. They are applied in various human communication tools through smart phones and personal computers such as in language learning system. However, the learner still have loss in naturalness pronunciation problem because of effects on mother tongue interference. Although the learner can realize that his/her speech is different from the trainer's, the learner still cannot detect or check the precise wrong part in their utterances speech. We indicate these differences by visualizing a learner's wrong and correct pronunciation with speech-to-animated text visualization tool. In this study, we focused on the media conversion process between speech prosodic information and animated text information by using AHP method as the mapping method. The pairwise comparison between both speech and text information were conducted and evaluated by Japanese native and nonnative speakers. Here, we investigated the comparison between native and nonnative speakers' perspective and determined the ideal matching elements between attributes of speech and text information.

KW - Animated text

KW - Nonlinguistic information

KW - Paralinguistic information

KW - Speech visualization

UR - http://www.scopus.com/inward/record.url?scp=84962199661&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=84962199661&partnerID=8YFLogxK

U2 - 10.1109/TENCON.2015.7372934

DO - 10.1109/TENCON.2015.7372934

M3 - Conference contribution

VL - 2016-January

BT - TENCON 2015 - 2015 IEEE Region 10 Conference

PB - Institute of Electrical and Electronics Engineers Inc.

ER -