Advancing Linguistic E-Learning with AI Innovations

In a recent publication in the journal Applied Sciences, researchers augmented the existing linguistic e-learning system for phonetic transcription with three artificial intelligence (AI)-driven enhancements.

Study: Advancing Linguistic E-Learning with AI Innovations. Image credit: Song_about_summer/Shutterstock
Study: Advancing Linguistic E-Learning with AI Innovations. Image credit: Song_about_summer/Shutterstock

Background

Phonetic transcription, a crucial component of linguistic education, employs International Phonetic Alphabet (IPA) characters to represent speech sounds. In their prior work, the authors introduced Automated Phonetic Transcription—The Grading Tool (APTgt), an interactive e-learning system for phonetic transcription and pronunciation. This system, tailored for IPA language students, offers phonetic transcription exams and automated grading for teachers. To enhance its intelligence and versatility, the authors propose three machine learning and deep learning-based improvements.

Empowering Linguistics Through E-Learning and AI Advancements

Amid the coronavirus disease 2019 (COVID-19) pandemic, e-learning saw a surge, affecting over one billion children globally, as per United Nations Educational, Scientific, and Cultural Organization (UNESCO) data. E-learning employs electronic devices such as smartphones and laptops to provide interactive, remote learning, offering flexibility and overcoming geographical constraints.

Linguistics, the study of human language, benefits from e-learning. APTgt, an interactive web-based system, focuses on phonetic transcription, representing speech sounds using unique symbols. The speech disorder classification process involves feature extraction and classification. Researchers extracted the Mel frequency cepstrum coefficient (MFCC) and linear prediction coding (LPC) features. These extracted features were used for the classification task.

A phoneme is the smallest spoken sound unit, while a grapheme denotes the smallest written language unit in linguistics. Words are converted from graphemes to phonemes (G2P), essential for natural language processing applications such as text-to-speech (TTS) and speech recognition. Hidden Markov and deep learning models have been investigated, with noticeable advancements in performance over time. Neural network-based end-to-end models, including Tacotron 1 and 2, Fast Speech, and Deep Voice, have become prominent in TTS research, comprising text analysis, acoustic modeling, and vocoding components.

Enhancements to Improve Linguistic E-Learning

MFCC and CNN-based Disorder Speech Classification: The initial enhancement introduced is a speech classification module that can differentiate between disordered and non-disordered speech. This classification entails two fundamental subproblems: feature extraction and classification. Researchers considered MFCC in image format speech features, and a CNN model handled classification. Feature extraction involved converting audio signals into MFCCs, utilizing MFC as a representation. The process included pre-emphasizing the audio, segmenting it into overlapping windows, Fourier transforming to the frequency domain, computing the Mel spectrum, and applying discrete cosine transformation to yield MFCCs.

Data for model training were sourced from the Speech Exemplar and Evaluation Database (SEED), comprising over 16,000 speech samples categorized by age and speech health status, both with and without speech disorders.

Implementation involved selecting about 1000 SEED samples, splitting them into training and validation sets, and processing MFCC values using Python's Librosa library. The disordered speech classification problem was translated into an image classification challenge addressed by a CNN model. The model achieved an average classification accuracy of approximately 83 percent, indicating efficiency and versatility in distinguishing disordered speech, irrespective of content or recorder.

Transformer-based Multilingual G2P Converter: The primary function of their e-learning system revolves around interactive phonetic transcription exams, posing challenges for teachers when generating IPA characters from written language. Hence, they devised a G2P converter. G2P conversion involves translating words into IPA format, which is essential for phonetic transcription. They opted for IPA symbols to represent pronunciation in their system and explored French and Spanish variants for multilingual support. Utilizing various datasets, they employed a Transformer model, which processes input comprehensively, avoiding dependency issues.

The six-layer transformer model achieved remarkable results. The English G2P converter yielded a 2.6 percent phoneme error rate (PER) and a 10.7 percent word error rate (WER). The French and Spanish converters attained PERs of 2.1 percent and 1.7 percent and WERs of 12.3 percent and 12.7 percent, respectively. Their models outperformed other G2P converters, showcasing their superiority in accuracy. This transformer-based G2P converter enhances phonetic transcription and presents potential for multilingual e-learning, streamlining the generation of IPA characters for teachers.

Tacotron2-Based IPA-to-Speech System: The existing e-learning system faces the challenge of acquiring high-quality speech audio and converting text into IPA format for phonetic transcription exams. To address this, researchers designed a TTS system capable of generating speech directly from IPA-formatted text.

Their IPA-to-speech system workflow involves converting English sentences from the LJSpeech dataset into IPA format using their G2P converter. They then predict and calculate Mel spectrograms with the Tacotron 2 model and employ WaveGlow as the vocoder, ensuring high-quality speech sound generation. The APTgt system employs Tacotron2, which uses simpler building blocks and predicts Mel spectrogram frames from character input sequences. A modified WaveNet generates waveform samples based on the predicted Mel spectrograms. This Tacotron2-based IPA-to-speech system offers teachers a solution to the challenges of recording speech and enhances students' understanding of IPA characters and word pronunciation.

Conclusion

In summary, researchers introduced three AI enhancements for the existing linguistic e-learning system. These enhancements elevate the system's functionality, making it more comprehensive. They improve speech classification, G2P conversion, and speech synthesis. Future work includes objective system evaluation and addressing potential privacy concerns related to recorded speech sounds.

Journal reference:
Dr. Sampath Lonka

Written by

Dr. Sampath Lonka

Dr. Sampath Lonka is a scientific writer based in Bangalore, India, with a strong academic background in Mathematics and extensive experience in content writing. He has a Ph.D. in Mathematics from the University of Hyderabad and is deeply passionate about teaching, writing, and research. Sampath enjoys teaching Mathematics, Statistics, and AI to both undergraduate and postgraduate students. What sets him apart is his unique approach to teaching Mathematics through programming, making the subject more engaging and practical for students.

Citations

Please use one of the following formats to cite this article in your essay, paper or report:

  • APA

    Lonka, Sampath. (2023, September 29). Advancing Linguistic E-Learning with AI Innovations. AZoAi. Retrieved on November 21, 2024 from https://www.azoai.com/news/20230929/Advancing-Linguistic-E-Learning-with-AI-Innovations.aspx.

  • MLA

    Lonka, Sampath. "Advancing Linguistic E-Learning with AI Innovations". AZoAi. 21 November 2024. <https://www.azoai.com/news/20230929/Advancing-Linguistic-E-Learning-with-AI-Innovations.aspx>.

  • Chicago

    Lonka, Sampath. "Advancing Linguistic E-Learning with AI Innovations". AZoAi. https://www.azoai.com/news/20230929/Advancing-Linguistic-E-Learning-with-AI-Innovations.aspx. (accessed November 21, 2024).

  • Harvard

    Lonka, Sampath. 2023. Advancing Linguistic E-Learning with AI Innovations. AZoAi, viewed 21 November 2024, https://www.azoai.com/news/20230929/Advancing-Linguistic-E-Learning-with-AI-Innovations.aspx.

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of AZoAi.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
Machine Learning Powering Breakthroughs in Climate Forecasting and Modeling