본문 바로가기 주메뉴 바로가기
국회도서관 홈으로 정보검색 소장정보 검색

초록보기

This paper presents the development of language tutoring systems for nonnative speakers by leveraging advanced end-to-end automatic speech recognition (ASR) and proficiency evaluation. Given the frequent errors in non-native speech, high-performance spontaneous speech recognition must be applied. Our systems accurately evaluate pronunciation and speaking fluency and provide feedback on errors by relying on precise transcriptions. End-to-end ASR is implemented and enhanced by using diverse non-native speaker speech data for model training. For performance enhancement, we combine semisupervised and transfer learning techniques using labeled and unlabeled speech data. Automatic proficiency evaluation is performed by a model trained to maximize the statistical correlation between the fluency score manually determined by a human expert and a calculated fluency score. We developed an English tutoring system for Korean elementary students called EBS AI Peng-Talk and a Korean tutoring system for foreigners called KSI Korean AI Tutor. Both systems were deployed by South Korean government agencies.

권호기사

권호기사 목록 테이블로 기사명, 저자명, 페이지, 원문, 기사목차 순으로 되어있습니다.
기사명 저자명 페이지 원문 목차
Towards a small language model powered chain-of-reasoning for open-domain question answering Jihyeon Roh, Minho Kim, Kyoungman Bae p. 11-21

Multimodal audiovisual speech recognition architecture using a three-feature multi-fusion method for noise-robust systems Sanghun Jeon, Jieun Lee, Dohyeon Yeo, Yong-Ju Lee, SeungJun Kim p. 22-34

CR-M-SpanBERT : multiple embedding-based DNN coreference resolution using self-attention SpanBERT Joon-young Jung p. 35-47

AI-based language tutoring systems with end-to-end automatic speech recognition and proficiency evaluation Byung Ok Kang, Hyung-Bae Jeon, Yun Kyung Lee p. 48-58

Named entity recognition using transfer learning and small human- and meta-pseudo-labeled datasets Kyoungman Bae, Joon-Ho Lim p. 59-70

KMSAV : Korean multi-speaker spontaneous audiovisual dataset Kiyoung Park, Changhan Oh, Sunghee Dong p. 71-81

Dual-scale BERT using multi-trait representations for holistic and trait-specific essay grading Minsoo Cho, Jin-Xia Huang, Oh-Woog Kwon p. 82-95

Alzheimer’s disease recognition from spontaneous speech using large language models Jeong-Uk Bang, Seung-Hoon Han, Byung-Ok Kang p. 96-105

Framework for evaluating code generation ability of large language models Sangyeop Yeo, Yu-Seung Ma, Sang Cheol Kim, Hyungkook Jun, Taeho Kim p. 106-117

Joint streaming model for backchannel prediction and automatic speech recognition Yong-Seok Choi, Jeong-Uk Bang, Seung Hi Kim p. 118-126

Spoken-to-written text conversion for enhancement of Korean–English readability and machine translation HyunJung Choi, Muyeol Choi, Seonhui Kim, Yohan Lim, Minkyu Lee, Seung Yun, Donghyun Kim, Sang Hun Kim p. 127-136

Transformer-based reranking for improving Korean morphological analysis systems Jihee Ryu, Soojong Lim, Oh-Woog Kwon, Seung-Hoon Na p. 137-153

참고문헌 (35건) : 자료제공( 네이버학술정보 )

참고문헌 목록에 대한 테이블로 번호, 참고문헌, 국회도서관 소장유무로 구성되어 있습니다.
번호 참고문헌 국회도서관 소장유무
1 SRILM - an extensible language modeling toolkit 미소장
2 Lightly supervised and unsupervised acoustic model training 미소장
3 KL-divergence regularized deep neural network adaptation for improved large vocabulary speech recognition 미소장
4 Hybrid speech recognition with Deep Bidirectional LSTM 미소장
5 Language Model Adaptation based on Topic Probability of Latent Dirichlet Allocation 미소장
6 GenieTutor: A Computer-Assisted Second-Language Learning System Based on Spoken Language Understanding 미소장
7 Speaker Adaptation for Multichannel End-to-End Speech Recognition 미소장
8 Discussing with a computer to practice a foreign language: research synthesis and conceptual framework of dialogue-based CALL 미소장
9 Semi-supervised Training for End-to-end Models via Weak Distillation 미소장
10 ESPnet: End-to-End Speech Processing Toolkit 미소장
11 An Analysis of Incorporating an External Language Model into a Sequence-to-Sequence Model 미소장
12 Cycle-consistency Training for End-to-end Speech Recognition 미소장
13 Improving Transformer-Based End-to-End Speech Recognition with Connectionist Temporal Classification and Language Model Integration 미소장
14 Transformer-Based Online CTC/Attention End-To-End Speech Recognition Architecture 미소장
15 End-To-End Multi-Speaker Speech Recognition With Transformer 미소장
16 Automatic proficiency assessment of Korean speech read aloud by non‐natives using bidirectional LSTM‐based speech recognition 미소장
17 wav2vec 2.0: A Framework for Self-Supervised Learning of Speech Representations 미소장
18 Speech Recognition for Task Domains with Sparse Matched Training Data 미소장
19 Semi-supervised Training for Sequence-to-Sequence Speech Recognition Using Reinforcement Learning 미소장
20 Semi-Supervised ASR by End-to-End Self-Training 미소장
21 Improving Cross-Lingual Transfer Learning for End-to-End Speech Recognition with Speech Translation 미소장
22 Transformer-Based Long-Context End-to-End Speech Recognition 미소장
23 Conformer: Convolution-augmented Transformer for Speech Recognition 미소장
24 Multimodal Unsupervised Speech Translation for Recognizing and Evaluating Second Language Speech 미소장
25 Elsa Speak App: Automatic Speech Recognition (ASR) for Supplementing English Pronunciation Skills 미소장
26 HuBERT: Self-Supervised Speech Representation Learning by Masked Prediction of Hidden Units 미소장
27 Fast offline transformer‐based end‐to‐end automatic speech recognition for real‐world applications 미소장
28 Transformer-Based Multi-Aspect Multi-Granularity Non-Native English Speaker Pronunciation Assessment 미소장
29 English–Korean speech translation corpus (EnKoST‐C): Construction procedure and evaluation results 미소장
30 E-Branchformer: Branchformer with Enhanced Merging for Speech Recognition 미소장
31 Towards End-to-End Unsupervised Speech Recognition 미소장
32 Automated speech scoring of dialogue response by Japanese learners of English as a foreign language 미소장
33 Attention Is All You Need 미소장
34 The Use of ELSA Speak as a Mobile-Assisted Language Learning (MALL) towards EFL Students Pronunciation 미소장
35 Unsupervised versus supervised training of acoustic models 미소장