본문 바로가기 주메뉴 바로가기
국회도서관 홈으로 정보검색 소장정보 검색

초록보기

We introduce a high-performance named entity recognition (NER) model for written and spoken language. To overcome challenges related to labeled data scarcity and domain shifts, we use transfer learning to leverage our previously developed KorBERT as the base model. We also adopt a meta-pseudo-label method using a teacher/student framework with labeled and unlabeled data. Our model presents two modifications. First, the student model is updated with an average loss from both human- and pseudo-labeled data. Second, the influence of noisy pseudo-labeled data is mitigated by considering feedback scores and updating the teacher model only when below a threshold (0.0005). We achieve the target NER performance in the spoken language domain and improve that in the written language domain by proposing a straightforward rollback method that reverts to the best model based on scarce human-labeled data. Further improvement is achieved by adjusting the label vector weights in the named entity dictionary.

권호기사

권호기사 목록 테이블로 기사명, 저자명, 페이지, 원문, 기사목차 순으로 되어있습니다.
기사명 저자명 페이지 원문 목차
Towards a small language model powered chain-of-reasoning for open-domain question answering Jihyeon Roh, Minho Kim, Kyoungman Bae p. 11-21

Multimodal audiovisual speech recognition architecture using a three-feature multi-fusion method for noise-robust systems Sanghun Jeon, Jieun Lee, Dohyeon Yeo, Yong-Ju Lee, SeungJun Kim p. 22-34

CR-M-SpanBERT : multiple embedding-based DNN coreference resolution using self-attention SpanBERT Joon-young Jung p. 35-47

AI-based language tutoring systems with end-to-end automatic speech recognition and proficiency evaluation Byung Ok Kang, Hyung-Bae Jeon, Yun Kyung Lee p. 48-58

Named entity recognition using transfer learning and small human- and meta-pseudo-labeled datasets Kyoungman Bae, Joon-Ho Lim p. 59-70

KMSAV : Korean multi-speaker spontaneous audiovisual dataset Kiyoung Park, Changhan Oh, Sunghee Dong p. 71-81

Dual-scale BERT using multi-trait representations for holistic and trait-specific essay grading Minsoo Cho, Jin-Xia Huang, Oh-Woog Kwon p. 82-95

Alzheimer’s disease recognition from spontaneous speech using large language models Jeong-Uk Bang, Seung-Hoon Han, Byung-Ok Kang p. 96-105

Framework for evaluating code generation ability of large language models Sangyeop Yeo, Yu-Seung Ma, Sang Cheol Kim, Hyungkook Jun, Taeho Kim p. 106-117

Joint streaming model for backchannel prediction and automatic speech recognition Yong-Seok Choi, Jeong-Uk Bang, Seung Hi Kim p. 118-126

Spoken-to-written text conversion for enhancement of Korean–English readability and machine translation HyunJung Choi, Muyeol Choi, Seonhui Kim, Yohan Lim, Minkyu Lee, Seung Yun, Donghyun Kim, Sang Hun Kim p. 127-136

Transformer-based reranking for improving Korean morphological analysis systems Jihee Ryu, Soojong Lim, Oh-Woog Kwon, Seung-Hoon Na p. 137-153

참고문헌 (27건) : 자료제공( 네이버학술정보 )

참고문헌 목록에 대한 테이블로 번호, 참고문헌, 국회도서관 소장유무로 구성되어 있습니다.
번호 참고문헌 국회도서관 소장유무
1 Self-training from labeled features for sentiment analysis 미소장
2 Importance weighting and unsupervised domain adaptation of POS taggers: a negative result 미소장
3 Domain Adaptation via Transfer Component Analysis 미소장
4 Effective self-training for parsing 미소장
5 Named Entity Recognition for Novel Types by Transfer Learning 미소장
6 Empower Sequence Labeling with Task-Aware Neural Language Model 미소장
7 A Multi-lingual Multi-task Architecture for Low-resource Sequence Labeling 미소장
8 Natural Language Processing (almost) from Scratch 미소장
9 Transfer Learning in Natural Language Processing 미소장
10 End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF 미소장
11 Label-Aware Double Transfer Learning for Cross-Specialty Medical Named Entity Recognition 미소장
12 Dynamic Data Selection for Neural Machine Translation 미소장
13 A Survey on Deep Learning for Named Entity Recognition 미소장
14 MetaNER: Named Entity Recognition with Meta-Learning 미소장
15 Dice Loss for Data-imbalanced NLP Tasks 미소장
16 A Small-Scale Korean-Specific BERT Language Model 미소장
17 Entity Enhanced BERT Pre-training for Chinese NER 미소장
18 LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention 미소장
19 EER-ASSL: Combining Rollback Learning and Deep Learning for Rapid Adaptive Object Detection 미소장
20 A Decade Survey of Transfer Learning (2010–2020) 미소장
21 Learning from Noisy Labels for Entity-Centric Information Extraction 미소장
22 Few-shot Learning for Named Entity Recognition Based on BERT and Two-level Model Fusion 미소장
23 Meta Self-training for Few-shot Neural Sequence Labeling 미소장
24 Improving Named Entity Recognition by External Context Retrieving and Cooperative Learning 미소장
25 Meta Pseudo Labels 미소장
26 A Survey on Deep Semi-Supervised Learning 미소장
27 Meta-Based Self-Training and Re-Weighting for Aspect-Based Sentiment Analysis 미소장