The papers were implemented in using korean corpus
- Using the Naver sentiment movie corpus v1.0
- Hyper-parameter was arbitrarily selected. (epoch: 5, mini_batch: 128)
Train ACC (120,000) | Validation ACC (30,000) | Test ACC (50,000) | |
---|---|---|---|
SenCNN | 92.87% | 86.87% | 86.38% |
CharCNN | 85.63% | 81.58% | 81.58% |
ConvRec | 86.80% | 82.66% | 82.29% |
VDCNN | 86.31% | 83.87% | 83.90% |
SAN | 93.90% | 86.52% | 86.35% |
- Convolutional Neural Networks for Sentence Classification (as SenCNN)
- Character-level Convolutional Networks for Text Classification (as CharCNN)
- Efficient Character-level Document Classification by Combining Convolution and Recurrent Layers (as ConvRec)
- Very Deep Convolutional Networks for Text Classification (as VDCNN)
- A Structured Self-attentive Sentence Embedding (as SAN)
- Creating dataset from https://github.com/songys/Question_pair
- Hyper-parameter was arbitrarily selected. (epoch: 5, mini_batch: 64)
Train ACC (6,060) | Validation ACC (1,516) | |
---|---|---|
SAN | 91.93% | 81.46% |
- Character-Aware Neural Language Models
- Using the Naver nlp-challange corpus for NER
- Hyper-parameter was arbitrarily selected.
- Bidirectional LSTM-CRF Models for Sequence Tagging
- End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF
- Neural Architectures for Named Entity Recognition
- Effective Approaches to Attention-based Neural Machine Translation
- Attention Is All You Need
- Bi-directional attention flow for machine comprehension
- Deep contextualized word representations
- Improving Language Understanding by Generative Pre-Training
- BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
- Language Models are Unsupervised Multitask Learners