earticle

논문검색

Subword Neural Language Generation with Unlikelihood Training

초록

영어

A Language model with neural networks commonly trained with likelihood loss. Such that the model can learn the sequence of human text. State-of-the-art results achieved in various language generation tasks, e.g., text summarization, dialogue response generation, and text generation, by utilizing the language model’s next token output probabilities. Monotonous and boring outputs are a well-known problem of this model, yet only a few solutions proposed to address this problem. Several decoding techniques proposed to suppress repetitive tokens. Unlikelihood training approached this problem by penalizing candidate tokens probabilities if the tokens already seen in previous steps. While the method successfully showed a less repetitive generated token, the method has a large memory consumption because of the training need a big vocabulary size. We effectively reduced memory footprint by encoding words as sequences of subword units. Finally, we report competitive results with token level unlikelihood training in several automatic evaluations compared to the previous work.

목차

Abstract
1. INTRODUCTION
2. METHODOLOGIES
2.1 Subword Tokenization
2.2 Unlikelihood Training
3. EXPERIMENTS
3.1 Experiment Setups
3.2 Evaluation Metrics
4. RESULTS AND DISCUSSIONS
5. CONCLUSION
Acknowledgment
References

저자정보

  • Salahuddin Muhammad Iqbal Master Student, Department of Computer Engineering, Dongseo University, Korea
  • Dae-Ki Kang Professor, Department of Computer Engineering, Dongseo University, Korea

참고문헌

자료제공 : 네이버학술정보

    함께 이용한 논문

      ※ 원문제공기관과의 협약기간이 종료되어 열람이 제한될 수 있습니다.

      0개의 논문이 장바구니에 담겼습니다.