earticle

논문검색

Subword Neural Language Generation with Unlikelihood Training

초록

영어

A Language model with neural networks commonly trained with likelihood loss. Such that the model can learn the sequence of human text. State-of-the-art results achieved in various language generation tasks, e.g., text summarization, dialogue response generation, and text generation, by utilizing the language model’s next token output probabilities. Monotonous and boring outputs are a well-known problem of this model, yet only a few solutions proposed to address this problem. Several decoding techniques proposed to suppress repetitive tokens. Unlikelihood training approached this problem by penalizing candidate tokens probabilities if the tokens already seen in previous steps. While the method successfully showed a less repetitive generated token, the method has a large memory consumption because of the training need a big vocabulary size. We effectively reduced memory footprint by encoding words as sequences of subword units. Finally, we report competitive results with token level unlikelihood training in several automatic evaluations compared to the previous work.

목차

Abstract
1. INTRODUCTION
2. METHODOLOGIES
2.1 Subword Tokenization
2.2 Unlikelihood Training
3. EXPERIMENTS
3.1 Experiment Setups
3.2 Evaluation Metrics
4. RESULTS AND DISCUSSIONS
5. CONCLUSION
Acknowledgment
References

저자정보

  • Salahuddin Muhammad Iqbal Master Student, Department of Computer Engineering, Dongseo University, Korea
  • Dae-Ki Kang Professor, Department of Computer Engineering, Dongseo University, Korea

참고문헌

자료제공 : 네이버학술정보

    함께 이용한 논문

      ※ 기관로그인 시 무료 이용이 가능합니다.

      • 4,000원

      0개의 논문이 장바구니에 담겼습니다.