Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 41 tok/s Pro
GPT-5 High 39 tok/s Pro
GPT-4o 89 tok/s Pro
Kimi K2 192 tok/s Pro
GPT OSS 120B 437 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

CharPoet: A Chinese Classical Poetry Generation System Based on Token-free LLM (2401.03512v3)

Published 7 Jan 2024 in cs.CL, cs.AI, and cs.LG

Abstract: Automatic Chinese classical poetry generation has attracted much research interest, but achieving effective control over format and content simultaneously remains challenging. Traditional systems usually accept keywords as user inputs, resulting in limited control over content. LLMs improve content control by allowing unrestricted user instructions, but the token-by-token generation process frequently makes format errors. Motivated by this, we propose CharPoet, a Chinese classical poetry generation system based on token-free LLM, which provides effective control over both format and content. Our token-free architecture generates in a character-by-character manner, enabling precise control over the number of characters. Pruned from existing token-based LLMs, CharPoet inherits their pretrained capabilities and can generate poetry following instructions like "Write me a poem for my mother's birthday." CharPoet achieves format accuracy above 0.96, outperforming Jiuge-GPT-2 (0.91) and GPT-4 (0.38). In terms of content quality, CharPoet surpasses traditional systems including Jiuge, and is comparable to other LLMs. Our system is open source and available at https://modelscope.cn/models/CharPoet/CharPoet. A video demonstration of CharPoet is available at https://youtu.be/voZ25qEp3Dc.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (16)
  1. Qwen Technical Report. arXiv:2309.16609 [cs.CL]
  2. Jonas Belouadi and Steffen Eger. 2023. ByGPT5: End-to-End Style-conditioned Poetry Generation with Token-free Language Models. In 61st Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics (ACL).
  3. Jinyi Hu and Maosong Sun. 2020. Generating Major Types of Chinese Classical Poetry in a Uniformed Framework. In 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics (ACL).
  4. Itay Itzhak and Omer Levy. 2022. Models In a Spelling Bee: Language Models Implicitly Learn the Character Composition of Tokens. In 60th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics (ACL).
  5. Ayush Kaushal and Kyle Mahowald. 2022. What do tokens know about their characters and how do they know it?. In 60th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics (ACL).
  6. OpenAI. 2016. Introducing ChatGPT. https://openai.com/blog/chatgpt
  7. Training language models to follow instructions with human feedback. arXiv:2203.02155 [cs.CL]
  8. Mike Schuster and Kaisuke Nakajima. 2012. Japanese and Korean Voice Search. In International Conference on Acoustics, Speech and Signal Processing. 5149–5152.
  9. Neural Machine Translation of Rare Words with Subword Units. In 54th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics (ACL), 1715–1725.
  10. Chinese Poetry Generation with Planning based Neural Network. In Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers. 1051–1060.
  11. Rui Yan. 2016. i, poet: automatic poetry composition through recurrent neural networks with iterative polishing schema. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence. 2238–2244.
  12. Generating chinese classical poems with rnn encoder-decoder. In Chinese Computational Linguistics and Natural Language Processing Based on Naturally Annotated Big Data: 16th China National Conference, CCL 2017, and 5th International Symposium, NLP-NABD 2017, Nanjing, China, October 13-15, 2017, Proceedings 16. Springer, 211–223.
  13. Chinese poetry generation with a working memory model. In Proceedings of the 27th International Joint Conference on Artificial Intelligence. 4553–4559.
  14. Flexible and Creative Chinese Poetry Generation Using Neural Memory. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 1364–1373.
  15. Xingxing Zhang and Mirella Lapata. 2014. Chinese poetry generation with recurrent neural networks. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). 670–680.
  16. Jiuge: A human-machine collaborative chinese classical poetry generation system. In Proceedings of the 57th annual meeting of the association for computational linguistics: system demonstrations. 25–30.
Citations (2)

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 2 tweets and received 0 likes.

Upgrade to Pro to view all of the tweets about this paper: