Skip to main content
Start main content

Conference Paper Published

Rearch

PhonoThink: Improving Large Language Models’ Reasoning on Chinese Phonological Ambiguities

Ma, J., Feng, Z., Chersoni, E., Song, H., & Zhang, Z. (2025). PhonoThink: Improving Large Language Models' Reasoning on Chinese Phonological Ambiguities. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP 2025), 19018-19033.
 
DOI:  https://doi.org/10.18653/v1/2025.emnlp-main.961

 

Abstract

Effectively resolving phonological ambiguities is crucial for robust natural language processing, as these ambiguities are pervasive in tasks ranging from speech-to-text, spelling correction, to offensive language detection. However, current Large Language Models (LLMs) frequently struggle to resolve such ambiguities.To address this challenge, we present a framework to enhances LLMs’ phonological capability through a multiple-stage training approach. Our method begins with supervised fine-tuning on well-constructed datasets, including three subtask datasets designed to enhance the model’s foundational phonological knowledge, along with a synthetic dataset of step-by-step reasoning chains. Following this, we apply reinforcement learning to incentivize and stabilize its reasoning.Results show that our framework enables the base model to achieve relatively comparable performance to a much larger model. Our ablation studies reveal that subtask datasets and the synthetic dataset can simultaneously impact as complementary modular enhancers to strengthen LLMs’ integrated application.

 

 

 

 








Your browser is not the latest version. If you continue to browse our website, Some pages may not function properly.

You are recommended to upgrade to a newer version or switch to a different browser. A list of the web browsers that we support can be found here