Discriminative training for near-synonym substitution

Liang Chih Yu, Hsiu Min Shih, Yu Ling Lai, Jui Feng Yeh, Chung Hsien Wu

研究成果: Paper同行評審

8 引文 斯高帕斯(Scopus)


Near-synonyms are useful knowledge resources for many natural language applications such as query expansion for information retrieval (IR) and paraphrasing for text generation. However, near-synonyms are not necessarily interchangeable in contexts due to their specific usage and syntactic constraints. Accordingly, it is worth to develop algorithms to verify whether near-synonyms do match the given contexts. In this paper, we consider the near-synonym substitution task as a classification task, where a classifier is trained for each near-synonym set to classify test examples into one of the near-synonyms in the set. We also propose the use of discriminative training to improve classifiers by distinguishing positive and negative features for each nearsynonym. Experimental results show that the proposed method achieves higher accuracy than both pointwise mutual information (PMI) and n-gram-based methods that have been used in previous studies.

出版狀態Published - 2010
事件23rd International Conference on Computational Linguistics, Coling 2010 - Beijing, China
持續時間: 2010 8月 232010 8月 27


Other23rd International Conference on Computational Linguistics, Coling 2010

All Science Journal Classification (ASJC) codes

  • 語言與語言學
  • 計算機理論與數學
  • 語言和語言學


深入研究「Discriminative training for near-synonym substitution」主題。共同形成了獨特的指紋。