A corpus-based approach to the Chinese word segmentation [Elektronische Ressource] / vorgelegt von Lezhong Liu
258 pages
English

Découvre YouScribe en t'inscrivant gratuitement

Je m'inscris

A corpus-based approach to the Chinese word segmentation [Elektronische Ressource] / vorgelegt von Lezhong Liu

Découvre YouScribe en t'inscrivant gratuitement

Je m'inscris
Obtenez un accès à la bibliothèque pour le consulter en ligne
En savoir plus
258 pages
English
Obtenez un accès à la bibliothèque pour le consulter en ligne
En savoir plus

Description

A Corpus-based Approach tothe Chinese Word SegmentationInaugural-Dissertationzur Erlangung des Doktorgradesder Philosophie an der Ludwig-Maximilians-UniversitätMünchenvorgelegt vonLezhong LiuCentrum für Information- und SprachverarbeitungReferent:Prof. Dr. Franz GünthnerKorreferent:Prof. Dr. Klaus U. SchulzTag der mündlichen Prüfung: 05.07.20051AcknowledgementsI would like to take this opportunity to express my graditute and appreciation foreach and every person who has made the work for this project possible.To Professor Franz Guenthner, for giving me an opportunity to work under hissupervision and for his unreserved support and guidance during the period of thisresearch project.To my fellow students Sebastian Nagel, Christopher Winestock, Johannes Goller,Felix Golcher, Aleksandra Wasiak, Li Yunbei, for providing technical support,helping me with translation and for lightening the load of work that has gone intothis thesis.To Professor Liu Qun, Dr. Zhang huaping, Xiong Deyi, Liu Yang, Chui Shiqi,Li Shuanglong from the Chinese Academy of Science, for providing me withcountless pieces of advice.To the Chinese Student Group at the CIS for their work in building the word list.To Jake Benilov for his sizeable contribution to the translation work.

Informations

Publié par
Publié le 01 janvier 2005
Nombre de lectures 16
Langue English
Poids de l'ouvrage 1 Mo

Extrait

A Corpus-based Approach to
the Chinese Word Segmentation
Inaugural-Dissertation
zur Erlangung des Doktorgrades
der Philosophie an der Ludwig-Maximilians-Universität
München
vorgelegt von
Lezhong Liu
Centrum für Information- und SprachverarbeitungReferent:
Prof. Dr. Franz Günthner
Korreferent:
Prof. Dr. Klaus U. Schulz
Tag der mündlichen Prüfung: 05.07.2005
1Acknowledgements
I would like to take this opportunity to express my graditute and appreciation for
each and every person who has made the work for this project possible.
To Professor Franz Guenthner, for giving me an opportunity to work under his
supervision and for his unreserved support and guidance during the period of this
research project.
To my fellow students Sebastian Nagel, Christopher Winestock, Johannes Goller,
Felix Golcher, Aleksandra Wasiak, Li Yunbei, for providing technical support,
helping me with translation and for lightening the load of work that has gone into
this thesis.
To Professor Liu Qun, Dr. Zhang huaping, Xiong Deyi, Liu Yang, Chui Shiqi,
Li Shuanglong from the Chinese Academy of Science, for providing me with
countless pieces of advice.
To the Chinese Student Group at the CIS for their work in building the word list.
To Jake Benilov for his sizeable contribution to the translation work.
To my parents, who give meaning to my life, and my father, for probably working
just as hard as I did to make this thesis possible, and for this I express my deepest
respects to him and my mother.
2Abstract
For a society based upon laws and reason, it has become too easy for us to believe
that we live in a world without them. And given that our linguistics wisdom was
originally motivated by the search for rules, it seems strange that we now consider
these rules to be the exceptions and take exceptions as the norm.
The current task of contemporary computational linguistics is to describe these
exceptions. In particular, it suffices for most language processing needs, to just
describe the argument and predicate within an elementary sentence, under the
framework of local grammar. Therefore, a corpus-based approach to the Chinese
Word Segmentation problem is proposed, as the first step towards a local grammar
for the Chinese language.
The two main issues with existing lexicon-based approaches are (a) the clas-
sification of unknown character sequences, i.e. sequences that are not listed in
the lexicon, and (b) the disambiguation of situations where two candidate words
overlap.
For (a), we propose an automatic method of enriching the lexicon by comparing
candidate sequences to occurrences of the same strings in a manually segmented
reference corpus, and using methods of machine learning to select the optimal
segmentation for them. These methods are developed in the course of the the-
sis specifically for this task. The possibility of applying these machine learning
method will be discussed in NP-extraction and alignment domain.
(b) is approached by designing a general processing framework for Chinese text,
which will be called multi-level processing. Under this framework, sentences are
recursively split into fragments, according to a language-specific, but domain-
independent heuristics. The resulting fragments then define the ultimate bound-
aries between candidate words and therefore resolve any segmentation ambiguity
caused by overlapping sequences. A new shallow semantical annotation is also
proposed under the frame work of multi-level processing.
A word segmentation algorithm based on these principles has been implemented
and tested; results of the evaluation are given and compared to the performance of
previous approaches as reported in the literature.
The first chapter of this thesis discusses the goals of segmentation and introduces
some background concepts. The second chapter analyses the current state-of-the-
art approach to Chinese language segmentation. Chapter 3 proposes a new corpus-
based approach to the identification of unknown words. In chapter 4, a new shal-
low semantical annotation is also proposed under the framework of multi-level
processing.
3Contents
1 Introduction 2
1.1 Comments on the Existing Standards . . . . . . . . . . . . . . . 3
1.1.1 The Beijing University Standard . . . . . . . . . . . . . . 3
1.1.2 The ROCLING Standard . . . . . . . . . . . . . . . . . . 8
1.2 The Goal of Chinese Word Segmentation . . . . . . . . . . . . . 9
1.2.1 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.3 The CIS’s Chinese Lexicon Construction . . . . . . . . . . . . . 15
1.4 Overview of the Recent Chinese Corpus in China . . . . . . . . . 19
2 State-of-the-art Segmentation 24
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
2.1.1 Description of the Unknown Word Problem and the Am-
biguity Problem . . . . . . . . . . . . . . . . . . . . . . . 25
2.2 The Chinese Word Segmentation as a Unified Problem does not
Exist . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
2.3 Segmentation as the Combination of Training Corpus and Methods 27
2.3.1 Training Corpus . . . . . . . . . . . . . . . . . . . . . . 27
2.3.2 Methods . . . . . . . . . . . . . . . . . . . . . . . . . . 27
2.3.3 HMM family: such as HMM, finite state automata, source
channel. . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
2.3.4 Transformation-based Learning . . . . . . . . . . . . . . 31
2.3.5 The Complicated statistical methods . . . . . . . . . . . . 33
2.3.6 Statistical enhanced Heuristic . . . . . . . . . . . . . . . 34
2.3.7 Iteration and Ranking Techniques . . . . . . . . . . . . . 34
2.4 Description of The Two Best Systems . . . . . . . . . . . . . . . 37
2.5 A Brief overview of the Chinese Word Tokenizer Developed at
the CIS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
2.6 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
3 Domain Adaptive Segmentation 41
3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
13.2 Unknown Word and Domain-specific . . . . . . . . . . . . . . . . 43
3.2.1 General vs. Domain-specific . . . . . . . . . . . . . . . . 47
3.3 System description . . . . . . . . . . . . . . . . . . . . . . . . . 47
3.4 Proposed method . . . . . . . . . . . . . . . . . . . . . . . . . . 48
3.4.1 Pure fragment filter . . . . . . . . . . . . . . . . . . . . . 50
3.4.2 Frequent single element filter . . . . . . . . . . . . . . . . 51
3.4.3 N-gram filter . . . . . . . . . . . . . . . . . . . . . . . . 53
3.5 Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
3.6 Examples of Segment Results Comparisons . . . . . . . . . . . . 56
3.6.1 Organization Name Comparison of Our System and ICT-
CLAS from the Chinese Academy . . . . . . . . . . . . . 56
3.6.2 Place Name Comparison of Our System and ICTCLAS
from the Chinese Academy . . . . . . . . . . . . . . . . . 57
3.6.3 Date Comparison of Our System and ICTCLAS from the
Chinese Academy . . . . . . . . . . . . . . . . . . . . . 58
3.6.4 Terminology Comparison of Our System and ICTCLAS
from Chinese Academy . . . . . . . . . . . . . . . . . . . 58
3.6.5 Running Text Comparison of Our System and ICTCLAS
from the Chinese Academy . . . . . . . . . . . . . . . . . 58
3.7 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
4 Segmentation in Chinese Language Processing 65
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
4.2 Comparison of Different Tagsets . . . . . . . . . . . . . . . . . . 67
4.2.1 From Chinese Grammar to Local Grammar . . . . . . . . 72
4.2.2 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . 74
4.3 A Semantical Tagset and its Application to the China Daily Corpus 75
4.3.1 A Semantical Tagset and its Application to China Daily
Corpus . . . . . . . . . . . . . . . . . . . . . . . . . . . 77
4.3.2 Our Annotation Applied to Two Complex Sentence Ex-
amples from the Original China Daily Corpus with Trans-
lations . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
4.3.3 Explanation of Our Annotation to the Two Complex Sen-
tences . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82
4.3.4 Our Annotation to an Example Text from the Original
China Daily Corpus with translations . . . . . . . . . . . 85
4.3.5 Two Category Assumption(TCA) and Multi-level Frame-
work for Disambiguity . . . . . . . . . . . . . . . . . . . 89
4.3.6 Multi-level Processing of Chinese Text with TCA . . . . . 90
2A Comparison of Our System and ICTCLAS 106
A.1 Result of Our System for Chemical Text . . . . . . . . . . . . . . 106
A.2 Result of ICTCLAS for Chemical Text . . . . . . . . . . . . . . . 163
B Mistakes in Some Example Texts 219
B.1 Texts which Contains a lot of Terminological Terms . . . . . . . . 219
B.2 Texts of News . . . . . . . . . . . . . . . . . . . . . . . . . . . . 223
B.3 Texts of Technic . . . . . . . . . . . . . . . . . . . . . . . . . . . 229
B.4 Texts of Politic . . . . . . . . . . . . . . . . . . . . . . . . . . . 234
B.5 Mistakes Analyse . . . . . . . . . . . . . . . . . . . . . . . . . . 243
C Resume 248
3Chapter 1
Introduction
The first step of any NLP task for any natural language is tokenization. For
some languages, such as English or Germa

  • Univers Univers
  • Ebooks Ebooks
  • Livres audio Livres audio
  • Presse Presse
  • Podcasts Podcasts
  • BD BD
  • Documents Documents