Grapheme-to-Phoneme Conversion using Multiple Unbounded Overlapping Chunks

Computer Science – Computation and Language

Scientific paper

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

11 pages, Postscript only, Proceedings of NeMLaP II

Scientific paper

We present in this paper an original extension of two data-driven algorithms for the transcription of a sequence of graphemes into the corresponding sequence of phonemes. In particular, our approach generalizes the algorithm originally proposed by Dedina and Nusbaum (D&N) (1991), which had originally been promoted as a model of the human ability to pronounce unknown words by analogy to familiar lexical items. We will show that DN's algorithm performs comparatively poorly when evaluated on a realistic test set, and that our extension allows us to improve substantially the performance of the analogy-based model. We will also suggest that both algorithms can be reformulated in a much more general framework, which allows us to anticipate other useful extensions. However, considering the inability to define in these models important notions like lexical neighborhood, we conclude that both approaches fail to offer a proper model of the analogical processes involved in reading aloud.

No associations

LandOfFree

Say what you really think

Search LandOfFree.com for scientists and scientific papers. Rate them and share your experience with other people.

Rating

Grapheme-to-Phoneme Conversion using Multiple Unbounded Overlapping Chunks does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.

If you have personal experience with Grapheme-to-Phoneme Conversion using Multiple Unbounded Overlapping Chunks, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Grapheme-to-Phoneme Conversion using Multiple Unbounded Overlapping Chunks will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFWR-SCP-O-358202

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.