References - INFLIBNETshodhganga.inflibnet.ac.in/bitstream/10603/50844/17/17_reference.pdf · 163...

28
163 References: [1] Lyberman F., Blumstein S,” Speech Physiology Perception and Acoustic Phonetics”, Cambridge University Press,1988. [2] Lawrence Rabiner, Biing-Hwang Juang, “Fundamentals of speech Recognition”, Prentice –Hall International, Inc.1993. [3] STEFANIE TOMKO, THOMAS K. HARRIS, ARTHUR TOTH, JAMES SANDERS,ALEXANDER RUDNICKY, and RONI OSENFELD,” Towards Efficient Human Machine Speech Communication: The Speech Graffiti Project”, ACM Transactions on Speech and Language Processing, Vol. 2, No. 1, February 2005, Article 2. [4] Jurafsky, D., Martin, J. H.,” Speech and Language Processing: An Introduction to Natural Language Processing”, Computational Linguistics, and Speech Recognition, Prentice Hall,New Jersey 2000. [5] Weinschenk, S., Barker, D. T.,”Designing Effective Speech Interfaces” , John Wiley \& Sons Inc., New York 2000. [6] Kacalak, W., Majewski, M.,” Automatic recognition and safety estimation of Voice commands in natural language given by the operator of the technical device using artificial neural networks”, Proceedings of the ANNIE 2004 Conference, Artificial Neural Networks in Engineering ANNIE 2004, Vol. 14: Smart

Transcript of References - INFLIBNETshodhganga.inflibnet.ac.in/bitstream/10603/50844/17/17_reference.pdf · 163...

163

References:

[1] Lyberman F., Blumstein S,” Speech Physiology Perception and Acoustic

Phonetics”, Cambridge University Press,1988.

[2] Lawrence Rabiner, Biing-Hwang Juang, “Fundamentals of speech Recognition”,

Prentice –Hall International, Inc.1993.

[3] STEFANIE TOMKO, THOMAS K. HARRIS, ARTHUR TOTH, JAMES

SANDERS,ALEXANDER RUDNICKY, and RONI OSENFELD,” Towards

Efficient Human Machine Speech Communication: The Speech Graffiti Project”,

ACM Transactions on Speech and Language Processing, Vol. 2, No. 1, February

2005, Article 2.

[4] Jurafsky, D., Martin, J. H.,” Speech and Language Processing: An Introduction to

Natural Language Processing”, Computational Linguistics, and Speech

Recognition, Prentice Hall,New Jersey 2000.

[5] Weinschenk, S., Barker, D. T.,”Designing Effective Speech

Interfaces” , John Wiley \& Sons Inc., New York 2000.

[6] Kacalak, W., Majewski, M.,” Automatic recognition and safety estimation of

Voice commands in natural language given by the operator of the technical device

using artificial neural networks”, Proceedings of the ANNIE 2004 Conference,

Artificial Neural Networks in Engineering ANNIE 2004, Vol. 14: Smart

164

Engineering Systems Design, St. Louis,ASME Press, New York 2004, 831-836.

[7] Kacalak, W., Majewski, M.,” Intelligent Layer of Two-Way Voice Communication

Of the Technological Device with the Operator”, Lectures Notes in Artificial

Intelligence 3070, Subseries of Lecture Notes in Computer Science, Springer-

Verlag Berlin Heidelberg New York 2004, 610-615.

[8] Dennis Klatt (1987), "Review of text-to-speech conversion for English" J. Acous.

Soc. Amer. 82, 737-793.

[9] X. Huang, A. Acero, and H.W. Hon, “Spoken Language Processing”, Prentice Hall

PTR,2001.

[10] Lacture, ” Anatomy of speech production Phonetic Transcription “,

week2,LING120,Introduction to Speech Analysis,2010.

[11] Helenca Duxans i Barrobes,” Voice Conversion applied to Text-to-Speech systems

“,Ph.D. Thesis, Universitat Polit_ecnica de Catalunya Barcelona, May 2006.

[12] Schroeder M. (1993). “A Brief History of Synthetic Speech”. Speech

Communication vol.13, pp. 231-237.

[13] Flanagan J. (1972).” Speech Analysis, Synthesis, and Perception. Springer-

Verlag”, Berlin-Heidelberg-New York.

[14] Flanagan J., Rabiner L. (Editors) (1973).” Speech Synthesis”. Dowden,

165

Hutchinson & Ross, Inc., Pennsylvania.

[15] Murtaza Bulut , Shrikanth S. Narayanan , Ann K. Syrdal, “Expressive speech

synthesis using a concatenative synthesizer”, International Conference on

Spoken Language Processing,2002.

[16] Santen J., Sproat R., Olive J., Hirschberg J. (editors) (1997).” Progress in Speech

Synthesis”, Springer-Verlag New York Inc.

[17] Kleijn K., Paliwal K. (Editors) (1998).,“Speech Coding and Synthesis”. Elsevier

Science B.V., The Netherlands.

[18] Rajeswari K C, Uma Maheswari P , “ Prosody Modeling Techniques for Text-to-

Speech Synthesis Systems - A Survey”, International Journal of Computer

Applications (0975 – 8887) ,Volume 39– No.16, February 2012.

[19] Anupam Basu, Debasish Sen , Shiraj Sen and Soumen Chakraborty,” An Indian

Language Speech Synthesizer –Techniques and Applications”, National

Systems Conference, NSC 2003, Kharagpur.

[20] Murray, I.R., Arnott, J.L., and Rohwer, E.A., “Emotional Stress in Synthetic

Speech: Progress and Future Directions”, Speech Communication, Vol. 20, pp. 3-

12 November 1996.

166

[21] Donovan R. (1996). “Trainable Speech Synthesis”. PhD. Thesis. Cambridge

University Engineering Department, England.

[22] O'Saughnessy D. (1987).” Speech Communication - Human and Machine”,

Addison- Wesley.

[23] Witten I. (1982). “Principles of Computer Speech”, Academic Press Inc.

[24] Allen J., Hunnicutt S., Klatt D. (1987). “From Text to Speech: The MITalk

System”.Cambridge University Press, Inc.

[25] Frances Alias, Xavier Servillano, Joan Claudi socoro and Xavier Gonzalvo

“Towards High-Quality Next Generation Text-to-Speech Synthesis:A multi

domain Approach by Automatic Domain Classification”,IEEE Transactions on

Audio,Speech and Languag Processing, Vol16, No,7 September 2008.

[26] M. Z. Rashad, Hazem M. El-Bakry, Islam R., Nikos Mastorakis,” An Overview of

Text-To-Speech Synthesis Techniques”, preceding of Latest Trends On

Communications and Information Technology, ISBN: 978-960- 474-207-3.

[27] Rahim M., Goodyear C., Kleijn B., Schroeter J., Sondhi M. (1993). “On the Use of

Neural Networks in Articulatory Speech Synthesis”. Journal of the Acoustical

Society of America, JASA vol. 93 (2): 1109-1121.

167

[28] Kröger B. (1992). “Minimal Rules for Articulatory Speech Synthesis”.

Proceedings of EUSIPCO92 (1): 331-334.

[29] Suresh Kumar Thakur, K.J. Satao,” Study of Various kinds of Speech Synthesizer

Technologies and Expression For Expressive Text To Speech Conversion

System”, (IJAEST) International Journal Of Advanced Engineering Sciences And

Technologies Vol No. 8, Issue No. 2, 301 – 305,2011.

[30] Susan Bartlett, Grzegorz Kondrak, and Colin Cherry.2008. “Automatic

syllabification with structured SVMs for letter-to-phoneme conversion”. In

Proceedings of ACL-08: HLT, pages 568–576, Columbus, Ohio, June.ACL.

[31] James Allen, “Natural Language Understanding”, 1995. Addison-Wesley

Professional

[32] R. Kaplan and M. Kay, “Regular Models of Phonological RuleSystems”,

Computational Linguistics Vol. 20 No. 3 331-378.

[33] T. Dutoit, “An Introduction to Text-to-Speech Synthesis”. Kluwer Academic

Publishers,1997.

[34] P. Taylor (2005), “Grapheme-to-Phoneme conversion using Hidden Markov

Models”. Proc. Interspeech 2005.

168

[35] Asanka Wasala, Ruvan Weerasinghe and Kumudu Gamage “Sinhala Grapheme-

to-Phoneme Conversion and Rules for Schwa Epenthesis”, Working Papers 2004-

2007.

[36] Ralph Penny,” Variation and Change in Spanish”, Cambridge University Press.

[37] Kayleen Schumacher,” Speaking English: A Geographical Analysis of Dialect

Distribution in Massachusetts”, The Ruth & Ted Braun Awards for Writing

Excellence at Saginaw Valley State University .

[38] Wolfram, Walt, and Natalie Schilling-Estes. American English: Dialects and

Variation. Malden:Blackwell Publishing, 1998.

[39] Deshmukh, N., Ganapathiraju, A, Picone J., (1999), “Hierarchical Search for

Large Vocabulary Conversational Speech Recognition”. IEEE Signal Processing

Magazine, 1(5):84-107.

[40] Kandasamy, S., (1995),”Speech recognition systems”. SURPRISE Journal,1(1).

[41] Pinker, S., (1994), “The Language Instinct”, Harper Collins, New York City, New

York,USA.

[42] Tiago H. Falk, Sebastian Möller.” Towards Signal-Based Instrumental Quality

Diagnosis for Text-to-Speech Systems”, Ieee Signal Processing

169

Letters, Vol. 15, 2008,Pages 781-784.

[43] Juergen Schroeter ,” Text-to-Speech (TTS) Synthesis”, The Electrical Engineering

Handbook- Circuit, Signals and Speech and Image processing, Chapter-16.

[44] Thierry Dutoit, Miloˇs Cerˇnak ” TTSBOX: a MATLAB toolbox for teaching text-

to- speech synthesis “, Acoustics, Speech, and Signal Processing, 2005.

Proceedings.(ICASSP '05). IEEE International Conference , 18-23 March

2005,pages v/537 - v/540 Vol. 5.

[45] Spyros Raptis, Aimilios Chalamandaris, Pirros Tsiakoulis, Sotiris Karabetsos "The

ILSP Text-to-Speech System for the Blizzard Challenge 2012", The Blizzard

Challenge 2012 workshop was held 14th September 2012, Portland.

[46] Costas Christogiannis, Theodora Varvarigou, Agatha Zappa, Yiannis

Vamvakoulas,” Construction of the acoustic inventory for a Greek text-to-speech

concatenative synthesis system”, Acoustics, Speech, and Signal Processing, 2000.

ICASSP '00. Proceedings. 2000 IEEE International Conference on

2000.Vol.2.pages II929 - II932 vol.2.

[47] D. Jurafsky and J.H. Martin, “Speech and Language Processing”. Pearson

Education,2000.

[48] M.D. Riley, “Tree-based modeling for speech synthesis,” Talking Machines:

170

Theories,Models and Designs, pp. 265–273, 1992.

[49] Samuel Thomas,” Natural Sounding Text-To-Speech Synthesis Based On

Syllable-Like Units”,MS Thesis, IIT Madras, May 2007.

[50] D. H. Klatt: “The KLATTalk text-to-speech conversion system”. proc. IEEE Int.

Conf. Acoustics., Speech, signal processing, 1982, pp. 1589-1592.

[51] B. S. Atal, “A new model for LPC excitation for producing natural sounding

speech at low bit-rate”, IEEE-ICASSP, pp. 611-614, 1982.

[52] S. Maeda, “A digital simulation method of the vocal tract system”,Speech

Communication, vol.1, no.3, pp. 199-229, 1982.

[53] K. N. Stevense, S. Kasowaki and G. Fant, “An electrical analogue of the vocal

tract”, JASA, no.25, pp. 734-742, 1953.

[54] P. Rubin, T. Baer and P. Mermelstein, “An articulatory synthesizer for perceptual

research,” Journal of the Acoustical Society of America, vol. 70, pp. 321–328,

1981.

[55] D. Klatt, “Software for a cascade/parallel formant synthesizer,” Journal of the

Acoustical Society of America, vol. 67, pp. 971–995, 1980.

[56] A.W. Black and P. Taylor, “Automatically clustering similar units for unit

171

selection in speech synthesis,” in Proceedings of EUROSPEECH, pp. 601–604,

1997.

[57] D. Jurafsky and J.H. Martin, “Speech and Language Processing. Pearson

Education Asia”, Second Edition 2009.

[58] A.W. Black and K.A. Lenzo, “Building synthetic voices.”

http://festvox.org/bsv/,2003.

[59] S. Lemmetty, “Review of Speech Synthesis Technology”. M.S Dissertation,

Laboratory of Acoustics and Audio Signal Processing, Helsinki University of

Technology, 1999.

[60] Jeremy Bradbury,” Linear Predictive Coding”,December 5,2000, University of

Ontario Institute of Technology (UOIT).

[61] F. Charpentier and M. Stella, “Diphone synthesis using an overlap-add technique

for speech waveforms concatenation,” in Proceedings of IEEE Int. Conf. Acoust.,

Speech and Signal Processing, pp. 2015–2018, 1986.

[62] J. Laroche, Y. Stylianou and E. Moulines, “HNS: Speech modification based on a

harmonic+noise model,” in Proceedings of IEEE Int. Conf. Acoust., Speech, and

Signal Processing, pp. 550–553, 1993.

172

[63] Eric Moulines, Francis Charpentier,” Pitch-synchronous waveform processing

techniques for text-to-speech synthesis using diphones”, Speech Communication

,Vol 9, Volume 9, Issues 5–6, December 1990, Pages 453–467.

[64] T. Dutoit, H. Leich, “MBR-PSOLA: Text-To-Speech synthesis based on an MBE

re-synthesis of the segments database”, Speech communication,Volume 13, Issues

3–4, December 1993, Pages 435–440.

[65] Ricardo Gutierrez-Osuna ,”Introduction to Speech Processing”,Lecture Notes,

Perception Sensing Instrumentation Lab, CSE@TAMU.

[66] Charpentier F., Moulines E. (1989).” Pitch-Synchronous Waveform Prosessing

Techniques for Text-to-Speech Synthesis Using Diphones”. Proceedings of

Eurospeech 89 (2): 13-19.

[67] M. Dolson, “The phase vocoder: A tutorial,” Computer Music Journal, vol. 10,pp.

145-27, 1986.

[68] R. McAulay, T Quatieri, “Speech analysis/synthesis based on a sinusoidal

representation”, IEEE Transactions on Acoustics, Speech and Signal Processing,

vol.34(4), pp.744–754, 1986.

[69] Dorran, David; Lawlor, Robert; Coyle, Eugene,” A Hybrid Time—Frequency

Domain Approach to Audio Time-Scale Modification”, JAES Volume 54 Issue

173

1/2 pp. 21-31; February 2006.

[70] David Dorran1, Robert Lawlor2, and Eugene Coyle,” A Comparison of Time-

Domain Time-Scale Modification Algorithms”, Convention Paper, 120th

Convention of Audio Engineering Society, may 2006,Paris, France.

[71] A.J. Hunt and A.W. Black, “Unit selection in a concatenative speech synthesis

system using a large speech database,” in Proceedings of IEEE Int. Conf. Acoust.,

Speech, and Signal Processing, vol. 1, pp. 373–376, 1996.

[72] Macchi, M. Bellcore, NJ. ,” Issues in text-to-speech synthesis”, Intelligence and

Systems, 1998. Proceedings., IEEE International Joint Symposia,May 1998,pp

318-325.

[73] Yi, Jon Rong-Wei. (1998).” Natural-Sounding Speech Synthesis Using Variable-

Length Units.Master thesis”. Dept. of Electrical Engineering and Computer

Science, Massachusetts Institute Of Technology.

[74] Ferencz A., Zaiu D., Ferencz M., Ratiu T., Toderean G. (1989). “A Text-To-

Speech System for the Romanian Language”, Recent Advances in Romanian

Language Technology, Center for Advanced Research in Machine Learning,

Natural Language Processing & Conceptual ModellingAwareness Campaign,

pp.163. ISBN 973- 27-0626-0.

174

[75] Kazuko AOKI,” Effective Use of Text-to-Speech Technology for Japanese

Learners of English”, NTUT Education of Disabilities, 2009 Vol. 7.

[76] Portele T., Krämer J. (1996). “Adapting a TTS System to a Reading Machine for

the Blind”. Proceedings of ICSLP 96 (1).

[77] Abedjieva E., Murray I., Arnott J. (1993).,” Applying Analysis of Human

Emotion Speech to Enhance Synthetic Speech”. Proceedings of Eurospeech 93

(2): 909-912.

[78] I.R. Murray, J.L. Arnott, N. Alm and A.F. Newell, "A Communication System for

the Disabled with Emotional Synthetic Speech Produced by Rule", Proceedings of

Eurospeech '91, 2nd European Conference on Speech Communication and

Technology, Genova, Italy, September 1991, pp. 65-68.

[79] Keller, E. & Zellner-Keller, B. (2000). “New Uses for Speech Synthesis”, The

Phonetician, 81, 35-40.

[80] Pardeep Gera,” Text-To-Speech Synthesis For Punjabi Language”,ME Thesis,

Thapar Institute of Engineering and Technology, Patiala,2006.

[81] Hunnicut S., "Grapheme-to-Phoneme rules: a Review", Speech Transmission

Laboratory, Royal Institute of Technology, Stockholm, Sweden, QPSR 2-3, pp.

38-60.

175

[82] Belrhari R., Auberge V., Boe L.J., "From lexicon to rules: towards a descriptive

method of French text-to-phonetics transcription", Proc. ICSLP 92, Alberta, pp.

1183-1186.

[83] Monojit Choudhury,”Rule Based Grapheme-to-Phoneme Mapping for Hindi

Speech Synthesis”, Department of Computer Science and Engineering Indian

Institute of Technology, Kharagpur.

[84] Chandan Sarma, Prof. P.H Talukdar,” Dialect variation in Boro Language and

Grapheme-to-Phoneme conversion rules to handle lexical lookup fails in Boro

TTS System”, International Journal of Scientific and Research Publications,

Volume 2, Issue 9, September 2012 1 ISSN 2250-3153.

[85] Susan Bartlett, Grzegorz Kondrak, and Colin Cherry.2008.” Automatic

Syllabification with structured SVMs for letter-to-phoneme conversion”. In

Proceedings of ACL-08: HLT, pages 568–576, Columbus, Ohio, June.ACL.

[86] R. I. Damper , Y. Marchand , M. J. Adamson , K. Gustafson,” Evaluating the

Pronunciation Component of Text-to-Speech Systems for English: A Performance

Comparison of Different Approaches”, In Speech And Language

Technology (Salt) Club Workshop On Evaluation In Speech And Language

Technology,1997.

[87] Chomsky, N. and M. Halle (1968). “The Sound Pattern of English”. New York,

176

NY: Harper and Row, page4.

[88] Elovitz, H. S., R. Johnson, A. McHugh, and J. E. Shore (1976). “Letter-to-sound

rules for automatic translation of English text to phonetics”. IEEE Transactions

on Acoustics, Speech and Signal Processing ASSP-24, 446–459.

[89] Beata Megyesi,” Data-Driven Syntactic Analysis Methods and Applications for

Swedish”, Doctoral Dissertation, Kungl Tekniska Hogskolan,2002.

[90] Anil Kr. Jain, Jianchang Mao,K.M. Mohiuddin, “Artificial neural networks: a

tutorial”,IEE-Computer Volume:29,Issue:3 Page(s): 31 – 44,1996.

[91] Eniko Beatrice Bilcu,” Text-To-Phoneme Mapping Using Neural Networks”,

Doctoral Thesis,Tampere University of Technology,2008.

[92] Cover, T.M., & Hart, P.E. (1967).,” Nearest neighbor pattern classification”.

Institute of Electrical and ElectronicsEngineers Transactions on Information

Theory, 13, 21-27.

[93] Hart, P.E. (1968),” The condensed nearest neighbor rule”. Institute of Electrical

and Electronics Engineers and Transactions on Information Theory, 14, 515-516.

[94] Gates, G.W. (1972), “The reduced nearest neighbor rule”. IEEE Transactions on

Information Theory, 431-433.

177

[95] Dasarathy, B.V. (1980),”Nosing around the neighborhood: A new system structure

and classification rule for recognition in partially exposed environments”. Pattern

Analysis and Machine Intelligence, 2, 67-71.

[96] David W. Aha ,Dennis Kibler ,Marc K. Albert,” Instance-Based

Learning Algorithms”, Machine Learning, 6, 37-66 ,Kluwer Academic Publishers,

Boston. Manufactured in The Netherlands.1991.

[97] Tom Mitchell, ” Machine Learning”, McGraw Hill, 1997.

[98] Breiman, Friedman, Olshen, and Stone,“Classification and Regression Trees”.

Wadsworth Inc., 1984.

[99] Gelfand, S., Ravishankar, C., and Delp, E., “An Iterative Growing and Pruning

Algorithm for Classification Tree Design”. In IEEE Pattern Analysis and Machine

Intelligence, pages 163-174, Feb. 1991.

[100] Kuhn, R., Lazaridès, A., Normandin, Y., and Brousseau, J.,“Improved decision

trees for phonetic modeling”. In ICASSP95, pages 552–555, 1995.

[101] Pagel V., K. Lenzo, A.Black. 1998. “Letter-to-sound rules for accented lexicon

Compression”. In Proc. of ICSLP98, vol 5,2015-2020, Sydney, Australia.

178

[102] Lawrence R. Rabiner.” A tutorial on hidden Markov models and selected

applications in speech recognition”. Proc. IEEE, 77:257–286, 1989.

[103] R. Durbin, S. R. Eddy, A. Krogh, and G. J. Mitchison. “Biological Sequence

Analysis: Probabilistic Models of Proteins and Nucleic Acids”. Cambridge

University Press, Cambridge UK,1998.

[104] Ghahramani , Z., “ An Introduction to Hidden Markov Model and Bayesian

Networks”, Journal of Pattern Recognition and Artificial Intelligence, 15 (1):9-

42.2001.

[105] D.B. Paul,” .Speech Recognition Using Hidden Markov Models”, The Liru:oln

Laboratory Journal, Volume 3, Number 1 (l990).

[106] L E Baum,” An inequality and an associated maximization technique in statistical

estimation of probabilistic functions of a Markov process”, by L E Baum,

Inequalities (1972) ,Volume: 3, Issue: 1, Publisher: Academic Press, Pages: 1-8.

[107] Leonard E Baum, J A Eagonin ,”An inequality with applications to statistical

estimation for probabilistic functions of Markov processes and to a model for

ecology”,Bulletin of the American Mathematical Society(1967).

[108] L E Baum, T Petriein ,” Statistical inference for probabilistic functions of finite

state Markov chains”, The Annals of Mathematical Statistics(1966).

179

[109] L. E. Baum, T. Petrie, G. Soules, and N. Weiss, "A maximization technique

occurring in the statistical analysis of probabilistic functions of Markov chains",

Ann. Math. Statist., vol. 41, no. 1, pp. 164–171, 1970.

[110] J.K.Baker, ” The Dragon System “, IEEE.Trans.Acous. Speech Signal Process.

ASSP-23, 24(1975).

[111] J.M. Baker, “Personal Communication”. Jan 1990.

[112] F.Jelinck. ,” Continuous Speech Recognition by Statistical Methods”, proc.IEEE

64,532(1970).

[113] S.E Levinson,I.R. Rabiner and M.M Sandhi ,” An Introduction to the Application

of the theory of Probabilistic Function of a Markov Process to Automatic Speech

Recognition”,Bellsys.Tech J02,1035(1983).

[114] Tatyana Polyakova, Antonio Bonafonte,” Main issues in grapheme-to-phoneme

conversion for TTS”, Volume: 1, Issue: 35, Pages: 29-34, ISSN: 11355948.

[115] Walter Daelemans and Antal Van Den Bosch.” Language-independent data-

oriented grapheme-to-phoneme conversion “.In Progressin speech synthesis ,

pages 77–89. Springer, New York.1997.

[116] Arthur Dempster, Nan Laird, and Donald Rubin. 1977. “Maximum likelihood

180

from incomplete data via the EM algorithm”. In Journal of the Royal Statistical

Society,pages B:1–38.

[117] Black A.,K. Lenzo K. And V.Pagel.” Issues In building general letter to sound

rules.” In Proc.of the 3rd ESCA workshop on speech synthesis.,77-80, Jenolah

Caves,Australia.1998.

[118] Sittichai Jiampojamarn, Grzegorz Kondrak, and Tarek Sherif. 2007.” Applying

many-to-many alignments and hidden markov models to letter-to-phoneme

conversion”. In HLT 2007: The Conference of the NAACL;Proceedings of the

Main Conference, pages 372–379,Rochester, New York, April. ACL.

[119] Bisani M. and H. Ney. ”Investigations on joint –multigram model for grapheme-

to-phoneme conversion.” In Proceedings of the 7th Int. Conf. on Spoken

Language Processing, Denver, CO, vol.1, 105-108.

[120] Sittichai Jiampojamarn and Grzegorz Kondrak,” Letter-Phoneme Alignment: An

Exploration”, Proceedings of the 48th Annual Meeting of the Association for

Computational Linguistics, pages 780–788,Uppsala, Sweden, 11-16 July 2010. c

2010 Association for Computational Linguistics.

[121] Michel Divay,Anthony J. Vitale,” Algorithms for Grapheme-Phoneme

Translation for English and French: Applications for Database Searches and

Speech Synthesis”, Computational Linguistics ,Volume 23, Number 4,1997.

181

[122] Ainsworth, W. A. 1973.” A system for converting English text into speech”. In

IEEE Transactions of Audio and Electroacoustics, pages 288-290.

[123] Bakiri, G., and T. G. Dietterich. 1991.”Converting English Text to Speech: A

Machine Learning Approach”. Ph.D. thesis.Rep. No. 91-30-1. Department of

Computer Science, Oregon State University.

[124] Bernstein, J. and L. Nessly. 1981.” Performance comparison of component

Algorithms for the phonemicization of orthography”. In Proceedings of the 19th

Annual Meeting, Stanford University.Association for Computational Linguistics.

[125] Elovitz, H. S., R. W. Johnson, A. McHugh,and J. E. Shore. 1976. “ Automatic

translation of English text to phonetics by means of letter-to-sound rules “. NRL

Report 7948, Naval Research Laboratory,Washington, D.C.

[126] Hunnicut, S. 1976. “Phonological rules for a text-to-speech system. American

Journal of Computational Linguistics”, Microfiche 57.

[127] Levin, H. 1963. “A basic research program on reading”. Final Report,

Cooperative Research Project No. 639, Cornell University.

[128] McCormick, S. and S. R. Hertz. 1989. “A new approach to English text-to-

phoneme conversion using delta”, Version 2. 117th Meeting. Journal of the

Acoustical Society of America, Supplement 1, Vol. 85, $124.

182

[129] McIlroy, M. D. 1974. “Synthetic English speech by rules”. Bell Telephone

Laboratories Memo.

[130] O'Malley, M. H. 1990.” Text-to-speech conversion technology”. Computer

IEEE, page 17.

[131] Venezky, R. L. 1962. “A Computer Program for Deriving Spelling to Sound

Correlations”. MA thesis, Cornell University. Published in part in A Basic

Research Program on Reading.

[132] Venezky, R. L. 1967a. “English orthography:Its graphical structure and its

Relation to sound”. Reading Research Quarterly, II.

[133] Venezky, R. L. 1967b. “Reading:Grapheme-phonerne relationships”.Education

87: 519-524.

[134] Venezky, R. L. 1967c.” The basis of English orthography “. Acta Linguistica

10: 145- 159.

[135] Venezky, R. L. 1970. “The Structure of English Orthography”. Mouton, The

Hague.

[136] Venezky, R. L. and R. Weir. 1966. “A study of selected spelling-to-sound

correspondence patterns”. Final Report, Cooperative Research project No. 3090,

183

Stanford University.

[137] Vitale, A. J. 1991. “An algorithm for high accuracy name pronunciation by

parametric speech synthesizer”. Computational Linguistics 17(3).

[138] Weir, R. 1964. “Formulation of grapheme-phoneme correspondence rules

to aid the teaching of reading”. Final Report, Cooperative Research project

No. S-039 Stanford University.

[139] Aubergd, V. 1991. La synthese de la parole:"des regles au lexique". These,

Universitd Stendhal, Grenoble.

[140] Bechet F., T. Spriet, and M. E1-Beze. 1996.”Traitement spdcifique des noms

propres dans un systeme de transcription grapheme-phoneme”. JST Avignon.

[141] Catach, N. 1989.” Informatique: Traitement automatique du Langage”. Bulletin

Liaisons-Heso, September.

[142] Catach, N. and L. Catach 1992. Pr4sentation du logiciel VOISINETTE, "Un

correcteur entree phonetique". CNRS-INFOS.

[143] Cotto, D. 1992. Traitement automatique des textes en vue de la synthese vocale.

These,Universite Paul Sabatier, Toulouse III.

184

[144] Divay, M. 1984. “ De l'dcrit vers l'oral ou contribution a l'etude des traitements

des textes ecrits en vue de leur prononciation sur synthetiseur de parole. These

d'Etat, Universit~ de Rennes, France.

[145] Divay, M. 1985.” A text-processing expert system”. 5eme Congres

Reconnaissance des formes et Intelligence Artificielle ,Novembre 1985,

Grenoble, France.

[146] Divay, M. 1990a.” Traitement du langage naturel: la phonetisation ou comment

apprendre a l'ordinateur a lire un texte Francais”. MICRO-SYSTEMES, March.

[147] Divay, M. 1990b. A written processing expert system for text to phoneme

conversion. In Proceedings of the International Conference on Spoken Language

(ICSLP 90), Kobe, Japan.

[148] Divay, M. 1991. “CD-ROM Electronic Dictionary”, November.

[149] Laporte, E. 1988.” Methodes algorithmiques et lexicales de phonetisation de

Texts” .These, Universit4 Paris 7, May.

[150] Prouts, B. 1980,” Contribution a la synthese de la parole a partir de texte;

transcription graphemo-phondtique en temps reel sur microprocesseur”. These de

Docteur-Ingenieur, Universit4 de Paris Sud, Orsay.

[151] Yvon, F. 1996. “Prononcer par analogie: motivation, formalisation et fvaluation”.

185

These,Ecole nationale des Telecommunications,Paris.

[152] Dedina, M. J. and H. C. Nusbaum. 1991. “PRONOUNCE: A program for

pronunciation by analogy”. Computer Speech and Language 5:55-64.

[153] Sejnowski, T. J. and C. R. Rosenberg. 1987. “NETtalk: Parallel networks that

learn to pronounce English text”. Complex Systems 1:145-168.

[154] Lucas, S. M. and R. I. Damper. 1992.” Syntactic neural networks for bi-

directional text-phonetics translation”.In G. Bailly and C. Benoit, editors,

Talking Machines, Theories, Models and Designs.North-Holland Publishers.

[155] Hochberg, J., S. M. Mniszewski, T. Calleja,and G. J. Papcun. 1991. “A default

hierarchy for pronouncing English”. IEEETransactions on Pattern Matching and

Machine Intelligence 13(9): 957-964.

[156] Klatt, D. H. and D. W. Shipman. 1982.”Letter-to-phoneme rules: A semi-

Automatic discovery procedure”.Journal of the Acoustical Society of

America.82: 737-793.

[157] Lucassen, J. M. and R. L. Mercer. 1984. “An information theoretic approach to

the automatic determination of phonemic baseforms”. In Proceedings of

lCASSP-84,pages 42.5.1-42.5.3, San Diego.

[158] Parfitt, S. and R. Sharman. 1991. “A bi-directional model of English

186

pronunciation”. In Proceedings of Eurospeech,volume 2, pages 801-804.

[159] Golding, A. R. 1991. “Pronouncing Names by a Combination of Case-based and

Rule-based Reasoning”. Ph.D. Thesis, Stanford University.

[160] Meng, H. M. 1995. “Phonological Parsing for Bi-Directional Letter-to-Sound

and Sound-to-Letter Generation”. Ph.D. Thesis,MIT, Cambridge, MA.

[161] H.B Hudgson“Essay the First: on the Koch, Bodo and Dhimal Tribes”,1946.

[162] Pratima Brahma, ” L a n g u a g e i n I n d ia “, volume 1 2 , 2 0 1 2

A s s a m e s e I n f l u e n c e o n B o d o, ISSN 1930-2940.

[163] Madhu Ram Baro,”Structure of Boro language”,N.L Publications,2008.

[164] http://www.iitg.ernet.in/rcilts/

[165] Madhu Ram Baro, “The Historical Development of Boro Langguage”, N.L

Publications ,2007.

[166] Priyankoo Sarmah,” Some Aspects Of The Tonal Phonology Of Bodo”, A

Dissertation Of Master Of Philosophy In Linguistics Central Institute Of

English And Foreign Languages,2004.

[167] Weidert, Alfons.1987. Tibeto-Burman Tonology. John Benjamins Publishing

187

Company,Amsterdam/Philadelphia.

[168] Maassen Ben and Povel Dirk-Jan 1984,” The effect of correcting fundamental

frequency on the intelligibility of deaf speech and its interaction with temporal

aspects “. Journal of the Acoustical Society of America, 76 (6): 1673–1681.

[169] Laures, Jacqueline S. and Weismer, Gary 1999. The effects of a flattened

fundamental frequency on intelligibility at the sentence level. Journal of Speech

Language and Hearing Research, 42(5): 1148a–1156.

[170] Ineke Mennen,” Phonetic and Phonological influences in non-native intonation :

an overview for language teachers.”, QMUC Speech Science Research Centre

Working Paper WP9,2006.

[171] Professor Pat Kamalani Hurley,” Introduction to the Study of Language”,

University of Hawai'i - Leeward Community College.

[172] Paul Kerswill,” Migration and language”, An international handbook of the

Science of language and society”, 2nd edn., Vol 3. Berlin: De Gruyter.2006.

[173] O'Grady, William, John Archibald, Mark Aronoff, and Jane Rees-Miller. eds.

(2001) ,” Contemporary Linguistics”. Boston: Bedford/St. Martin's.

[174] Edward J. Vajda, ” Dialectology”, course material, Linguistics 201, Western

188

Washington University,2001.

[175] Dr. Pramod Chandra Bhattacharjya, “A Descriptive Analysis of the Boro

Language”, Doctoral Thesis, 1977.

[176] Phukan. Basumatary, (2005). “An Introduction to the Boro Language”, Mittali

Publications,New Delhi.

[177] Vera Demberg, Helmut Schmid, Gregor M¨ohler,” Phonological Constraints and

Morphological Preprocessing for Grapheme-to-Phoneme Conversion”,

Proceedings of the 45th Annual Meeting of the Association of Computational

Linguistics, pages 96–103,Prague, Czech Republic, June 2007. c 2007

Association for Computational Linguistics.

[178] ChandanSarma,U.Sharma,C.K.Nath,S.Kalita,P.H.Talukdar,” Selection of Units

and Development of Speech Database for Natural Sounding Bodo TTS

System”, CISP Guwahati ,March 2012.

[179] Parminder Singh, Gurpreet Singh Lehal, “Syllables Selection for the

Development of Speech Database for Punjabi TTS System”, IJCSI

International Journal of Computer Science Issues, Vol. 7, Issue 6, November

2010.

[180] R.A. Krakow,” Physiological organization of syllables: a review”, Journal of

189

Phonetics, Vol. 27, 1999, pp. 23-54.

[181] Susan Bartlett, Grzegorz Kondrak, Colin Cherry, “On the Syllabification of

Phonemes”, Human Language Technologies: The 2009 Annual Conference of

the North American Chapter of the ACL, pages 308– 316,Boulder, Colorado,

June 2009. c 2009 Association for Computational Linguistics.

[182] Y. A. El-Imam, “Phonetization of arabic: rules and algorithms”, Computer

Speech & Language, vol. 18, pp. 339–373, October 2004.

[183] R. Dale et al. (Eds.), “A Rule Based Syllabification Algorithm for Sinhala”,

IJCNLP 2005, LNAI 3651, pp. 438 – 449, 2005.© Springer-Verlag Berlin

Heidelberg 2005.

[184] Juliette Blevins,”The syllable in phonological theory”,1995.

[185] George Kiraz and Bernd M¨obius, “Multilingual syllabification using weighted

finite-state transducers”.In Proceedings of the 3rd Workshop on Speech

Synthesis,1998.

[186] Robert Damper. 2001. “Learning about speech from data: Beyond NETtalk”. In

Data-Driven Techniques in Speech Synthesis, pages 1–25. Kluwer Academic

Publishers.

190

[187] John R. Rickford,“How Linguists Approach the Study of Language and Dialect”

, ms. January 2002, for students in Ling 73, AAVE, Stanfo.

[188] Othman O. Khalifa, Zakiah Hanim Ahmad, Aisha-Hassan A. Hashim &Teddy

Suya Gunawan,” SMaTalk: Standard Malay Text to Speech Talk System”,

Signal Processing: An International Journal, Volume (2) : Issue (5).August 2008