References - INFLIBNETshodhganga.inflibnet.ac.in/bitstream/10603/50844/17/17_reference.pdf · 163...
Transcript of References - INFLIBNETshodhganga.inflibnet.ac.in/bitstream/10603/50844/17/17_reference.pdf · 163...
163
References:
[1] Lyberman F., Blumstein S,” Speech Physiology Perception and Acoustic
Phonetics”, Cambridge University Press,1988.
[2] Lawrence Rabiner, Biing-Hwang Juang, “Fundamentals of speech Recognition”,
Prentice –Hall International, Inc.1993.
[3] STEFANIE TOMKO, THOMAS K. HARRIS, ARTHUR TOTH, JAMES
SANDERS,ALEXANDER RUDNICKY, and RONI OSENFELD,” Towards
Efficient Human Machine Speech Communication: The Speech Graffiti Project”,
ACM Transactions on Speech and Language Processing, Vol. 2, No. 1, February
2005, Article 2.
[4] Jurafsky, D., Martin, J. H.,” Speech and Language Processing: An Introduction to
Natural Language Processing”, Computational Linguistics, and Speech
Recognition, Prentice Hall,New Jersey 2000.
[5] Weinschenk, S., Barker, D. T.,”Designing Effective Speech
Interfaces” , John Wiley \& Sons Inc., New York 2000.
[6] Kacalak, W., Majewski, M.,” Automatic recognition and safety estimation of
Voice commands in natural language given by the operator of the technical device
using artificial neural networks”, Proceedings of the ANNIE 2004 Conference,
Artificial Neural Networks in Engineering ANNIE 2004, Vol. 14: Smart
164
Engineering Systems Design, St. Louis,ASME Press, New York 2004, 831-836.
[7] Kacalak, W., Majewski, M.,” Intelligent Layer of Two-Way Voice Communication
Of the Technological Device with the Operator”, Lectures Notes in Artificial
Intelligence 3070, Subseries of Lecture Notes in Computer Science, Springer-
Verlag Berlin Heidelberg New York 2004, 610-615.
[8] Dennis Klatt (1987), "Review of text-to-speech conversion for English" J. Acous.
Soc. Amer. 82, 737-793.
[9] X. Huang, A. Acero, and H.W. Hon, “Spoken Language Processing”, Prentice Hall
PTR,2001.
[10] Lacture, ” Anatomy of speech production Phonetic Transcription “,
week2,LING120,Introduction to Speech Analysis,2010.
[11] Helenca Duxans i Barrobes,” Voice Conversion applied to Text-to-Speech systems
“,Ph.D. Thesis, Universitat Polit_ecnica de Catalunya Barcelona, May 2006.
[12] Schroeder M. (1993). “A Brief History of Synthetic Speech”. Speech
Communication vol.13, pp. 231-237.
[13] Flanagan J. (1972).” Speech Analysis, Synthesis, and Perception. Springer-
Verlag”, Berlin-Heidelberg-New York.
[14] Flanagan J., Rabiner L. (Editors) (1973).” Speech Synthesis”. Dowden,
165
Hutchinson & Ross, Inc., Pennsylvania.
[15] Murtaza Bulut , Shrikanth S. Narayanan , Ann K. Syrdal, “Expressive speech
synthesis using a concatenative synthesizer”, International Conference on
Spoken Language Processing,2002.
[16] Santen J., Sproat R., Olive J., Hirschberg J. (editors) (1997).” Progress in Speech
Synthesis”, Springer-Verlag New York Inc.
[17] Kleijn K., Paliwal K. (Editors) (1998).,“Speech Coding and Synthesis”. Elsevier
Science B.V., The Netherlands.
[18] Rajeswari K C, Uma Maheswari P , “ Prosody Modeling Techniques for Text-to-
Speech Synthesis Systems - A Survey”, International Journal of Computer
Applications (0975 – 8887) ,Volume 39– No.16, February 2012.
[19] Anupam Basu, Debasish Sen , Shiraj Sen and Soumen Chakraborty,” An Indian
Language Speech Synthesizer –Techniques and Applications”, National
Systems Conference, NSC 2003, Kharagpur.
[20] Murray, I.R., Arnott, J.L., and Rohwer, E.A., “Emotional Stress in Synthetic
Speech: Progress and Future Directions”, Speech Communication, Vol. 20, pp. 3-
12 November 1996.
166
[21] Donovan R. (1996). “Trainable Speech Synthesis”. PhD. Thesis. Cambridge
University Engineering Department, England.
[22] O'Saughnessy D. (1987).” Speech Communication - Human and Machine”,
Addison- Wesley.
[23] Witten I. (1982). “Principles of Computer Speech”, Academic Press Inc.
[24] Allen J., Hunnicutt S., Klatt D. (1987). “From Text to Speech: The MITalk
System”.Cambridge University Press, Inc.
[25] Frances Alias, Xavier Servillano, Joan Claudi socoro and Xavier Gonzalvo
“Towards High-Quality Next Generation Text-to-Speech Synthesis:A multi
domain Approach by Automatic Domain Classification”,IEEE Transactions on
Audio,Speech and Languag Processing, Vol16, No,7 September 2008.
[26] M. Z. Rashad, Hazem M. El-Bakry, Islam R., Nikos Mastorakis,” An Overview of
Text-To-Speech Synthesis Techniques”, preceding of Latest Trends On
Communications and Information Technology, ISBN: 978-960- 474-207-3.
[27] Rahim M., Goodyear C., Kleijn B., Schroeter J., Sondhi M. (1993). “On the Use of
Neural Networks in Articulatory Speech Synthesis”. Journal of the Acoustical
Society of America, JASA vol. 93 (2): 1109-1121.
167
[28] Kröger B. (1992). “Minimal Rules for Articulatory Speech Synthesis”.
Proceedings of EUSIPCO92 (1): 331-334.
[29] Suresh Kumar Thakur, K.J. Satao,” Study of Various kinds of Speech Synthesizer
Technologies and Expression For Expressive Text To Speech Conversion
System”, (IJAEST) International Journal Of Advanced Engineering Sciences And
Technologies Vol No. 8, Issue No. 2, 301 – 305,2011.
[30] Susan Bartlett, Grzegorz Kondrak, and Colin Cherry.2008. “Automatic
syllabification with structured SVMs for letter-to-phoneme conversion”. In
Proceedings of ACL-08: HLT, pages 568–576, Columbus, Ohio, June.ACL.
[31] James Allen, “Natural Language Understanding”, 1995. Addison-Wesley
Professional
[32] R. Kaplan and M. Kay, “Regular Models of Phonological RuleSystems”,
Computational Linguistics Vol. 20 No. 3 331-378.
[33] T. Dutoit, “An Introduction to Text-to-Speech Synthesis”. Kluwer Academic
Publishers,1997.
[34] P. Taylor (2005), “Grapheme-to-Phoneme conversion using Hidden Markov
Models”. Proc. Interspeech 2005.
168
[35] Asanka Wasala, Ruvan Weerasinghe and Kumudu Gamage “Sinhala Grapheme-
to-Phoneme Conversion and Rules for Schwa Epenthesis”, Working Papers 2004-
2007.
[36] Ralph Penny,” Variation and Change in Spanish”, Cambridge University Press.
[37] Kayleen Schumacher,” Speaking English: A Geographical Analysis of Dialect
Distribution in Massachusetts”, The Ruth & Ted Braun Awards for Writing
Excellence at Saginaw Valley State University .
[38] Wolfram, Walt, and Natalie Schilling-Estes. American English: Dialects and
Variation. Malden:Blackwell Publishing, 1998.
[39] Deshmukh, N., Ganapathiraju, A, Picone J., (1999), “Hierarchical Search for
Large Vocabulary Conversational Speech Recognition”. IEEE Signal Processing
Magazine, 1(5):84-107.
[40] Kandasamy, S., (1995),”Speech recognition systems”. SURPRISE Journal,1(1).
[41] Pinker, S., (1994), “The Language Instinct”, Harper Collins, New York City, New
York,USA.
[42] Tiago H. Falk, Sebastian Möller.” Towards Signal-Based Instrumental Quality
Diagnosis for Text-to-Speech Systems”, Ieee Signal Processing
169
Letters, Vol. 15, 2008,Pages 781-784.
[43] Juergen Schroeter ,” Text-to-Speech (TTS) Synthesis”, The Electrical Engineering
Handbook- Circuit, Signals and Speech and Image processing, Chapter-16.
[44] Thierry Dutoit, Miloˇs Cerˇnak ” TTSBOX: a MATLAB toolbox for teaching text-
to- speech synthesis “, Acoustics, Speech, and Signal Processing, 2005.
Proceedings.(ICASSP '05). IEEE International Conference , 18-23 March
2005,pages v/537 - v/540 Vol. 5.
[45] Spyros Raptis, Aimilios Chalamandaris, Pirros Tsiakoulis, Sotiris Karabetsos "The
ILSP Text-to-Speech System for the Blizzard Challenge 2012", The Blizzard
Challenge 2012 workshop was held 14th September 2012, Portland.
[46] Costas Christogiannis, Theodora Varvarigou, Agatha Zappa, Yiannis
Vamvakoulas,” Construction of the acoustic inventory for a Greek text-to-speech
concatenative synthesis system”, Acoustics, Speech, and Signal Processing, 2000.
ICASSP '00. Proceedings. 2000 IEEE International Conference on
2000.Vol.2.pages II929 - II932 vol.2.
[47] D. Jurafsky and J.H. Martin, “Speech and Language Processing”. Pearson
Education,2000.
[48] M.D. Riley, “Tree-based modeling for speech synthesis,” Talking Machines:
170
Theories,Models and Designs, pp. 265–273, 1992.
[49] Samuel Thomas,” Natural Sounding Text-To-Speech Synthesis Based On
Syllable-Like Units”,MS Thesis, IIT Madras, May 2007.
[50] D. H. Klatt: “The KLATTalk text-to-speech conversion system”. proc. IEEE Int.
Conf. Acoustics., Speech, signal processing, 1982, pp. 1589-1592.
[51] B. S. Atal, “A new model for LPC excitation for producing natural sounding
speech at low bit-rate”, IEEE-ICASSP, pp. 611-614, 1982.
[52] S. Maeda, “A digital simulation method of the vocal tract system”,Speech
Communication, vol.1, no.3, pp. 199-229, 1982.
[53] K. N. Stevense, S. Kasowaki and G. Fant, “An electrical analogue of the vocal
tract”, JASA, no.25, pp. 734-742, 1953.
[54] P. Rubin, T. Baer and P. Mermelstein, “An articulatory synthesizer for perceptual
research,” Journal of the Acoustical Society of America, vol. 70, pp. 321–328,
1981.
[55] D. Klatt, “Software for a cascade/parallel formant synthesizer,” Journal of the
Acoustical Society of America, vol. 67, pp. 971–995, 1980.
[56] A.W. Black and P. Taylor, “Automatically clustering similar units for unit
171
selection in speech synthesis,” in Proceedings of EUROSPEECH, pp. 601–604,
1997.
[57] D. Jurafsky and J.H. Martin, “Speech and Language Processing. Pearson
Education Asia”, Second Edition 2009.
[58] A.W. Black and K.A. Lenzo, “Building synthetic voices.”
http://festvox.org/bsv/,2003.
[59] S. Lemmetty, “Review of Speech Synthesis Technology”. M.S Dissertation,
Laboratory of Acoustics and Audio Signal Processing, Helsinki University of
Technology, 1999.
[60] Jeremy Bradbury,” Linear Predictive Coding”,December 5,2000, University of
Ontario Institute of Technology (UOIT).
[61] F. Charpentier and M. Stella, “Diphone synthesis using an overlap-add technique
for speech waveforms concatenation,” in Proceedings of IEEE Int. Conf. Acoust.,
Speech and Signal Processing, pp. 2015–2018, 1986.
[62] J. Laroche, Y. Stylianou and E. Moulines, “HNS: Speech modification based on a
harmonic+noise model,” in Proceedings of IEEE Int. Conf. Acoust., Speech, and
Signal Processing, pp. 550–553, 1993.
172
[63] Eric Moulines, Francis Charpentier,” Pitch-synchronous waveform processing
techniques for text-to-speech synthesis using diphones”, Speech Communication
,Vol 9, Volume 9, Issues 5–6, December 1990, Pages 453–467.
[64] T. Dutoit, H. Leich, “MBR-PSOLA: Text-To-Speech synthesis based on an MBE
re-synthesis of the segments database”, Speech communication,Volume 13, Issues
3–4, December 1993, Pages 435–440.
[65] Ricardo Gutierrez-Osuna ,”Introduction to Speech Processing”,Lecture Notes,
Perception Sensing Instrumentation Lab, CSE@TAMU.
[66] Charpentier F., Moulines E. (1989).” Pitch-Synchronous Waveform Prosessing
Techniques for Text-to-Speech Synthesis Using Diphones”. Proceedings of
Eurospeech 89 (2): 13-19.
[67] M. Dolson, “The phase vocoder: A tutorial,” Computer Music Journal, vol. 10,pp.
145-27, 1986.
[68] R. McAulay, T Quatieri, “Speech analysis/synthesis based on a sinusoidal
representation”, IEEE Transactions on Acoustics, Speech and Signal Processing,
vol.34(4), pp.744–754, 1986.
[69] Dorran, David; Lawlor, Robert; Coyle, Eugene,” A Hybrid Time—Frequency
Domain Approach to Audio Time-Scale Modification”, JAES Volume 54 Issue
173
1/2 pp. 21-31; February 2006.
[70] David Dorran1, Robert Lawlor2, and Eugene Coyle,” A Comparison of Time-
Domain Time-Scale Modification Algorithms”, Convention Paper, 120th
Convention of Audio Engineering Society, may 2006,Paris, France.
[71] A.J. Hunt and A.W. Black, “Unit selection in a concatenative speech synthesis
system using a large speech database,” in Proceedings of IEEE Int. Conf. Acoust.,
Speech, and Signal Processing, vol. 1, pp. 373–376, 1996.
[72] Macchi, M. Bellcore, NJ. ,” Issues in text-to-speech synthesis”, Intelligence and
Systems, 1998. Proceedings., IEEE International Joint Symposia,May 1998,pp
318-325.
[73] Yi, Jon Rong-Wei. (1998).” Natural-Sounding Speech Synthesis Using Variable-
Length Units.Master thesis”. Dept. of Electrical Engineering and Computer
Science, Massachusetts Institute Of Technology.
[74] Ferencz A., Zaiu D., Ferencz M., Ratiu T., Toderean G. (1989). “A Text-To-
Speech System for the Romanian Language”, Recent Advances in Romanian
Language Technology, Center for Advanced Research in Machine Learning,
Natural Language Processing & Conceptual ModellingAwareness Campaign,
pp.163. ISBN 973- 27-0626-0.
174
[75] Kazuko AOKI,” Effective Use of Text-to-Speech Technology for Japanese
Learners of English”, NTUT Education of Disabilities, 2009 Vol. 7.
[76] Portele T., Krämer J. (1996). “Adapting a TTS System to a Reading Machine for
the Blind”. Proceedings of ICSLP 96 (1).
[77] Abedjieva E., Murray I., Arnott J. (1993).,” Applying Analysis of Human
Emotion Speech to Enhance Synthetic Speech”. Proceedings of Eurospeech 93
(2): 909-912.
[78] I.R. Murray, J.L. Arnott, N. Alm and A.F. Newell, "A Communication System for
the Disabled with Emotional Synthetic Speech Produced by Rule", Proceedings of
Eurospeech '91, 2nd European Conference on Speech Communication and
Technology, Genova, Italy, September 1991, pp. 65-68.
[79] Keller, E. & Zellner-Keller, B. (2000). “New Uses for Speech Synthesis”, The
Phonetician, 81, 35-40.
[80] Pardeep Gera,” Text-To-Speech Synthesis For Punjabi Language”,ME Thesis,
Thapar Institute of Engineering and Technology, Patiala,2006.
[81] Hunnicut S., "Grapheme-to-Phoneme rules: a Review", Speech Transmission
Laboratory, Royal Institute of Technology, Stockholm, Sweden, QPSR 2-3, pp.
38-60.
175
[82] Belrhari R., Auberge V., Boe L.J., "From lexicon to rules: towards a descriptive
method of French text-to-phonetics transcription", Proc. ICSLP 92, Alberta, pp.
1183-1186.
[83] Monojit Choudhury,”Rule Based Grapheme-to-Phoneme Mapping for Hindi
Speech Synthesis”, Department of Computer Science and Engineering Indian
Institute of Technology, Kharagpur.
[84] Chandan Sarma, Prof. P.H Talukdar,” Dialect variation in Boro Language and
Grapheme-to-Phoneme conversion rules to handle lexical lookup fails in Boro
TTS System”, International Journal of Scientific and Research Publications,
Volume 2, Issue 9, September 2012 1 ISSN 2250-3153.
[85] Susan Bartlett, Grzegorz Kondrak, and Colin Cherry.2008.” Automatic
Syllabification with structured SVMs for letter-to-phoneme conversion”. In
Proceedings of ACL-08: HLT, pages 568–576, Columbus, Ohio, June.ACL.
[86] R. I. Damper , Y. Marchand , M. J. Adamson , K. Gustafson,” Evaluating the
Pronunciation Component of Text-to-Speech Systems for English: A Performance
Comparison of Different Approaches”, In Speech And Language
Technology (Salt) Club Workshop On Evaluation In Speech And Language
Technology,1997.
[87] Chomsky, N. and M. Halle (1968). “The Sound Pattern of English”. New York,
176
NY: Harper and Row, page4.
[88] Elovitz, H. S., R. Johnson, A. McHugh, and J. E. Shore (1976). “Letter-to-sound
rules for automatic translation of English text to phonetics”. IEEE Transactions
on Acoustics, Speech and Signal Processing ASSP-24, 446–459.
[89] Beata Megyesi,” Data-Driven Syntactic Analysis Methods and Applications for
Swedish”, Doctoral Dissertation, Kungl Tekniska Hogskolan,2002.
[90] Anil Kr. Jain, Jianchang Mao,K.M. Mohiuddin, “Artificial neural networks: a
tutorial”,IEE-Computer Volume:29,Issue:3 Page(s): 31 – 44,1996.
[91] Eniko Beatrice Bilcu,” Text-To-Phoneme Mapping Using Neural Networks”,
Doctoral Thesis,Tampere University of Technology,2008.
[92] Cover, T.M., & Hart, P.E. (1967).,” Nearest neighbor pattern classification”.
Institute of Electrical and ElectronicsEngineers Transactions on Information
Theory, 13, 21-27.
[93] Hart, P.E. (1968),” The condensed nearest neighbor rule”. Institute of Electrical
and Electronics Engineers and Transactions on Information Theory, 14, 515-516.
[94] Gates, G.W. (1972), “The reduced nearest neighbor rule”. IEEE Transactions on
Information Theory, 431-433.
177
[95] Dasarathy, B.V. (1980),”Nosing around the neighborhood: A new system structure
and classification rule for recognition in partially exposed environments”. Pattern
Analysis and Machine Intelligence, 2, 67-71.
[96] David W. Aha ,Dennis Kibler ,Marc K. Albert,” Instance-Based
Learning Algorithms”, Machine Learning, 6, 37-66 ,Kluwer Academic Publishers,
Boston. Manufactured in The Netherlands.1991.
[97] Tom Mitchell, ” Machine Learning”, McGraw Hill, 1997.
[98] Breiman, Friedman, Olshen, and Stone,“Classification and Regression Trees”.
Wadsworth Inc., 1984.
[99] Gelfand, S., Ravishankar, C., and Delp, E., “An Iterative Growing and Pruning
Algorithm for Classification Tree Design”. In IEEE Pattern Analysis and Machine
Intelligence, pages 163-174, Feb. 1991.
[100] Kuhn, R., Lazaridès, A., Normandin, Y., and Brousseau, J.,“Improved decision
trees for phonetic modeling”. In ICASSP95, pages 552–555, 1995.
[101] Pagel V., K. Lenzo, A.Black. 1998. “Letter-to-sound rules for accented lexicon
Compression”. In Proc. of ICSLP98, vol 5,2015-2020, Sydney, Australia.
178
[102] Lawrence R. Rabiner.” A tutorial on hidden Markov models and selected
applications in speech recognition”. Proc. IEEE, 77:257–286, 1989.
[103] R. Durbin, S. R. Eddy, A. Krogh, and G. J. Mitchison. “Biological Sequence
Analysis: Probabilistic Models of Proteins and Nucleic Acids”. Cambridge
University Press, Cambridge UK,1998.
[104] Ghahramani , Z., “ An Introduction to Hidden Markov Model and Bayesian
Networks”, Journal of Pattern Recognition and Artificial Intelligence, 15 (1):9-
42.2001.
[105] D.B. Paul,” .Speech Recognition Using Hidden Markov Models”, The Liru:oln
Laboratory Journal, Volume 3, Number 1 (l990).
[106] L E Baum,” An inequality and an associated maximization technique in statistical
estimation of probabilistic functions of a Markov process”, by L E Baum,
Inequalities (1972) ,Volume: 3, Issue: 1, Publisher: Academic Press, Pages: 1-8.
[107] Leonard E Baum, J A Eagonin ,”An inequality with applications to statistical
estimation for probabilistic functions of Markov processes and to a model for
ecology”,Bulletin of the American Mathematical Society(1967).
[108] L E Baum, T Petriein ,” Statistical inference for probabilistic functions of finite
state Markov chains”, The Annals of Mathematical Statistics(1966).
179
[109] L. E. Baum, T. Petrie, G. Soules, and N. Weiss, "A maximization technique
occurring in the statistical analysis of probabilistic functions of Markov chains",
Ann. Math. Statist., vol. 41, no. 1, pp. 164–171, 1970.
[110] J.K.Baker, ” The Dragon System “, IEEE.Trans.Acous. Speech Signal Process.
ASSP-23, 24(1975).
[111] J.M. Baker, “Personal Communication”. Jan 1990.
[112] F.Jelinck. ,” Continuous Speech Recognition by Statistical Methods”, proc.IEEE
64,532(1970).
[113] S.E Levinson,I.R. Rabiner and M.M Sandhi ,” An Introduction to the Application
of the theory of Probabilistic Function of a Markov Process to Automatic Speech
Recognition”,Bellsys.Tech J02,1035(1983).
[114] Tatyana Polyakova, Antonio Bonafonte,” Main issues in grapheme-to-phoneme
conversion for TTS”, Volume: 1, Issue: 35, Pages: 29-34, ISSN: 11355948.
[115] Walter Daelemans and Antal Van Den Bosch.” Language-independent data-
oriented grapheme-to-phoneme conversion “.In Progressin speech synthesis ,
pages 77–89. Springer, New York.1997.
[116] Arthur Dempster, Nan Laird, and Donald Rubin. 1977. “Maximum likelihood
180
from incomplete data via the EM algorithm”. In Journal of the Royal Statistical
Society,pages B:1–38.
[117] Black A.,K. Lenzo K. And V.Pagel.” Issues In building general letter to sound
rules.” In Proc.of the 3rd ESCA workshop on speech synthesis.,77-80, Jenolah
Caves,Australia.1998.
[118] Sittichai Jiampojamarn, Grzegorz Kondrak, and Tarek Sherif. 2007.” Applying
many-to-many alignments and hidden markov models to letter-to-phoneme
conversion”. In HLT 2007: The Conference of the NAACL;Proceedings of the
Main Conference, pages 372–379,Rochester, New York, April. ACL.
[119] Bisani M. and H. Ney. ”Investigations on joint –multigram model for grapheme-
to-phoneme conversion.” In Proceedings of the 7th Int. Conf. on Spoken
Language Processing, Denver, CO, vol.1, 105-108.
[120] Sittichai Jiampojamarn and Grzegorz Kondrak,” Letter-Phoneme Alignment: An
Exploration”, Proceedings of the 48th Annual Meeting of the Association for
Computational Linguistics, pages 780–788,Uppsala, Sweden, 11-16 July 2010. c
2010 Association for Computational Linguistics.
[121] Michel Divay,Anthony J. Vitale,” Algorithms for Grapheme-Phoneme
Translation for English and French: Applications for Database Searches and
Speech Synthesis”, Computational Linguistics ,Volume 23, Number 4,1997.
181
[122] Ainsworth, W. A. 1973.” A system for converting English text into speech”. In
IEEE Transactions of Audio and Electroacoustics, pages 288-290.
[123] Bakiri, G., and T. G. Dietterich. 1991.”Converting English Text to Speech: A
Machine Learning Approach”. Ph.D. thesis.Rep. No. 91-30-1. Department of
Computer Science, Oregon State University.
[124] Bernstein, J. and L. Nessly. 1981.” Performance comparison of component
Algorithms for the phonemicization of orthography”. In Proceedings of the 19th
Annual Meeting, Stanford University.Association for Computational Linguistics.
[125] Elovitz, H. S., R. W. Johnson, A. McHugh,and J. E. Shore. 1976. “ Automatic
translation of English text to phonetics by means of letter-to-sound rules “. NRL
Report 7948, Naval Research Laboratory,Washington, D.C.
[126] Hunnicut, S. 1976. “Phonological rules for a text-to-speech system. American
Journal of Computational Linguistics”, Microfiche 57.
[127] Levin, H. 1963. “A basic research program on reading”. Final Report,
Cooperative Research Project No. 639, Cornell University.
[128] McCormick, S. and S. R. Hertz. 1989. “A new approach to English text-to-
phoneme conversion using delta”, Version 2. 117th Meeting. Journal of the
Acoustical Society of America, Supplement 1, Vol. 85, $124.
182
[129] McIlroy, M. D. 1974. “Synthetic English speech by rules”. Bell Telephone
Laboratories Memo.
[130] O'Malley, M. H. 1990.” Text-to-speech conversion technology”. Computer
IEEE, page 17.
[131] Venezky, R. L. 1962. “A Computer Program for Deriving Spelling to Sound
Correlations”. MA thesis, Cornell University. Published in part in A Basic
Research Program on Reading.
[132] Venezky, R. L. 1967a. “English orthography:Its graphical structure and its
Relation to sound”. Reading Research Quarterly, II.
[133] Venezky, R. L. 1967b. “Reading:Grapheme-phonerne relationships”.Education
87: 519-524.
[134] Venezky, R. L. 1967c.” The basis of English orthography “. Acta Linguistica
10: 145- 159.
[135] Venezky, R. L. 1970. “The Structure of English Orthography”. Mouton, The
Hague.
[136] Venezky, R. L. and R. Weir. 1966. “A study of selected spelling-to-sound
correspondence patterns”. Final Report, Cooperative Research project No. 3090,
183
Stanford University.
[137] Vitale, A. J. 1991. “An algorithm for high accuracy name pronunciation by
parametric speech synthesizer”. Computational Linguistics 17(3).
[138] Weir, R. 1964. “Formulation of grapheme-phoneme correspondence rules
to aid the teaching of reading”. Final Report, Cooperative Research project
No. S-039 Stanford University.
[139] Aubergd, V. 1991. La synthese de la parole:"des regles au lexique". These,
Universitd Stendhal, Grenoble.
[140] Bechet F., T. Spriet, and M. E1-Beze. 1996.”Traitement spdcifique des noms
propres dans un systeme de transcription grapheme-phoneme”. JST Avignon.
[141] Catach, N. 1989.” Informatique: Traitement automatique du Langage”. Bulletin
Liaisons-Heso, September.
[142] Catach, N. and L. Catach 1992. Pr4sentation du logiciel VOISINETTE, "Un
correcteur entree phonetique". CNRS-INFOS.
[143] Cotto, D. 1992. Traitement automatique des textes en vue de la synthese vocale.
These,Universite Paul Sabatier, Toulouse III.
184
[144] Divay, M. 1984. “ De l'dcrit vers l'oral ou contribution a l'etude des traitements
des textes ecrits en vue de leur prononciation sur synthetiseur de parole. These
d'Etat, Universit~ de Rennes, France.
[145] Divay, M. 1985.” A text-processing expert system”. 5eme Congres
Reconnaissance des formes et Intelligence Artificielle ,Novembre 1985,
Grenoble, France.
[146] Divay, M. 1990a.” Traitement du langage naturel: la phonetisation ou comment
apprendre a l'ordinateur a lire un texte Francais”. MICRO-SYSTEMES, March.
[147] Divay, M. 1990b. A written processing expert system for text to phoneme
conversion. In Proceedings of the International Conference on Spoken Language
(ICSLP 90), Kobe, Japan.
[148] Divay, M. 1991. “CD-ROM Electronic Dictionary”, November.
[149] Laporte, E. 1988.” Methodes algorithmiques et lexicales de phonetisation de
Texts” .These, Universit4 Paris 7, May.
[150] Prouts, B. 1980,” Contribution a la synthese de la parole a partir de texte;
transcription graphemo-phondtique en temps reel sur microprocesseur”. These de
Docteur-Ingenieur, Universit4 de Paris Sud, Orsay.
[151] Yvon, F. 1996. “Prononcer par analogie: motivation, formalisation et fvaluation”.
185
These,Ecole nationale des Telecommunications,Paris.
[152] Dedina, M. J. and H. C. Nusbaum. 1991. “PRONOUNCE: A program for
pronunciation by analogy”. Computer Speech and Language 5:55-64.
[153] Sejnowski, T. J. and C. R. Rosenberg. 1987. “NETtalk: Parallel networks that
learn to pronounce English text”. Complex Systems 1:145-168.
[154] Lucas, S. M. and R. I. Damper. 1992.” Syntactic neural networks for bi-
directional text-phonetics translation”.In G. Bailly and C. Benoit, editors,
Talking Machines, Theories, Models and Designs.North-Holland Publishers.
[155] Hochberg, J., S. M. Mniszewski, T. Calleja,and G. J. Papcun. 1991. “A default
hierarchy for pronouncing English”. IEEETransactions on Pattern Matching and
Machine Intelligence 13(9): 957-964.
[156] Klatt, D. H. and D. W. Shipman. 1982.”Letter-to-phoneme rules: A semi-
Automatic discovery procedure”.Journal of the Acoustical Society of
America.82: 737-793.
[157] Lucassen, J. M. and R. L. Mercer. 1984. “An information theoretic approach to
the automatic determination of phonemic baseforms”. In Proceedings of
lCASSP-84,pages 42.5.1-42.5.3, San Diego.
[158] Parfitt, S. and R. Sharman. 1991. “A bi-directional model of English
186
pronunciation”. In Proceedings of Eurospeech,volume 2, pages 801-804.
[159] Golding, A. R. 1991. “Pronouncing Names by a Combination of Case-based and
Rule-based Reasoning”. Ph.D. Thesis, Stanford University.
[160] Meng, H. M. 1995. “Phonological Parsing for Bi-Directional Letter-to-Sound
and Sound-to-Letter Generation”. Ph.D. Thesis,MIT, Cambridge, MA.
[161] H.B Hudgson“Essay the First: on the Koch, Bodo and Dhimal Tribes”,1946.
[162] Pratima Brahma, ” L a n g u a g e i n I n d ia “, volume 1 2 , 2 0 1 2
A s s a m e s e I n f l u e n c e o n B o d o, ISSN 1930-2940.
[163] Madhu Ram Baro,”Structure of Boro language”,N.L Publications,2008.
[164] http://www.iitg.ernet.in/rcilts/
[165] Madhu Ram Baro, “The Historical Development of Boro Langguage”, N.L
Publications ,2007.
[166] Priyankoo Sarmah,” Some Aspects Of The Tonal Phonology Of Bodo”, A
Dissertation Of Master Of Philosophy In Linguistics Central Institute Of
English And Foreign Languages,2004.
[167] Weidert, Alfons.1987. Tibeto-Burman Tonology. John Benjamins Publishing
187
Company,Amsterdam/Philadelphia.
[168] Maassen Ben and Povel Dirk-Jan 1984,” The effect of correcting fundamental
frequency on the intelligibility of deaf speech and its interaction with temporal
aspects “. Journal of the Acoustical Society of America, 76 (6): 1673–1681.
[169] Laures, Jacqueline S. and Weismer, Gary 1999. The effects of a flattened
fundamental frequency on intelligibility at the sentence level. Journal of Speech
Language and Hearing Research, 42(5): 1148a–1156.
[170] Ineke Mennen,” Phonetic and Phonological influences in non-native intonation :
an overview for language teachers.”, QMUC Speech Science Research Centre
Working Paper WP9,2006.
[171] Professor Pat Kamalani Hurley,” Introduction to the Study of Language”,
University of Hawai'i - Leeward Community College.
[172] Paul Kerswill,” Migration and language”, An international handbook of the
Science of language and society”, 2nd edn., Vol 3. Berlin: De Gruyter.2006.
[173] O'Grady, William, John Archibald, Mark Aronoff, and Jane Rees-Miller. eds.
(2001) ,” Contemporary Linguistics”. Boston: Bedford/St. Martin's.
[174] Edward J. Vajda, ” Dialectology”, course material, Linguistics 201, Western
188
Washington University,2001.
[175] Dr. Pramod Chandra Bhattacharjya, “A Descriptive Analysis of the Boro
Language”, Doctoral Thesis, 1977.
[176] Phukan. Basumatary, (2005). “An Introduction to the Boro Language”, Mittali
Publications,New Delhi.
[177] Vera Demberg, Helmut Schmid, Gregor M¨ohler,” Phonological Constraints and
Morphological Preprocessing for Grapheme-to-Phoneme Conversion”,
Proceedings of the 45th Annual Meeting of the Association of Computational
Linguistics, pages 96–103,Prague, Czech Republic, June 2007. c 2007
Association for Computational Linguistics.
[178] ChandanSarma,U.Sharma,C.K.Nath,S.Kalita,P.H.Talukdar,” Selection of Units
and Development of Speech Database for Natural Sounding Bodo TTS
System”, CISP Guwahati ,March 2012.
[179] Parminder Singh, Gurpreet Singh Lehal, “Syllables Selection for the
Development of Speech Database for Punjabi TTS System”, IJCSI
International Journal of Computer Science Issues, Vol. 7, Issue 6, November
2010.
[180] R.A. Krakow,” Physiological organization of syllables: a review”, Journal of
189
Phonetics, Vol. 27, 1999, pp. 23-54.
[181] Susan Bartlett, Grzegorz Kondrak, Colin Cherry, “On the Syllabification of
Phonemes”, Human Language Technologies: The 2009 Annual Conference of
the North American Chapter of the ACL, pages 308– 316,Boulder, Colorado,
June 2009. c 2009 Association for Computational Linguistics.
[182] Y. A. El-Imam, “Phonetization of arabic: rules and algorithms”, Computer
Speech & Language, vol. 18, pp. 339–373, October 2004.
[183] R. Dale et al. (Eds.), “A Rule Based Syllabification Algorithm for Sinhala”,
IJCNLP 2005, LNAI 3651, pp. 438 – 449, 2005.© Springer-Verlag Berlin
Heidelberg 2005.
[184] Juliette Blevins,”The syllable in phonological theory”,1995.
[185] George Kiraz and Bernd M¨obius, “Multilingual syllabification using weighted
finite-state transducers”.In Proceedings of the 3rd Workshop on Speech
Synthesis,1998.
[186] Robert Damper. 2001. “Learning about speech from data: Beyond NETtalk”. In
Data-Driven Techniques in Speech Synthesis, pages 1–25. Kluwer Academic
Publishers.
190
[187] John R. Rickford,“How Linguists Approach the Study of Language and Dialect”
, ms. January 2002, for students in Ling 73, AAVE, Stanfo.
[188] Othman O. Khalifa, Zakiah Hanim Ahmad, Aisha-Hassan A. Hashim &Teddy
Suya Gunawan,” SMaTalk: Standard Malay Text to Speech Talk System”,
Signal Processing: An International Journal, Volume (2) : Issue (5).August 2008