HOME > 상세정보

상세정보

Neural networks and speech processing

Neural networks and speech processing (2회 대출)

자료유형
단행본
개인저자
Morgan, David P., 1961- Scofield, Christopher L., 1957-.
서명 / 저자사항
Neural networks and speech processing / by David P. Morgan, Christopher L. Scofield ; foreword by Leon N. Cooper.
발행사항
Boston :   Kluwer Academic Publishers,   c1991.  
형태사항
xvi, 391 p. : ill. ; 25 cm.
총서사항
The Kluwer international series in engineering and computer science.VLSI, computer architecture, and digital signal processing.
ISBN
0792391446 (alk. paper)
서지주기
Includes bibliographical references (p. 359-382) and index.
일반주제명
Neural networks (Computer science). Speech processing systems.
000 00974camuuu200241 a 4500
001 000000921506
005 19990120093404.0
008 901207s1991 maua b 00110 eng
020 ▼a 0792391446 (alk. paper)
040 ▼a DLC ▼c DLC ▼d DLC ▼d 244002
049 0 ▼l 151031972
050 0 0 ▼a QA76.87 ▼b .M67 1991
082 0 0 ▼a 006.3 ▼2 20
090 ▼a 006.3 ▼b M847n
100 1 ▼a Morgan, David P., ▼d 1961-
245 1 0 ▼a Neural networks and speech processing / ▼c by David P. Morgan, Christopher L. Scofield ; foreword by Leon N. Cooper.
260 ▼a Boston : ▼b Kluwer Academic Publishers, ▼c c1991.
300 ▼a xvi, 391 p. : ▼b ill. ; ▼c 25 cm.
440 4 ▼a The Kluwer international series in engineering and computer science. ▼p VLSI, computer architecture, and digital signal processing.
504 ▼a Includes bibliographical references (p. 359-382) and index.
650 0 ▼a Neural networks (Computer science).
650 0 ▼a Speech processing systems.
700 1 ▼a Scofield, Christopher L., ▼d 1957-.

No. 소장처 청구기호 등록번호 도서상태 반납예정일 예약 서비스
No. 1 소장처 과학도서관/Sci-Info(2층서고)/ 청구기호 006.3 M847n 등록번호 121162331 도서상태 대출가능 반납예정일 예약 서비스 B M
No. 2 소장처 세종학술정보원/과학기술실(5층)/ 청구기호 006.3 M847n 등록번호 151031972 (2회 대출) 도서상태 대출가능 반납예정일 예약 서비스 B M ?
No. 소장처 청구기호 등록번호 도서상태 반납예정일 예약 서비스
No. 1 소장처 과학도서관/Sci-Info(2층서고)/ 청구기호 006.3 M847n 등록번호 121162331 도서상태 대출가능 반납예정일 예약 서비스 B M
No. 소장처 청구기호 등록번호 도서상태 반납예정일 예약 서비스
No. 1 소장처 세종학술정보원/과학기술실(5층)/ 청구기호 006.3 M847n 등록번호 151031972 (2회 대출) 도서상태 대출가능 반납예정일 예약 서비스 B M ?

컨텐츠정보

책소개

We would like to take this opportunity to thank all of those individ­ uals who helped us assemble this text, including the people of Lockheed Sanders and Nestor, Inc., whose encouragement and support were greatly appreciated. In addition, we would like to thank the members of the Lab­ oratory for Engineering Man-Machine Systems (LEMS) and the Center for Neural Science at Brown University for their frequent and helpful discussions on a number of topics discussed in this text. Although we both attended Brown from 1983 to 1985, and had offices in the same building, it is surprising that we did not meet until 1988. We also wish to thank Kluwer Academic Publishers for their profes­ sionalism and patience, and the reviewers for their constructive criticism. Thanks to John McCarthy for performing the final proof, and to John Adcock, Chip Bachmann, Deborah Farrow, Nathan Intrator, Michael Perrone, Ed Real, Lance Riek and Paul Zemany for their comments and assistance. We would also like to thank Khrisna Nathan, our most unbi­ ased and critical reviewer, for his suggestions for improving the content and accuracy of this text. A special thanks goes to Steve Hoffman, who was instrumental in helping us perform the experiments described in Chapter 9.

We would like to take this opportunity to thank all of those individ­ uals who helped us assemble this text, including the people of Lockheed Sanders and Nestor, Inc., whose encouragement and support were greatly appreciated. In addition, we would like to thank the members of the Lab­ oratory for Engineering Man-Machine Systems (LEMS) and the Center for Neural Science at Brown University for their frequent and helpful discussions on a number of topics discussed in this text. Although we both attended Brown from 1983 to 1985, and had offices in the same building, it is surprising that we did not meet until 1988. We also wish to thank Kluwer Academic Publishers for their profes­ sionalism and patience, and the reviewers for their constructive criticism. Thanks to John McCarthy for performing the final proof, and to John Adcock, Chip Bachmann, Deborah Farrow, Nathan Intrator, Michael Perrone, Ed Real, Lance Riek and Paul Zemany for their comments and assistance. We would also like to thank Khrisna Nathan, our most unbi­ ased and critical reviewer, for his suggestions for improving the content and accuracy of this text. A special thanks goes to Steve Hoffman, who was instrumental in helping us perform the experiments described in Chapter 9.


정보제공 : Aladin

목차


CONTENTS
Preface = xiii
foreword = xv
1 Introduction = 1
 1.1 Motivation = 1
 1.2 A Few Words on Speech Recognition = 3
 1.3 A Few Words on Neural Networks = 4
 1.4 Contents = 6
2 The Mammalian Auditory System = 9
 2.1 Introduction to Auditory Processing = 9
 2.2 The Anatomy and Physiology of Neurons = 9
 2.3 Neuroanatomy of the Auditory System = 12
  2.3.1 The Ear = 13
  2.3.2 The Cochlea = 15
  2.3.3 The Eighth Nerve = 18
  2.3.4 The Cochlear Nucleus = 24
  2.3.5 The Superior Olivary Complex = 28
  2.3.6 The Inferior Colliculus = 31
  2.3.7 The Medial Geniculate Nucleus = 32
  2.3.8 The Auditory Cortex = 35
 2.4 Recurrent Connectivity in the Auditory - Pathway = 37
 2.5 Summary = 39
3. An Artificial Neural Network Primer = 41
 3.1 A Neural Network Primer for Speech Scientists = 41
 3.2 Elements of Artificial Neural Networks = 42
  3.2.1 Similarity Measures and Activation Functions = 43
  3.2.2 Networks and Mappings = 48
 3.3 Learning in Neural Networks = 50
 3.4 Supervised Learning = 51
  3.4.1 The Perceptron and Gradient-Descent Learning = 51
  3.4.2 Associative Memories = 54
  3.4.3 The Hopfield Network = 57
 3.5 Multi-Layer Networks = 59
  3.5.1 The Restricted Coulomb Energy Network = 63
  3.5.2 The Backward Error Propagation Network = 69
  3.5.3 The Charge Clustering Network = 75
  3.5.4 Recurrent Back Propagation = 78
 3.6 Unsupervised Learning = 80
  3.6.1 The BCM Network = 81
  3.6.2 The Kohonen Feature Map = 84
 3.7 Summary = 88
4 A Speech Technology Primer = 91
 4.1 A Speech Primer for Neural Scientists = 91
 4.2 Human Speech Production / Perception = 92
  4.2.1 Information in the Speech Signal = 96
 4.3 ASR Technology = 102
  4.3.1 A General Speech Recognition Model = 103
 4.4 Signal Processing and Feature Extraction = 106
  4.4.1 Linear Predictive Coding = 110
  4.4.2 Feature Extraction and Modeling = 113
  4.4.3 Vector Quantization = 114
 4.5 Time Alignment and Pattern Matching = 115
  4.5.1 Dynamic Time Warping = 117
  4.5.2 Hidden Markov Models = 119
  4.5.3 Pronunciation Network Word Models = 126
 4.6 Language Models = 127
  4.6.1 Parsers = 128
  4.6.2 Statistical Models = 129
 4.7 Summary = 130
5 Methods in Neural Network Applications = 131
 5.1 The Allure of Neural Networks for Speech Processing = 131
 5.2 The Computational Properties of ANNs = 132
  5.2.1 Computability and Network Size = 135
 5.3 ANN Limitations :  The Scaling Problem = 141
  5.3.1 The Scaling of Learning = 142
  5.3.2 The Scaling of Generalization = 143
 5.4 Structured ANN Solutions = 145
  5.4.1 Hierarchical Modules = 145
  5.4.2 Hybrid Systems = 149
  5.4.3 Multiple Neural Network Systems = 156
  5.4.4 Integrating Neural Speech Modules = 158
 5.5 Summary = 161
6 Signal Processing and Feature Extraction = 163
 6.1 The Importance of Signal Representations = 163
 6.2 The Signal Processing Problem Domain = 163
 6.3 Biologically Motivated Signal Processing = 166
  6.3.1 Review of Speech Representation in the Auditory Nerve = 166
  6.3.2 The Silicon Cochlea and Temporal-Place Representations for ASR = 169
  6.3.3 The Role of Automatic Gain Control in Noisy Environments = 173
 6.4 ANNs for Conventional Signal Processing = 179
  6.4.1 Adaptive Filtering = 179
  6.4.2 A Noise Reduction Network = 183
 6.5 Feature Representations = 190
  6.5.1 Unsupervised Feature Extraction for Phoneme Classification = 191
  6.5.2 Feature Maps = 196
 6.6 Summary = 201
7 Time Alignment and Pattern Matching = 203
 7.1 Modeling Spectro-Temporal Structure = 203
 7.2 Time Normalization Via Pre-Processing = 204
  7.2.1 Interpolation and Decimation Techniques = 204
  7.2.2 Feature-Set Transformations = 206
 7.3 The Dynamic Programming Neural Network = 209
  7.3.1 The DPNN Architecture = 210
  7.3.2 The Time Warping Structure = 211
  7.3.3 The DPNN Training Procedure = 213
  7.3.4 Application to Speaker-Independent Digit Recognition = 214
 7.4 HMM Motivated Networks = 215
  7.4.1 The Viterbi Network = 216
  7.4.2 The HMM Network = 220
 7.5 Recurrent Networks for Temporal Modeling = 226
  7.5.1 The Temporal Flow Model = 226
  7.5.2 Temporal Flow Experiments = 228
 7.6 The Time Delay Neural Network = 230
  7.6.1 The TDNN Temporal Architecture = 231
  7.6.2 TDNN Training = 234
  7.6.3 Application to Phoneme Classification = 234
  7.6.4 Interpreting the TDNN Spectro-Temporal Representation = 236
  7.6.5 Phoneme Classification Summary = 240
  7.6.6 TDNNs for Word Discrimination = 242
 7.7 Summary = 243
8 Natural Language Processing = 245
 8.1 The Importance of Language Processing = 245
 8.2 Syntactic Models = 248
  8.2.1 NETgrams : An ANN Word Category Predictor = 253
  8.2.2 An ANN for Word Category Disambiguation = 260
  8.2.3 Recurrent Networks and Formal Languages = 264
 8.3 Semantic Models = 273
  8.3.1 Pronoun Reference ANNs = 274
 8.4 Knowledge Representation = 277
  8.4.1 Knowledge Representation in a Hopfield Network = 279
 8.5 Summary = 288
9 ANN Keyword Recognition = 289
 9.1 Keyword Spotting = 289
 9.2 The Primary KWS System = 291
  9.2.1 Experimental Data = 294
 9.3 DUR Experiments = 296
  9.3.1 Selecting a Fixed-Length Feature Representation = 296
  9.3.2 Single and Multiple Networks = 297
  9.3.3 Experiments with Hybrid Systems = 302
 9.4 Secondary Processing Experiments = 305
  9.4.1 The Pattern Matching Approach = 305
  9.4.2 An Investigation of Temporal Models = 313
 9.5 Summary = 326
10 Neural Networks and Speech Processing = 329
 10.1 Speech Processing Applications = 329
  10.1.1 Speech Synthesis = 331
  10.1.2 Speech Coding = 334
  10.1.3 Speaker Separation = 335
  10.1.4 Speech Enhancement = 336
  10.1.5 Speaker Verification / Identification = 337
  10.1.6 Language Identification = 338
  10.1.7 Keyword / Keyphrase Spotting = 338
 10.2 Summary of Efforts in ASR = 339
  10.2.1 The Past : Institutions Involved in ASR = 340
  10.2.2 The Current Status of ANNs in ASR = 342
  10.2.3 The Future : Challenges and Goals = 345
 10.3 Concluding Remarks = 348
A Faster Back Propagation = 349
 A.1 Methods for the Successful Application of ANNS = 349
  A.1.1 Gradient-Descent Heuristics = 349
  A.1.2 Second-Order Methods = 353
B Acronyms = 357
Bibliography = 359
Index = 383


관련분야 신착자료

Negro, Alessandro (2026)
Dyer-Witheford, Nick (2026)