当前位置 :

新闻中心

关于乔治亚理工大学Chin-Hui Lee教授学术报告的通知

  讲座题目:

  Automatic Speech Attribute Transcription: Next-Generation Speech Processing

  主讲嘉宾:Chin-Hui Lee

  时 间:2013年7月16日上午10:00

  地 点:科大西区科技实验西楼一楼117会议室

  主办单位:语音及语言信息处理国家工程实验室

  报告摘要:

  The field of automatic speech recognition (ASR) has enjoyed more than 30 years of technology advancement due to the extensive utilization of the hidden Markov model (HMM) framework and a concentrated effort by the community to make available a vast amount of language resources. However the ASR problem is still far from being solved because not all information available in the speech knowledge hierarchy can be directly and effectively integrated into state-of-the-art systems to improve ASR performance and enhance system robustness. It is believed that some of the current knowledge insufficiency issues can be partially addressed by processing techniques that can take advantage of the full set of acoustic and language information in speech. On the other hand in human speech recognition (HSR) and spectrogram reading we often determine the linguistic identity of a sound based on detected cues and evidences that exist at various levels of the speech knowledge hierarchy, ranging from acoustic phonetics to syntax and semantics. This calls for a bottom-up knowledge integration framework that links speech processing with information extraction, by spotting speech cues with a bank of attribute detectors, weighing and combining acoustic evidences to form cognitive hypotheses, and verifying these theories until a consistent recognition decision can be reached. The recently proposed ASAT (automatic speech attribute transcription) framework is an attempt to mimic some HSR capabilities with asynchronous speech event detection followed by bottom-up speech knowledge integration and verification. In the last few years it has demonstrated potentials and offered insights in detection-based speech processing and information extraction.

  This presentation is intended to illustrate new possibilities of speech research via linking analysis and processing of raw speech signals with extracting multiple layers of useful speech information. By organizing these probabilistic evidences from the speech knowledge hierarchy, and integrating them into the already-powerful, top-down HMM framework we can facilitate a knowledge-rich, bottom-up and data-driven framework that will lower the entry barriers to ASR research and further enhance the capabilities and reduce some of the limitations in the state-of-the-art ASR systems. Everyone in and out of the current ASR community will be able to contribute to this worthwhile effort to building a collaborative ASR community of the 21st Century.

  嘉宾简介:

  Chin-Hui Lee is currently a Professor in the School of Electrical and Computer Engineering,Georgia Institute of Technology. His research interests include multimedia communication, multimedia signal and information processing, speech and speaker recognition, speech and language modeling, spoken dialogue processing, machine learning, biometric authentication, information retrieval, and bioinformatics. He is a member of the IEEE Signal Processing Society, Communication Society, Computer Society, and the European Speech Communication Association. He is also a lifetime member of the Computational Linguistics Society in Taiwan. He has published more than 300 papers and 25 patents on the subject of automatic speech and speaker recognition. For his long standing contributions to speech research he won in 2007 the IEEE Signal Processing Society's Technical Achievement Award for "Exceptional Contributions to the Field of Automatic Speech Recognition".