This book presents a summary of the cognitively inspired basis behind multimodal speech
enhancement covering the relationship between audio and visual modalities in speech as well
as recent research into audiovisual speech correlation. A number of audiovisual speech
filtering approaches that make use of this relationship are also discussed. A novel multimodal
speech enhancement system making use of both visual and audio information to filter speech is
presented and this book explores the extension of this system with the use of fuzzy logic to
demonstrate an initial implementation of an autonomous adaptive and context aware multimodal
system. This work also discusses the challenges presented with regard to testing such a system
the limitations with many current audiovisual speech corpora and discusses a suitable approach
towards development of a corpus designed to test this novel cognitively inspired speech
filtering system.