Journal Information
Computer Speech and Language
https://www.sciencedirect.com/journal/computer-speech-and-language
Impact Factor:
3.100
Publisher:
Elsevier
ISSN:
0885-2308
Viewed:
16947
Tracked:
30
Call For Papers
An official publication of the International Speech Communication Association (ISCA)

Computer Speech & Language publishes reports of original research related to the recognition, understanding, production, coding and mining of speech and language.

The speech and language sciences have a long history, but it is only relatively recently that large-scale implementation of and experimentation with complex models of speech and language processing has become feasible. Such research is often carried out somewhat separately by practitioners of artificial intelligence, computer science, electronic engineering, information retrieval, linguistics, phonetics, or psychology.

The journal provides a focus for this work, and encourages an interdisciplinary approach to speech and language research and technology. Thus contributions from all of the related fields are welcomed in the form of reports of theoretical or experimental studies, tutorials, reviews, and brief correspondence pertaining to models and their implementation, or reports of fundamental research leading to the improvement of such models.

Research Areas Include

    Algorithms and models for speech recognition and synthesis
    Natural language processing for speech understanding and generation
    Statistical computational linguistics
    Computational models of discourse and dialogue
    Information retrieval, extraction and summarization
    Speaker and language recognition
    Computational models of speech production and perception
    Signal processing for speech analysis, enhancement and transformation
    Evaluation of human and computer system performance
Last updated by Dou Sun in 2024-07-16
Special Issues
Special Issue on Security and Privacy in Speech Communication
Submission Date: 2024-08-05

Speech and voice play a crucial role in self-expression and communication. They are used for interacting with virtual assistants, conveying emotions, and establishing identity. Addressing security and privacy concerns in speech communication is increasingly important for various sectors, including commercial, forensic, and government applications. To enhance security and privacy in user-centric human/machine interactions, we must explore the intersection of speech technology with AI ethics, trust, safety, and regulation. This special issue emerged from the Symposium on Security & Privacy in Speech Communication (SPSC) and encourages interdisciplinary research collaboration. It focuses on advancing security and privacy for speech technology across diverse applications such as voice assistants, medical use cases, and public domains. We invite submissions of original research, review articles, and position papers that contribute to security and privacy methods, tools, and approaches. The landscape for security and privacy in speech technology is rapidly evolving. For instance, healthcare applications have often prioritized diagnostics over patient privacy and data security. Voice privacy now employs advanced deep learning techniques and extends to various applications. Audio deepfake threats continue to grow in complexity, making them harder to detect and posing ethical challenges as synthetic speech technology advances. In summary, this special issue seeks to foster innovation and research in the field of security and privacy in speech communication, in line with the evolving landscape of technology and ethics. Guest editors: Jun.-Prof. Ingo Siegert (Executive Guest Editor) Institute for Information and Communication Technology, Faculty of Electrical Engineering and Information Technology, Otto-von-Guericke-University Magdeburg, Germany Email: siegert@ovgu.de Areas of Expertise: Artificial Intelligence, Machine Learning, Audio Signal Processing, Speaker and Speech Analyses, Voice Conversion Asst. Prof. Jennifer Williams School of Electronics and Computer Science, University of Southampton, United Kingdom Email: j.williams@soton.ac.uk Areas of Expertise: Audio Signal Processing, Natural Language Processing, Machine Learning Asst. Prof. Sneha Das Section for Statistics and Data Analysis, Department of Applied Mathematics and Computer Science, Technical University of Denmark, Denmark Email: sned@dtu.dk Areas of Expertise: Low-resource Machine Learning, Speech Processing, Statistics, Trustworthy AI, Social Signal Processing Special issue information Potential topics include: Privacy-preserving speech communication Speech Recognition and Processing Speech Perception, Production, and Acquisition Speech Synthesis Speech Coding and Enhancement Speaker and Language Identification Phonetics, Phonology, and Prosody Speech Paralinguistics Voice Biometrics Voice data sharing Machine Learning for Speech Technology Federated Learning Disentangled Representations Differential Privacy Distributed Learning Human-Computer Interfaces with Speech as a Medium Usable Security and Privacy Ubiquitous and Pervasive Computing Cognitive Science Ethics & Law Privacy and Data Protection Identity Management Electronic Mobile Commerce Data in Digital Media Speech Technology Development and Practice Acceptance and Trust Studies User Experience research on practice Co-development across disciplines Future studies Situated Ethics STS perspectives Manuscript submission information: Tentative Dates: Submission Open Date: December 7, 2023 Final Manuscript Submission Deadline: August 5, 2024 Editorial Acceptance Deadline: September 13, 2024 Contributed full papers must be submitted via Computer Speech & Language online submission system (Editorial Manager®): https://www.editorialmanager.com/ycsla/default2.aspx. Please select the article type “VSI: Security and Privacy in Speech Communication” when submitting the manuscript online. Please refer to the Guide for Authors to prepare your manuscript: https://www.elsevier.com/journals/computer-speech-and-language/0885-2308/guide-for-authors For any further information, the authors may contact the Guest Editors. Keywords: Security, Privacy, Trust, Speech Communication, Ethics Law, Acceptance
Last updated by Dou Sun in 2024-07-16
Special Issue on Multi-Speaker, Multi-Microphone, and Multi-Modal Distant Speech Recognition
Submission Date: 2024-12-02

Automatic speech recognition (ASR) has significantly progressed in the single-speaker scenario, owing to extensive training data, sophisticated deep learning architectures, and abundant computing resources. Building on this success, the research community is now tackling real-world multi-speaker speech recognition, where the number and nature of the sound sources are unknown and changing over time. In this scenario, refining core multi-speaker speech processing technologies such as speech separation, speaker diarization, and robust speech recognition is essential, and the effective integration of these advancements becomes increasingly crucial. In addition, emerging approaches, such as end-to-end neural networks, speech foundation models, and advanced training methods (e.g., semi-supervised, self-supervised, and unsupervised training) incorporating multi-microphone and multi-modal information (such as video and accelerometer data), offer promising avenues to alleviate these challenges. This special issue gathers recent advances in multi-speaker, multi-microphone, and multi-modal speech processing studies to establish real-world conversational speech recognition. Guest editors: Assoc. Prof. Shinji Watanabe (Executive Guest Editor) Carnegie Mellon University, Pittsburgh, Pennsylvania, United States of America Email: shinjiw@ieee.org Areas of Expertise: Speech recognition, speech enhancement, and speaker diarization Dr. Michael Mandel Reality Labs, Meta, Menlo Park, California, United States of America Email: mmandel@meta.com Areas of Expertise: Source separation, noise robust ASR, electromyography Dr. Marc Delcroix NTT Corporation, Chiyoda-Ku, Japan Email: marc.delcroix@ieee.org; marc.delcroix@ntt.com Areas of Expertise: Robust speech recognition, speech enhancement, source separation and extraction Dr. Leibny Paola Garcia Perera Johns Hopkins University, Baltimore, Maryland, United States of America Email: lgarci27@jhu.edu Areas of Expertise: Speech recognition, speech enhancement, and speaker diarization, multimodal speech processing Dr. Katerina Zmolikova Meta, Menlo Park, California, United States of America Email: kzmolikova@meta.com Areas of Expertise: Speech separation and extraction, speech enhancement, robust speech recognition Dr. Samuele Cornell Carnegie Mellon University, Pittsburgh, Pennsylvania, United States of America Email: scornell@andrew.cmu.edu Areas of Expertise: Robust speech recognition, speech separation and enhancement Special issue information: Relevant research topics include (but are not limited to): Speaker identification and diarization Speaker localization and beamforming Single- or multi-microphone enhancement and source separation Robust features and feature transforms Robust acoustic and language modeling for distant or multi-talker ASR Traditional or end-to-end robust speech recognition Training schemes: data simulation and augmentation, semi-supervised, self-supervised, and unsupervised training for distant or multi-talker speech processing Pre-training and fine-tuning of speech and audio foundation models and their application to distant and multi-talker speech processing Robust speaker and language recognition Robust paralinguistics Cross-environment or cross-dataset performance analysis Environmental background noise modeling Multimodal speech processing Systems, resources, and tools for distant Speech Recognition In addition to traditional research papers, the special issue also hopes to include descriptions of successful conversational speech recognition systems where the contribution is more in the implementation than the techniques themselves, as well as successful applications of conversational speech recognition systems. For example, the recently concluded seventh and eighth CHiME challenges serve as a focus for discussion in this special issue. The challenge considered the problem of conversational speech separation, speech recognition, and speaker diarization in everyday home environments from multi-microphone and multi-modal input. Seventh and eighth CHiME challenges consist of multiple tasks based on 1) distant automatic speech recognition with multiple devices in diverse scenarios, 2) unsupervised domain adaptation for conversational speech enhancement, 3) distant diarization and ASR in natural conferencing environments, and 4) ASR for multimodal conversations in smart glasses. Papers reporting evaluation results on the CHiME-7/8 datasets or other datasets dealing with real-world conversational speech recognition are equally welcome. Manuscript submission information: Tentative Dates: Submission Open Date: August 19, 2024 Manuscript Submission Deadline: December 2, 2024 Editorial Acceptance Deadline: September 1, 2025 Contributed full papers must be submitted via Computer Speech & Language online submission system (Editorial Manager®): https://www.editorialmanager.com/ycsla/default2.aspx. Please select the article type “VSI: Multi-DSR” when submitting the manuscript online. Please refer to the Guide for Authors to prepare your manuscript: https://www.elsevier.com/journals/computer-speech-and-language/0885-2308/guide-for-authors For any further information, the authors may contact the Guest Editors. Keywords: Speech recognition, speech enhancement/separation, speaker diarization, multi-speaker, multi-microphone, multi-modal, Distant Speech Recognition, CHiME challenge
Last updated by Dou Sun in 2024-07-16
Related Journals
Related Conferences
CCFCOREQUALISShortFull NameSubmissionNotificationConference
FSPSEInternational Conference on Frontiers of Signal Processing and Software Engineering2022-11-152022-11-202022-11-25
aa*a1SODAACM-SIAM Symposium on Discrete Algorithms2024-07-05 2025-01-12
ICCISInternational Conference on Computational and Information Sciences2014-03-102014-03-202014-05-30
bab1SATInternational Conference on Theory and Applications of Satisfiability Testing2024-03-082024-05-222024-08-21
AAME'International Conference on Aerospace, Aerodynamics and Mechatronics Engineering2022-06-02 2022-07-22
ICETCAInternational Conference on Electronics Technology and Computer Applications2020-07-152020-07-202020-07-28
SaCoNeTInternational Conference on Smart Communications in Network Technologies2018-07-312018-09-052018-10-27
AmIEuropean Conference on Ambient Intelligence2019-07-192019-08-122019-11-13
HPTSInternational Workshop on High Performance Transaction Systems 2011-06-012011-10-23
WebSciACM Web Science2023-11-302024-01-312024-05-21
Recommendation