Subject: [AUDITORY] CFP: IEEE JSTSP Special Issue - deadline extended till 11 Nov 2024 From: "Hussain, Amir" <0000016168b2549a-dmarc-request@xxxxxxxx> Date: Mon, 21 Oct 2024 19:06:47 +0000--_000_LO6P302MB00143239CA4731E74522C7E4D3432LO6P302MB0014GBRP_ Content-Type: text/plain; charset="iso-8859-1" Content-Transfer-Encoding: quoted-printable Dear List (with apologies for any cross-postings) Due to numerous requests from authors, the deadline for the Special Issue (SI) of the IEEE Journal of Selected Topics in Signal Processing (JSTSP) on "Deep Multimodal Speech Enhancement and Separation" has been extended till 11 November 2024 - see CFP below and here: https://signalprocessingsociety.org/publications-resources/special-issue-de= adlines/ieee-jstsp-special-issue-deep-multimodal-speech-enhancement-and-sep= aration CFP: IEEE Journal of Selected Topics in Signal Processing (JSTSP) Special Issue (SI) on: Deep Multimodal Speech Enhancement and Separation Manuscripts Due: 11 November 2024 (final extension) SI Publication Date: May 2025 Scope: Voice is the most commonly used modality by humans to communicate and psychologically blend into society. Recent technological advances have triggered the development of various voice-related applications in the information and communications technology market. However, noise, reverberation, and interfering speech are detrimental for effective communications between humans and other humans or machines, leading to performance degradation of associated voice-enabled services. To address the formidable speech-in-noise challenge, a range of speech enhancement (SE) and speech separation (SS) techniques are normally employed as important front-end speech processing units to handle distortions in input signals in order to provide more intelligible speech for automatic speech recognition (ASR), synthesis and dialogue systems. Emerging advances in artificial intelligence (AI) and machine learning, particularly deep neural networks, have led to remarkable improvements in SE and SS based solutions. A growing number of researchers have explored various extensions of these methods by utilising a variety of modalities as auxiliary inputs to the main speech processing task to access additional information from heterogeneous signals. In particular, multi-modal SE and SS systems have been shown to deliver enhanced performance in challenging noisy environments by augmenting the conventional speech modality with complementary information from multi-sensory inputs, such as video, noise type, signal-to-noise ratio (SNR), bone-conducted speech (vibrations), speaker, text information, electromyography, and electromagnetic midsagittal articulometer (EMMA) data. Various integration schemes, including early and late fusions, cross-attention mechanisms, and self-supervised learning algorithms, have also been successfully explored. Topics: This timely special issue aims to collate latest advances in multi-modal SE and SS systems that exploit both conventional and unconventional modalities to further improve state-of-the-art performance in benchmark problems. We particularly welcome submissions for novel deep neural network based algorithms and architectures, including new feature processing methods for multimodal and cross-modal speech processing. We also encourage submissions that address practical issues related to multimodal data recording, energy-efficient system design and real-time low-latency solutions, such as for assistive hearing and speech communication applications. Special Issue research topics of interest relate to open problems needing addressed These include, but are not limited to, the following. - Novel acoustic features and architectures for multi-modal SE (MM-SE) and multi-modal SS (MM-SS) solutions. - Self-supervised and unsupervised learning techniques for MM-SE and MM-SS systems. - Adversarial learning for MM-SE and MM-SS. - Large language model-based Generative approaches for MM-SE and MM-SS - Low-delay, low-power, low-complexity MM-SE and MM-SS models - Integration of multiple data acquisition devices for multimodal learning and novel learning algorithms robust to imperfect data. - Few-shot/zero-shot learning and adaptation algorithms for MM-SE and MM-SS systems with a small amount of training and adaptation data. - Approaches that effectively reduce model size and inference cost without reducing the speech quality and intelligibility of processed signals. - Novel objective functions including psychoacoustics and perceptually motivated loss functions for MM-SE and MM-SS - Holistic evaluation metrics for MM-SE and MM-SS systems. - Real-world applications and use-cases of MM-SE and MM-SS, including human-human and human-machine communications - Challenges and solutions in the integration of MM-SE and MM-SS into existing systems We encourage submissions that not only propose novel approaches but also substantiate the findings with rigorous evaluations, including real-world datasets. Studies that provide insights into the challenges involved and the impact of MM-SE and MM-SS on end-users are particularly welcome. Submission Guidelines: Manuscripts should be original and should not have been previously published or currently under consideration for publication elsewhere. All submissions will be peer-reviewed according to the IEEE Signal Processing Society review process. Authors should prepare their manuscripts according to the Instructions for Authors available from the Signal Processing Society website. Important Dates Manuscript Submission Deadline: 11 November 2024 First Review Due: 15 December 2024 Revised Manuscript Due: 15 January 2024 Second Review Due: 15 February 2024 Final Decision: 28 February 2025 Guest Editors: Amir Hussain, Edinburgh Napier University, UK Yu Tsao, Academia Sinica, Taiwan John H.L. Hansen, University of Texas at Dallas, USA Naomi Harte, Trinity College Dublin, Ireland Shinji Watanabe, Carnegie Mellon University, USA Isabel Trancoso, Instituto Superior T=E9cnico, IST, Univ. Lisbon, Portugal Shixiong Zhang, Tencent AI Lab, USA We look forward to your submissions. Many thanks, On behalf of the Guest Editorial Team -- Prof Amir Hussain School of Computing, Edinburgh Napier University, Scotland, UK E-mail: A.Hussain@xxxxxxxx<mailto:A.Hussain@xxxxxxxx> http://cogmhear.org<http://cogmhear.org/> This message and its attachment(s) are intended for the addressee(s) only a= nd should not be read, copied, disclosed, forwarded or relied upon by any p= erson other than the intended addressee(s) without the permission of the se= nder. If you are not the intended addressee you must not take any action ba= sed on this message and its attachment(s) nor must you copy or show them to= anyone. Please respond to the sender and ensure that this message and its = attachment(s) are deleted. It is your responsibility to ensure that this message and its attachment(s)= are scanned for viruses or other defects. Edinburgh Napier University does= not accept liability for any loss or damage which may result from this mes= sage or its attachment(s), or for errors or omissions arising after it was = sent. Email is not a secure medium. Emails entering Edinburgh Napier Univer= sity's system are subject to routine monitoring and filtering by Edinburgh = Napier University. Edinburgh Napier University is a registered Scottish charity. Registration = number SC018373 BSL users can contact us via contactSCOTLAND-BSL, the on-line British Sign = Language interpreting service. Find out more on the contactSCOTLAND website= . --_000_LO6P302MB00143239CA4731E74522C7E4D3432LO6P302MB0014GBRP_ Content-Type: text/html; charset="iso-8859-1" Content-Transfer-Encoding: quoted-printable <html> <head> <meta http-equiv=3D"Content-Type" content=3D"text/html; charset=3Diso-8859-= 1"> <style type=3D"text/css" style=3D"display:none;"> P {margin-top:0;margin-bo= ttom:0;} </style> </head> <body dir=3D"ltr"> <div class=3D"elementToProof" style=3D"font-family: Aptos, Aptos_EmbeddedFo= nt, Aptos_MSFontService, Calibri, Helvetica, sans-serif; font-size: 12pt; c= olor: rgb(0, 0, 0);"> Dear List (with apologies for any cross-postings)<br> <br> Due to numerous requests from authors, the deadline for the Special<br> Issue (SI) of the IEEE Journal of Selected Topics in Signal Processing<br> (JSTSP) on "Deep Multimodal Speech Enhancement and Separation" ha= s<br> been extended till 11 November 2024 - see CFP below and here:<br> <span style=3D"color: rgb(17, 85, 204);"><a href=3D"https://signalprocessin= gsociety.org/publications-resources/special-issue-deadlines/ieee-jstsp-spec= ial-issue-deep-multimodal-speech-enhancement-and-separation" target=3D"_bla= nk" id=3D"OWAff2f598a-dfb9-0c1f-883f-239ac527d2c3" class=3D"OWAAutoLink" re= l=3D"noreferrer" data-saferedirecturl=3D"https://www.google.com/url?q=3Dhtt= ps://signalprocessingsociety.org/publications-resources/special-issue-deadl= ines/ieee-jstsp-special-issue-deep-multimodal-speech-enhancement-and-separa= tion&source=3Dgmail&ust=3D1729623315416000&usg=3DAOvVaw0uYJlWuk= HSI9oOfCc31uLL" style=3D"color: rgb(17, 85, 204); text-align: left;">https:= //signalprocessingsociet<wbr style=3D"color:rgb(17, 85, 204)">y.org/publica= tions-resources/s<wbr style=3D"color:rgb(17, 85, 204)">pecial-issue-deadlin= es/ieee-js<wbr style=3D"color:rgb(17, 85, 204)">tsp-special-issue-deep-mult= imo<wbr style=3D"color:rgb(17, 85, 204)">dal-speech-enhancement-and-<wbr st= yle=3D"color:rgb(17, 85, 204)">separation</a></span><br> <br> CFP: IEEE Journal of Selected Topics in Signal Processing (JSTSP)<br> Special Issue (SI) on: Deep Multimodal Speech Enhancement and Separation<br= > <br> Manuscripts Due: 11 November 2024 (final extension)<br> SI Publication Date: May 2025<br> <br> Scope:<br> <br> Voice is the most commonly used modality by humans to communicate and<br> psychologically blend into society. Recent technological advances have<br> triggered the development of various voice-related applications in the<br> information and communications technology market. However, noise,<br> reverberation, and interfering speech are detrimental for effective<br> communications between humans and other humans or machines, leading to<br> performance degradation of associated voice-enabled services. To<br> address the formidable speech-in-noise challenge, a range of speech<br> enhancement (SE) and speech separation (SS) techniques are normally<br> employed as important front-end speech processing units to handle<br> distortions in input signals in order to provide more intelligible<br> speech for automatic speech recognition (ASR), synthesis and dialogue<br> systems. Emerging advances in artificial intelligence (AI) and machine<br> learning, particularly deep neural networks, have led to remarkable<br> improvements in SE and SS based solutions. A growing number of<br> researchers have explored various extensions of these methods by<br> utilising a variety of modalities as auxiliary inputs to the main<br> speech processing task to access additional information from<br> heterogeneous signals. In particular, multi-modal SE and SS systems<br> have been shown to deliver enhanced performance in challenging noisy<br> environments by augmenting the conventional speech modality with<br> complementary information from multi-sensory inputs, such as video,<br> noise type, signal-to-noise ratio (SNR), bone-conducted speech<br> (vibrations), speaker, text information, electromyography, and<br> electromagnetic midsagittal articulometer (EMMA) data. Various<br> integration schemes, including early and late fusions, cross-attention<br> mechanisms, and self-supervised learning algorithms, have also been<br> successfully explored.<br> <br> Topics:<br> <br> This timely special issue aims to collate latest advances in<br> multi-modal SE and SS systems that exploit both conventional and<br> unconventional modalities to further improve state-of-the-art<br> performance in benchmark problems. We particularly welcome submissions<br> for novel deep neural network based algorithms and architectures,<br> including new feature processing methods for multimodal and<br> cross-modal speech processing. We also encourage submissions that<br> address practical issues related to multimodal data recording,<br> energy-efficient system design and real-time low-latency solutions,<br> such as for assistive hearing and speech communication applications.<br> <br> Special Issue research topics of interest relate to open problems<br> needing addressed These include, but are not limited to, the<br> following.<br> - Novel acoustic features and architectures for multi-modal SE (MM-SE)<br> and multi-modal SS (MM-SS) solutions.<br> - Self-supervised and unsupervised learning techniques for MM-SE and<br> MM-SS systems.<br> - Adversarial learning for MM-SE and MM-SS.<br> - Large language model-based Generative approaches for MM-SE and MM-SS<br> - Low-delay, low-power, low-complexity MM-SE and MM-SS models<br> - Integration of multiple data acquisition devices for multimodal<br> learning and novel learning algorithms robust to imperfect data.<br> - Few-shot/zero-shot learning and adaptation algorithms for MM-SE and<br> MM-SS systems with a small amount of training and adaptation data.<br> - Approaches that effectively reduce model size and inference cost<br> without reducing the speech quality and intelligibility of processed<br> signals.<br> - Novel objective functions including psychoacoustics and perceptually<br> motivated loss functions for MM-SE and MM-SS<br> - Holistic evaluation metrics for MM-SE and MM-SS systems.<br> - Real-world applications and use-cases of MM-SE and MM-SS, including<br> human-human and human-machine communications<br> - Challenges and solutions in the integration of MM-SE and MM-SS into<br> existing systems<br> <br> We encourage submissions that not only propose novel approaches but<br> also substantiate the findings with rigorous evaluations, including<br> real-world datasets. Studies that provide insights into the challenges<br> involved and the impact of MM-SE and MM-SS on end-users are<br> particularly welcome.<br> <br> Submission Guidelines:<br> <br> Manuscripts should be original and should not have been previously<br> published or currently under consideration for publication elsewhere.<br> All submissions will be peer-reviewed according to the IEEE Signal<br> Processing Society review process. Authors should prepare their<br> manuscripts according to the Instructions for Authors available from<br> the Signal Processing Society website.<br> <br> Important Dates<br> <br> Manuscript Submission Deadline: 11 November 2024<br> First Review Due: 15 December 2024<br> Revised Manuscript Due: 15 January 2024<br> Second Review Due: 15 February 2024<br> Final Decision: 28 February 2025<br> <br> Guest Editors:<br> Amir Hussain, Edinburgh Napier University, UK<br> Yu Tsao, Academia Sinica, Taiwan<br> John H.L. Hansen, University of Texas at Dallas, USA<br> Naomi Harte, Trinity College Dublin, Ireland<br> Shinji Watanabe, Carnegie Mellon University, USA<br> Isabel Trancoso, Instituto Superior T=E9cnico, IST, Univ. Lisbon, Portugal<= br> Shixiong Zhang, Tencent AI Lab, USA<br> <br> We look forward to your submissions.<br> <br> Many thanks,</div> <div class=3D"elementToProof" style=3D"font-family: Aptos, Aptos_EmbeddedFo= nt, Aptos_MSFontService, Calibri, Helvetica, sans-serif; font-size: 12pt; c= olor: rgb(0, 0, 0);"> <br> On behalf of the Guest Editorial Team</div> <div class=3D"elementToProof" style=3D"font-family: Aptos, Aptos_EmbeddedFo= nt, Aptos_MSFontService, Calibri, Helvetica, sans-serif; font-size: 12pt; c= olor: rgb(0, 0, 0);"> <br> --<br> Prof Amir Hussain</div> <div class=3D"elementToProof" style=3D"font-family: Aptos, Aptos_EmbeddedFo= nt, Aptos_MSFontService, Calibri, Helvetica, sans-serif; font-size: 12pt; c= olor: rgb(0, 0, 0);"> School of Computing, </div> <div class=3D"elementToProof" style=3D"font-family: Aptos, Aptos_EmbeddedFo= nt, Aptos_MSFontService, Calibri, Helvetica, sans-serif; font-size: 12pt; c= olor: rgb(0, 0, 0);"> Edinburgh Napier University, Scotland, UK</div> <div class=3D"elementToProof" style=3D"font-family: Aptos, Aptos_EmbeddedFo= nt, Aptos_MSFontService, Calibri, Helvetica, sans-serif; font-size: 12pt; c= olor: rgb(0, 0, 0);"> E-mail: <span style=3D"color: rgb(17, 85, 204);"><a href=3D"mailto:A.Hussai= n@xxxxxxxx" target=3D"_blank" id=3D"OWA1c557dd5-6a29-0ca0-f6ca-aa7e731d= d15c" class=3D"OWAAutoLink" style=3D"color: rgb(17, 85, 204); text-align: l= eft;">A.Hussain@xxxxxxxx</a></span></div> <div class=3D"elementToProof" style=3D"font-family: Aptos, Aptos_EmbeddedFo= nt, Aptos_MSFontService, Calibri, Helvetica, sans-serif; font-size: 12pt; c= olor: rgb(17, 85, 204);"> <a href=3D"http://cogmhear.org/" target=3D"_blank" id=3D"OWAfd32a43d-1f17-5= 094-188d-8f033ebfce36" class=3D"OWAAutoLink" rel=3D"noreferrer" data-safere= directurl=3D"https://www.google.com/url?q=3Dhttp://cogmhear.org&source= =3Dgmail&ust=3D1729623315416000&usg=3DAOvVaw19vKc-2gdVfmZxmi1WJ8iZ"= style=3D"color: rgb(17, 85, 204); text-align: left;">http://cogmhear.org</= a></div> <div id=3D"x_x_Signature"> <div style=3D"background-color: rgb(255, 255, 255); margin: 0px;"></div> </div> <table width=3D"100%" cellspacing=3D"0" cellpadding=3D"=3D"0"" bo= rder=3D"0" align=3D"left" style=3D"font-family: Arial, Helvetica, san-serif= ; font-size:10pt;"> <tbody> <tr> <td style=3D"padding: 5px;"> <p>This message and its attachment(s) are intended for the addressee(s) onl= y and should not be read, copied, disclosed, forwarded or relied upon by an= y person other than the intended addressee(s) without the permission of the= sender. If you are not the intended addressee you must not take any action based on this message and its attac= hment(s) nor must you copy or show them to anyone. Please respond to the se= nder and ensure that this message and its attachment(s) are deleted.</p> <p>It is your responsibility to ensure that this message and its attachment= (s) are scanned for viruses or other defects. Edinburgh Napier University d= oes not accept liability for any loss or damage which may result from this = message or its attachment(s), or for errors or omissions arising after it was sent. Email is not a secure m= edium. Emails entering Edinburgh Napier University's system are subject to = routine monitoring and filtering by Edinburgh Napier University.</p> <p>Edinburgh Napier University is a registered Scottish charity. Registrati= on number SC018373</p> <p>BSL users can contact us via contactSCOTLAND-BSL, the on-line British Si= gn Language interpreting service. Find out more on the contactSCOTLAND webs= ite.</p> </td> </tr> </tbody> </table> </body> </html> --_000_LO6P302MB00143239CA4731E74522C7E4D3432LO6P302MB0014GBRP_--