Subject: [AUDITORY] Research Fellow post in Machine Learning for Audio Captioning, at University of Surrey, UK (deadline: 8th April 2021) From: Wenwu Wang <000000615c5e5fae-dmarc-request@xxxxxxxx> Date: Mon, 5 Apr 2021 23:48:32 +0000--_000_DB9PR06MB77242B191C50C31FBD4D7688BA779DB9PR06MB7724eurp_ Content-Type: text/plain; charset="Windows-1252" Content-Transfer-Encoding: quoted-printable Apologies for cross-posting. Please feel free to circulate the following va= cancy to those who might be interested. Many thanks Research Fellow in Machine Learning for Audio Captioning Applications are invited for a Research Fellow (RF) position for 22 months = within the Centre for Vision Speech and Signal Processing (CVSSP), Universi= ty of Surrey, UK, to work on a project titled =93Automated Captioning of Im= age and Audio for Visually and Hearing Impaired=94, which is a collaborativ= e project between the University of Surrey and the Izmir Katip Celebi Unive= rsity (IKCU), Turkey, with project partners from charities and industrial s= ectors working with the hearing and visually impaired. This project aims to= address fundamental challenges in audio and image captioning, develop new = algorithms to improve performance of audio and image captioning algorithms,= and application tools that could be used by the hearing and visually impai= red to access audio and image content. The work at Surrey will focus on new methods and algorithms of automated au= dio captioning and natural language description of audio. This work is buil= t on the significant contributions of CVSSP in the area of acoustic scene a= nalysis, audio event detection, environmental sound recognition, and audio = tagging, together with preliminary results on audio captioning. This new pr= oject offers an opportunity to take this work to the next stages, and demon= strate the benefit of such technologies for the hearing and visually impair= ed. A smartphone based prototype will be developed for audio and visual cap= tioning jointly by Surrey and IKCU. New data will also be gathered, includi= ng audio-visual data for captioning, and user feedback for the prototype sy= stem. The postholder will be responsible for investigating and developing audio s= ignal processing, machine learning algorithms for natural language descript= ion of sound, and implementing software for prototyping the concept and alg= orithms. The postholder should have a doctoral level (or equivalent) resear= ch and development experience in electronic engineering, applied mathematic= s, computer science, artificial intelligence, machine learning, natural lan= guage processing, or related subjects. The postholder should ideally have e= xperience in one of the following areas: audio captioning, machine descript= ion of audio, audio classification, audio tagging, image captioning, video = captioning, translations between audio/image and texts, and/or translation = between audio and video. The post-holder will be based in CVSSP, and work under the direction of the= Principal Investigator Prof Wenwu Wang, with co-supervision by Prof Sabine= Braun, Director of the Centre for Translation Studies, at University of Su= rrey, and in collaboration with Dr Volkan Kilic, from the IKCU, Turkey. CVSSP is an International Centre of Excellence for research in Audio-Visual= Machine Perception, with over 150 researchers, a grant portfolio of =A324M= (=A317.5M EPSRC) from EPSRC, EU, InnovateUK, charity and industry, and a t= urnover of =A37M/annum. The Centre has state-of-the-art acoustic capture an= d analysis facilities and a Visual Media Lab with video and audio capture f= acilities supporting research in real-time video and audio processing and v= isualisation. CVSSP has a compute facility with 120 GPUs and >1PB of high-s= peed secure storage. Please apply online using the following link: https://jobs.surrey.ac.uk/vacancy.aspx?ref=3D015821 Best wishes, Wenwu -- Professor Wenwu Wang Centre for Vision Speech and Signal Processing Department of Electronic Engineering University of Surrey Guildford GU2 7XH United Kingdom Phone: +44 (0) 1483 686039 Fax: +44 (0) 1483 686031 Email: w.wang@xxxxxxxx http://personal.ee.surrey.ac.uk/Personal/W.Wang/ --_000_DB9PR06MB77242B191C50C31FBD4D7688BA779DB9PR06MB7724eurp_ Content-Type: text/html; charset="Windows-1252" Content-Transfer-Encoding: quoted-printable <html> <head> <meta http-equiv=3D"Content-Type" content=3D"text/html; charset=3DWindows-1= 252"> <style type=3D"text/css" style=3D"display:none;"> P {margin-top:0;margin-bo= ttom:0;} </style> </head> <body dir=3D"ltr"> <div style=3D"font-family: Calibri, Arial, Helvetica, sans-serif; font-size= : 12pt; color: rgb(0, 51, 0);"> <span style=3D"color: black; font-size: 12pt;">Apologies for cross-posting.= Please feel free to circulate the following vacancy to those who might be = interested. Many thanks</span><br> </div> <div> <div dir=3D"ltr"> <div dir=3D"ltr"> <div dir=3D"ltr"> <div style=3D"font-family:Calibri,Arial,Helvetica,sans-serif; font-size:12p= t; color:rgb(0,51,0)"> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; font-size:11pt; font-family:Calibri,sans-serif; margin-botto= m:.0001pt"> <b><span style=3D"font-size:16.0pt; color:black"><br> </span></b></p> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; font-size:11pt; font-family:Calibri,sans-serif; margin-botto= m:.0001pt"> <b><span style=3D"font-size:16.0pt; color:black">Research Fellow in </span>= </b><b><span style=3D"font-size:16.0pt; color:black">Machine Learning for A= udio Captioning </span></b></p> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; line-height:115%; font-size:11pt; font-family:Calibri,sans-s= erif; margin-bottom:.0001pt"> <span style=3D"font-size:10.0pt; color:black"> </span></p> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; line-height:115%; font-size:11pt; font-family:Calibri,sans-s= erif; margin-bottom:.0001pt"> <span style=3D"color:black">Applications are invited for a Research Fellow = (RF) position for 22 months within the Centre for Vision Speech and Signal = Processing (CVSSP), University of Surrey, UK, to work on a project titled = =93Automated Captioning of Image and Audio for Visually and Hearing Impaired=94, which is a collaborative proje= ct between the University of Surrey and the Izmir Katip Celebi University (= IKCU), Turkey, with project partners from charities and industrial sectors = working with the hearing and visually impaired. This project aims to address fundamental challenges in audio and= image captioning, develop new algorithms to improve performance of audio a= nd image captioning algorithms, and application tools that could be used by= the hearing and visually impaired to access audio and image content. </span></p> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; line-height:115%; font-size:11pt; font-family:Calibri,sans-s= erif; margin-bottom:.0001pt"> <span style=3D"color:black"> </span></p> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; line-height:115%; font-size:11pt; font-family:Calibri,sans-s= erif; margin-bottom:.0001pt"> <span style=3D"color:black">The work at Surrey will focus on new methods an= d algorithms of automated audio captioning and natural language description= of audio. This work is built on the significant contributions of CVSSP in = the area of acoustic scene analysis, audio event detection, environmental sound recognition, and audio tagging,= together with preliminary results on audio captioning. This new project of= fers an opportunity to take this work to the next stages, and demonstrate t= he benefit of such technologies for the hearing and visually impaired. A smartphone based prototype will b= e developed for audio and visual captioning jointly by Surrey and IKCU. New= data will also be gathered, including audio-visual data for captioning, an= d user feedback for the prototype system. </span></p> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; line-height:115%; font-size:11pt; font-family:Calibri,sans-s= erif; margin-bottom:.0001pt"> <span style=3D"color:black"> </span></p> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; line-height:115%; font-size:11pt; font-family:Calibri,sans-s= erif; margin-bottom:.0001pt"> <span style=3D"color:black">The postholder will be responsible for investig= ating and developing audio signal processing, machine learning algorithms f= or natural language description of sound, and implementing software for pro= totyping the concept and algorithms. The postholder should have a doctoral level (or equivalent) research and d= evelopment experience in electronic engineering, applied mathematics, compu= ter science, artificial intelligence, machine learning, natural language pr= ocessing, or related subjects. The postholder should ideally have experience in one of the following areas: a= udio captioning, machine description of audio, audio classification, audio = tagging, image captioning, video captioning, translations between audio/ima= ge and texts, and/or translation between audio and video. </span></p> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; line-height:115%; font-size:11pt; font-family:Calibri,sans-s= erif; margin-bottom:.0001pt"> <span style=3D"color:black"> </span></p> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; line-height:115%; font-size:11pt; font-family:Calibri,sans-s= erif; margin-bottom:.0001pt"> <span style=3D"color:black">The post-holder will be based in CVSSP, and wor= k under the direction of the Principal Investigator Prof Wenwu Wang, with c= o-supervision by Prof Sabine Braun, Director of the Centre for Translation = Studies, at University of Surrey, and in collaboration with Dr Volkan Kilic, from the IKCU, Turkey. </s= pan></p> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; line-height:115%; font-size:11pt; font-family:Calibri,sans-s= erif; margin-bottom:.0001pt"> <span style=3D"color:black"><span style=3D""> </span><br> CVSSP is an International Centre of Excellence for research in Audio-Visual= Machine Perception, with over 150 researchers, a grant portfolio of =A324M= (=A317.5M EPSRC) from EPSRC, EU, InnovateUK, charity and industry, and a t= urnover of =A37M/annum. The Centre has state-of-the-art acoustic capture and analysis facilities and a Visual Med= ia Lab with video and audio capture facilities supporting research in real-= time video and audio processing and visualisation. CVSSP has a compute faci= lity with 120 GPUs and >1PB of high-speed secure storage. </span></p> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; line-height:115%; font-size:11pt; font-family:Calibri,sans-s= erif; margin-bottom:.0001pt"> <br> </p> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; font-size:11pt; font-family:Calibri,sans-serif; margin-botto= m:.0001pt"> <span style=3D"color:black">Please apply online using the following link: &= nbsp;</span></p> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; font-size:11pt; font-family:Calibri,sans-serif; margin-botto= m:.0001pt"> <a href=3D"https://jobs.surrey.ac.uk/vacancy.aspx?ref=3D015821" id=3D"LPlnk= ">https://jobs.surrey.ac.uk/vacancy.aspx?ref=3D015821</a><span style=3D"col= or:black"> </span><br> </p> <div style=3D"font-family:Calibri,Arial,Helvetica,sans-serif; font-size:12p= t; color:rgb(0,51,0)"> <br> </div> <p class=3D"x_x_x_MsoNormal" style=3D"margin-top: 0px; margin-bottom: 0px;m= argin-top:0px; margin-bottom:0px; margin-top:0px; margin-bottom:0px; margin= :0cm 0cm 10pt; font-size:11pt; font-family:Calibri,sans-serif; margin-botto= m:.0001pt"> <br> </p> </div> <div> <div id=3D"x_x_x_Signature"> <div> <meta content=3D"text/html; charset=3DUTF-8"> <div id=3D"x_x_x_divtagdefaultwrapper" dir=3D"ltr" style=3D"font-size:12pt;= color:#003300; font-family:Calibri,Arial,Helvetica,sans-serif"> <div name=3D"x_x_x_divtagdefaultwrapper" style=3D"font-family:Calibri,Arial= ,Helvetica,sans-serif; font-size:; margin:0"> <div class=3D"x_x_x_BodyFragment"><font size=3D"2"> <div class=3D"x_x_x_PlainText"><font size=3D"3">Best wishes,<br> <br> Wenwu<br> <br> <br> <br> --<br> Professor Wenwu Wang<br> Centre for Vision Speech and Signal Processing <br> Department of Electronic Engineering <br> University of Surrey <br> Guildford GU2 7XH <br> United Kingdom<br> Phone: +44 (0) 1483 686039 <br> Fax: +44 (0) 1483 686031 <br> Email: w.wang@xxxxxxxx<br> <a href=3D"http://personal.ee.surrey.ac.uk/Personal/W.Wang/" class=3D"x_x_x= _OWAAutoLink">http://personal.ee.surrey.ac.uk/Personal/W.Wang/</a></font></= div> <div class=3D"x_x_x_PlainText"><br> <br> </div> </font></div> </div> </div> </div> </div> </div> </div> </div> </div> </div> </body> </html> --_000_DB9PR06MB77242B191C50C31FBD4D7688BA779DB9PR06MB7724eurp_--