Subject: [AUDITORY] Releasing the test set of FSDKaggle2019 dataset (used in DCASE 2019 Task2) From: Eduardo Fonseca <eduardo.fonseca@xxxxxxxx> Date: Fri, 24 Jan 2020 18:40:16 +0100 List-Archive:<http://lists.mcgill.ca/scripts/wa.exe?LIST=AUDITORY>--000000000000ca59f1059ce64091 Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable =3D=3D=3D Apologies for cross-posting =3D=3D=3D Dear list, We=E2=80=99re glad to announce we have released the full test set & labels = of FSDKaggle2019. This dataset was used for DCASE 2019 Task 2 <http://dcase.community/challenge2019/task-audio-tagging>, which was hosted on the Kaggle platform as a competition titled Freesound Audio Tagging 2019 <https://www.kaggle.com/c/freesound-audio-tagging-2019>. FSDKaggle2019 includes almost 30k audio clips amounting over 100h of audio, encompassing 80 classes drawn from the AudioSet Ontology <https://research.google.com/audioset/ontology/index.html>. It includes a human curated train set from Freesound (~5k clips, ~11h), a noisy train set from Flickr (~20k clips, ~80h), and a test set from Freesound (~4.5k clips, ~13h). The dataset allows development and evaluation of machine listening methods in conditions of label noise, minimal supervision, and real-world acoustic mismatch. FSDKaggle2019 is freely available from Zenodo: https://doi.org/10.5281/zenodo.3612637 <https://doi.org/10.5281/zenodo.3612637> You can find more details in our DCASE 2019 paper: E. Fonseca, M. Plakal, F. Font, D. P. W. Ellis, and X. Serra. Audio tagging with noisy labels and minimal supervision <https://arxiv.org/abs/1906.02975>. Detection and Classification of Acoustic Scenes and Events (DCASE) Workshop, NYC, USA, 2019 Both competition and dataset have been a collaboration between the Music Technology Group of Universitat Pompeu Fabra, and the Sound Understanding team at Google AI Perception. This effort was kindly sponsored by a Google Faculty Research Award 2018 <https://ai.googleblog.com/2019/03/google-faculty-research-awards-2018.html= > . Best, Eduardo, Manoj, Frederic, Dan and Xavier -- Eduardo Fonseca Music Technology Group Universitat Pompeu Fabra -- --000000000000ca59f1059ce64091 Content-Type: text/html; charset="UTF-8" Content-Transfer-Encoding: quoted-printable <div dir=3D"ltr"><p dir=3D"ltr" style=3D"line-height:1.38;text-align:justif= y;margin-top:12pt;margin-bottom:12pt" id=3D"gmail-docs-internal-guid-367574= 69-7fff-bc27-4af3-ca4a07121066"><font size=3D"2"><span style=3D"font-family= :Arial;color:rgb(0,0,0);background-color:transparent;font-weight:400;font-s= tyle:normal;font-variant:normal;text-decoration:none;vertical-align:baselin= e;white-space:pre-wrap">=3D=3D=3D Apologies for cross-posting =3D=3D=3D</sp= an></font></p><p dir=3D"ltr" style=3D"line-height:1.38;text-align:justify;m= argin-top:12pt;margin-bottom:12pt"><font size=3D"2"><span style=3D"font-fam= ily:Arial;color:rgb(0,0,0);background-color:transparent;font-weight:400;fon= t-style:normal;font-variant:normal;text-decoration:none;vertical-align:base= line;white-space:pre-wrap">Dear list,</span></font></p><p dir=3D"ltr" style= =3D"line-height:1.38;text-align:justify;margin-top:12pt;margin-bottom:12pt"= ><font size=3D"2"><span style=3D"font-family:Arial;color:rgb(0,0,0);backgro= und-color:transparent;font-weight:400;font-style:normal;font-variant:normal= ;text-decoration:none;vertical-align:baseline;white-space:pre-wrap">We=E2= =80=99re glad to announce we have released the full test set & labels o= f FSDKaggle2019. This dataset was used for</span><a href=3D"http://dcase.co= mmunity/challenge2019/task-audio-tagging" style=3D"text-decoration:none"><s= pan style=3D"font-family:Arial;color:rgb(0,0,0);background-color:transparen= t;font-weight:400;font-style:normal;font-variant:normal;text-decoration:non= e;vertical-align:baseline;white-space:pre-wrap"> </span><span style=3D"font= -family:Arial;color:rgb(17,85,204);background-color:transparent;font-weight= :400;font-style:normal;font-variant:normal;text-decoration:underline;vertic= al-align:baseline;white-space:pre-wrap">DCASE 2019 Task 2</span></a><span s= tyle=3D"font-family:Arial;color:rgb(0,0,0);background-color:transparent;fon= t-weight:400;font-style:normal;font-variant:normal;text-decoration:none;ver= tical-align:baseline;white-space:pre-wrap">, which was hosted on the Kaggle= platform as a competition titled</span><a href=3D"https://www.kaggle.com/c= /freesound-audio-tagging-2019" style=3D"text-decoration:none"><span style= =3D"font-family:Arial;color:rgb(0,0,0);background-color:transparent;font-we= ight:400;font-style:normal;font-variant:normal;text-decoration:none;vertica= l-align:baseline;white-space:pre-wrap"> </span><span style=3D"font-family:A= rial;color:rgb(17,85,204);background-color:transparent;font-weight:400;font= -style:normal;font-variant:normal;text-decoration:underline;vertical-align:= baseline;white-space:pre-wrap">Freesound Audio Tagging 2019</span></a><span= style=3D"font-family:Arial;color:rgb(0,0,0);background-color:transparent;f= ont-weight:400;font-style:normal;font-variant:normal;text-decoration:none;v= ertical-align:baseline;white-space:pre-wrap">.</span></font></p><p dir=3D"l= tr" style=3D"line-height:1.38;text-align:justify;margin-top:12pt;margin-bot= tom:12pt"><font size=3D"2"><span style=3D"font-family:Arial;color:rgb(0,0,0= );background-color:transparent;font-weight:400;font-style:normal;font-varia= nt:normal;text-decoration:none;vertical-align:baseline;white-space:pre-wrap= ">FSDKaggle2019 includes almost 30k audio clips amounting over 100h of audi= o, encompassing 80 classes drawn from the</span><a href=3D"https://research= .google.com/audioset/ontology/index.html" style=3D"text-decoration:none"><s= pan style=3D"font-family:Arial;color:rgb(0,0,0);background-color:transparen= t;font-weight:400;font-style:normal;font-variant:normal;text-decoration:non= e;vertical-align:baseline;white-space:pre-wrap"> </span><span style=3D"font= -family:Arial;color:rgb(17,85,204);background-color:transparent;font-weight= :400;font-style:normal;font-variant:normal;text-decoration:underline;vertic= al-align:baseline;white-space:pre-wrap">AudioSet Ontology</span></a><span s= tyle=3D"font-family:Arial;color:rgb(0,0,0);background-color:transparent;fon= t-weight:400;font-style:normal;font-variant:normal;text-decoration:none;ver= tical-align:baseline;white-space:pre-wrap">. It includes a human curated tr= ain set from Freesound (~5k clips, ~11h), a noisy train set from Flickr (~2= 0k clips, ~80h), and a test set from Freesound (~4.5k clips, ~13h). The dat= aset allows development and evaluation of machine listening methods in cond= itions of label noise, minimal supervision, and real-world acoustic mismatc= h.</span></font></p>FSDKaggle2019 is freely available from Zenodo: <a href= =3D"https://doi.org/10.5281/zenodo.3612637">https://doi.org/10.5281/zenodo.= 3612637=E2=80=A8</a><br>You<font size=3D"2"><span style=3D"font-family:Aria= l;color:rgb(0,0,0);background-color:transparent;font-weight:400;font-style:= normal;font-variant:normal;text-decoration:none;vertical-align:baseline;whi= te-space:pre-wrap"> can find more details in our DCASE 2019 paper: E. Fonse= ca, M. Plakal, F. Font, D. P. W. Ellis, and X. Serra.</span><a href=3D"http= s://arxiv.org/abs/1906.02975" style=3D"text-decoration:none"><span style=3D= "font-family:Arial;color:rgb(0,0,0);background-color:transparent;font-weigh= t:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-a= lign:baseline;white-space:pre-wrap"> </span><span style=3D"font-family:Aria= l;color:rgb(17,85,204);background-color:transparent;font-weight:400;font-st= yle:normal;font-variant:normal;text-decoration:underline;vertical-align:bas= eline;white-space:pre-wrap">Audio tagging with noisy labels and minimal sup= ervision</span></a><span style=3D"font-family:Arial;color:rgb(0,0,0);backgr= ound-color:transparent;font-weight:400;font-style:normal;font-variant:norma= l;text-decoration:none;vertical-align:baseline;white-space:pre-wrap">. Dete= ction and Classification of Acoustic Scenes and Events (DCASE) Workshop, NY= C, USA, 2019</span></font><p dir=3D"ltr" style=3D"line-height:1.38;text-ali= gn:justify;margin-top:12pt;margin-bottom:12pt"><font size=3D"2"><span style= =3D"font-family:Arial;color:rgb(0,0,0);background-color:transparent;font-we= ight:400;font-style:normal;font-variant:normal;text-decoration:none;vertica= l-align:baseline;white-space:pre-wrap">Both competition and dataset have be= en a collaboration between the Music Technology Group of Universitat Pompeu= Fabra, and the Sound Understanding team at Google AI Perception. This effo= rt was kindly sponsored by a</span><a href=3D"https://ai.googleblog.com/201= 9/03/google-faculty-research-awards-2018.html" style=3D"text-decoration:non= e"><span style=3D"font-family:Arial;color:rgb(0,0,0);background-color:trans= parent;font-weight:400;font-style:normal;font-variant:normal;text-decoratio= n:none;vertical-align:baseline;white-space:pre-wrap"> </span><span style=3D= "font-family:Arial;color:rgb(17,85,204);background-color:transparent;font-w= eight:400;font-style:normal;font-variant:normal;text-decoration:underline;v= ertical-align:baseline;white-space:pre-wrap">Google Faculty Research Award = 2018</span></a><span style=3D"font-family:Arial;color:rgb(0,0,0);background= -color:transparent;font-weight:400;font-style:normal;font-variant:normal;te= xt-decoration:none;vertical-align:baseline;white-space:pre-wrap">.</span></= font></p><p dir=3D"ltr" style=3D"line-height:1.38;text-align:justify;margin= -top:12pt;margin-bottom:12pt"><font size=3D"2"><span style=3D"font-family:A= rial;color:rgb(0,0,0);background-color:transparent;font-weight:400;font-sty= le:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;= white-space:pre-wrap">Best,</span></font></p><p dir=3D"ltr" style=3D"line-h= eight:1.38;text-align:justify;margin-top:12pt;margin-bottom:12pt"><font siz= e=3D"2"><span style=3D"font-family:Arial;color:rgb(0,0,0);background-color:= transparent;font-weight:400;font-style:normal;font-variant:normal;text-deco= ration:none;vertical-align:baseline;white-space:pre-wrap">Eduardo, Manoj, F= rederic, Dan and Xavier</span></font></p><div><div dir=3D"ltr" class=3D"gma= il_signature" data-smartmail=3D"gmail_signature"><div dir=3D"ltr"><div><div= dir=3D"ltr"><div><div dir=3D"ltr"><div><div dir=3D"ltr"><div><div dir=3D"l= tr"><div><div dir=3D"ltr"><div><div dir=3D"ltr"><div><div dir=3D"ltr"><div>= <div dir=3D"ltr"><div><div dir=3D"ltr"><div><div dir=3D"ltr">--</div><div d= ir=3D"ltr"><font size=3D"1">Eduardo Fonseca<br>Music Technology Group<br>Un= iversitat Pompeu Fabra</font><div><span style=3D"color:rgb(0,0,0)"><br></sp= an></div><div><span style=3D"color:rgb(255,255,255)">--</span><br></div><br= ></div></div></div></div></div></div></div></div></div></div></div></div></= div></div></div></div></div></div></div></div></div></div></div></div> --000000000000ca59f1059ce64091--