| Quick reminder: DEADLINE for C4DM PhD studentships is next Tuesday, 31 Jan
        2012. -- ----- - -- ---- --- --- - - ------ - - Funded PhD studentships at theÂCentre
                    for Digital Music, Queen Mary University of
                  London Mathematical
                    Models for Musical Prosodic Gestures Supervisor:ÂÂProf. Elaine Chew (Centre for Digital Music) Co-supervisor:ÂÂDr. Oscar Bandtlow (School of Mathematical Sciences) Application deadline:Â31stÂJanuary 2012 In music performance studies, prosody is the musician-specific timing, stress, and sometimes intonation added when interpreting a notated score. Mid- to high-level music prosodic gestures, for example tempo trajectories, often invoke parallels in the physical world, such as a damped oscillator. This project seeks to identify and mathematically model such gestures. The mathematical descriptors will form the basis for a vocabulary of prosodic gestures for music. Â Ensemble
                    Interaction Over Distance Supervisor:ÂÂProf. Elaine Chew (Centre for Digital Music) Co-supervisor:ÂÂProf. Patrick Healey (Interaction Media and Communication) Application deadline:Â31stÂJanuary 2012 When a small group of musicians negotiate in performance (i.e. real-time) the shaping and execution of a collective interpretation, the communication is non-verbal; some of the cues can be embedded in the musical prosody, and some demonstrated through gestures. This project aims to capture, analyze, quantify, and model the cues necessary for effective and engaging ensemble performance, by studying both co-located as well as distributed (over the Internet) ensembles. ÂÂ Adaptive,
                    personalised digital musical instruments Supervisor:ÂÂDr. Andrew McPherson (Centre for Digital Music) Co-supervisor:ÂÂProf. Elaine Chew (Centre for Digital Music) Application deadline:Â31stÂJanuary 2012 A performer can take decades to learn a musical instrument. This studentship will focus on creating instruments that learn the capabilities and artistic preferences of the individual performer, with a particular focus on the relationship between physical gesture and sound production. The successful candidate will develop intelligent gesture-sound mapping strategies, which dynamically update based on feedback from the performer. User studies with professional and amateur musicians will be integral to all stages of the project, and the successful candidate will take a leading role in designing and conducting these studies. The project aims to make performance more accessible to beginning musicians while enabling new modes of _expression_ for experts. Â Probabilistic
                    Modelling of Temporal Expectations in Music Supervisor:ÂÂDr. Marcus Pearce (Centre for Digital Music) Co-supervisor:ÂÂDr. Matthew Purver (Interaction Media and Communication) Application deadline:Â31stÂJanuary 2012 The project's goal is to construct and evaluate computational models of human temporal expectation. It involves developing probabilistic models of temporal prediction, taking representational account of rhythm and metre. The models and parameters are optimised to maximise prediction performance and compared to human temporal expectations in empirical studies of listeners. Â Intelligent
                    Interfaces for Accelerating Intermediate Piano
                    LearningÂ(*) Supervisor:ÂÂProf. Elaine Chew (Centre for Digital Music) Sponsoring Company:ÂÂYamaha R&D Centre London Application deadline:Â31stÂJanuary 2012 Ubiquitous access to digital music, and the hours of practice required to master new pieces, has led to a decline in amateur instrument playing. Cognisance of music structure can facilitate planning and sequence production, and enhance music making pleasure. Machine intelligence can help diagnose areas of difficulty and offer targeted constructive assistance. The studentship will propose/evaluate score-based visualisations of music structure and gestures that accelerate intermediate piano piece mastery for young and adult learners. Candidates should be proficient at programming, experienced with user interface design, have some background in statistics, and possess at least amateur-level piano playing ability. Â Semantic
                    Audio: bringing audio signal analysis together with
                    future internet technologiesÂ(*) Supervisor:ÂÂProf. Mark Sandler (Centre for Digital Music) Sponsoring Company:ÂÂFocusrite/Novation Application deadline:Â31stÂJanuary 2012 The project is
                    concerned with analysis of musical content where it
                    is created (typically in studios), thus affording
                    much cleaner computer representation of
                    musicological information in the music that can then
                    be used both to enhance consumer experiences and
                    recording studio practices. We base the
                    representation on RDF and ontologies, which are the
                    technologies that underpin Open Data, Semantic Web
                    and the Internet of Things. We have collaborated
                    with organisations such as BBC, MusicBrainz and the
                    British Library in developing these principles. Content
                      (e.g. music, film, tv) recommendation and
                      discovery is reaching a level of maturity (for
                      example, last.fm and Genius). But today, these
                      content descriptions and semantics are derived
                      from theÂfinished productÂ(e.g. CD, MP3,
                      DVD). The research question explored in this PhD
                      relies on performing the audio signal content
                      analysis at theÂpoint of content creation.
                      By using ontologies and RDF (Resource Description
                      Framework  a superior version of XML), many new
                      user modalities are enabled. For example, new and
                      complex user queries/searches of the form, Âfind
                      songs in A minor, with lead and rhythm guitars,
                      less than 2m30secs, and a rhythm that modulates
                      between 90 and 120 beats per minuteÂ. Not only can
                      content semantics enhance the listener/consumer
                      experience, they also enhance the workflow in the
                      recording studio. This is where FocusriteÂs
                      interest lies, and especially integration with
                      OSC. ------In addition, there are 10(*)+1 or 2 funded studentships
              through theÂMedia
                  and Arts TechnologyÂdoctoral training centre.
              ÂLimited funding may be available for other PhD research
              at theÂCentre
                  for Digital Music. Marcus Pearce -- Lecturer in Sound and Music Processing Centre for Digital Music School of Electronic Engineering & Computer Science Queen Mary, University of London Mile End Road, London E1 4NS, UK Tel: +44 (0)20 7882 5352 http://webprojects.eecs.qmul.ac.uk/marcusp MSc in Digital Music
          ProcessingÂÂhttp://bit.ly/zb21lm MSc in Digital Signal
          ProcessingÂhttp://bit.ly/zdZ8cP BEng / MEng in Audio
          Systems EngineeringÂhttp://bit.ly/zL1TDS Ends of
                Audience Wkshp 30-31 MayÂhttp://bit.ly/zHYDdjÂ**
                Deadline 30 Jan 2012 ** CMMR 2012 Conference ÂÂ19-22 JunÂhttp://bit.ly/xJdUIaÂ** Deadline Â1 Feb 2012 ** |