[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

REMINDER: DEADLINE 31 Jan 2012 for digital music PhD studentships at Queen Mary University of London



Quick reminder: DEADLINE for C4DM PhD studentships is next Tuesday, 31 Jan 2012.

-- ----- - -- ---- --- --- - - ------ - -

Funded PhD studentships at theÂCentre for Digital Music, Queen Mary University of London
(details alsoÂhere). NB: Studentships marked with an asterisk (*) have UKÂresidencyÂrequirements.


Mathematical Models for Musical Prosodic Gestures
Supervisor:ÂÂProf. Elaine Chew (Centre for Digital Music)
Co-supervisor:ÂÂDr. Oscar Bandtlow (School of Mathematical Sciences)
Application deadline:Â31stÂJanuary 2012

In music performance studies, prosody is the musician-specific timing, stress, and sometimes intonation added when interpreting a notated score. Mid- to high-level music prosodic gestures, for example tempo trajectories, often invoke parallels in the physical world, such as a damped oscillator. This project seeks to identify and mathematically model such gestures. The mathematical descriptors will form the basis for a vocabulary of prosodic gestures for music. Â


Ensemble Interaction Over Distance
Supervisor:ÂÂProf. Elaine Chew (Centre for Digital Music)
Co-supervisor:ÂÂProf. Patrick Healey (Interaction Media and Communication)
Application deadline:Â31stÂJanuary 2012

When a small group of musicians negotiate in performance (i.e. real-time) the shaping and execution of a collective interpretation, the communication is non-verbal; some of the cues can be embedded in the musical prosody, and some demonstrated through gestures. This project aims to capture, analyze, quantify, and model the cues necessary for effective and engaging ensemble performance, by studying both co-located as well as distributed (over the Internet) ensembles. ÂÂ


Adaptive, personalised digital musical instruments
Supervisor:ÂÂDr. Andrew McPherson (Centre for Digital Music)
Co-supervisor:ÂÂProf. Elaine Chew (Centre for Digital Music)
Application deadline:Â31stÂJanuary 2012

A performer can take decades to learn a musical instrument. This studentship will focus on creating instruments that learn the capabilities and artistic preferences of the individual performer, with a particular focus on the relationship between physical gesture and sound production. The successful candidate will develop intelligent gesture-sound mapping strategies, which dynamically update based on feedback from the performer. User studies with professional and amateur musicians will be integral to all stages of the project, and the successful candidate will take a leading role in designing and conducting these studies. The project aims to make performance more accessible to beginning musicians while enabling new modes of _expression_ for experts. Â


Probabilistic Modelling of Temporal Expectations in Music
Supervisor:ÂÂDr. Marcus Pearce (Centre for Digital Music)
Co-supervisor:ÂÂDr. Matthew Purver (Interaction Media and Communication)
Application deadline:Â31stÂJanuary 2012

The project's goal is to construct and evaluate computational models of human temporal expectation. It involves developing probabilistic models of temporal prediction, taking representational account of rhythm and metre. The models and parameters are optimised to maximise prediction performance and compared to human temporal expectations in empirical studies of listeners. Â


Intelligent Interfaces for Accelerating Intermediate Piano LearningÂ(*)
Supervisor:ÂÂProf. Elaine Chew (Centre for Digital Music)
Sponsoring Company:ÂÂYamaha R&D Centre London
Application deadline:Â31stÂJanuary 2012

Ubiquitous access to digital music, and the hours of practice required to master new pieces, has led to a decline in amateur instrument playing. Cognisance of music structure can facilitate planning and sequence production, and enhance music making pleasure. Machine intelligence can help diagnose areas of difficulty and offer targeted constructive assistance. The studentship will propose/evaluate score-based visualisations of music structure and gestures that accelerate intermediate piano piece mastery for young and adult learners. Candidates should be proficient at programming, experienced with user interface design, have some background in statistics, and possess at least amateur-level piano playing ability. Â


Semantic Audio: bringing audio signal analysis together with future internet technologiesÂ(*)
Supervisor:ÂÂProf. Mark Sandler (Centre for Digital Music)
Sponsoring Company:ÂÂFocusrite/Novation
Application deadline:Â31stÂJanuary 2012

The project is concerned with analysis of musical content where it is created (typically in studios), thus affording much cleaner computer representation of musicological information in the music that can then be used both to enhance consumer experiences and recording studio practices. We base the representation on RDF and ontologies, which are the technologies that underpin Open Data, Semantic Web and the Internet of Things. We have collaborated with organisations such as BBC, MusicBrainz and the British Library in developing these principles.

Content (e.g. music, film, tv) recommendation and discovery is reaching a level of maturity (for example, last.fm and Genius). But today, these content descriptions and semantics are derived from theÂfinished productÂ(e.g. CD, MP3, DVD). The research question explored in this PhD relies on performing the audio signal content analysis at theÂpoint of content creation. By using ontologies and RDF (Resource Description Framework  a superior version of XML), many new user modalities are enabled. For example, new and complex user queries/searches of the form, Âfind songs in A minor, with lead and rhythm guitars, less than 2m30secs, and a rhythm that modulates between 90 and 120 beats per minuteÂ. Not only can content semantics enhance the listener/consumer experience, they also enhance the workflow in the recording studio. This is where FocusriteÂs interest lies, and especially integration with OSC.

------
In addition, there are 10(*)+1 or 2 funded studentships through theÂMedia and Arts TechnologyÂdoctoral training centre. ÂLimited funding may be available for other PhD research at theÂCentre for Digital Music.

Marcus Pearce
--
Lecturer in Sound and Music Processing
Centre for Digital Music
School of Electronic Engineering & Computer Science
Queen Mary, University of London
Mile End Road, London E1 4NS, UK
Tel: +44 (0)20 7882 5352
http://webprojects.eecs.qmul.ac.uk/marcusp

MSc in Digital Music ProcessingÂÂhttp://bit.ly/zb21lm
MSc in Digital Signal ProcessingÂhttp://bit.ly/zdZ8cP
BEng / MEng in Audio Systems EngineeringÂhttp://bit.ly/zL1TDS

Ends of Audience Wkshp 30-31 MayÂhttp://bit.ly/zHYDdjÂ** Deadline 30 Jan 2012 **
CMMR 2012 Conference ÂÂ19-22 JunÂhttp://bit.ly/xJdUIaÂ** Deadline Â1 Feb 2012 **