~INTEGRATED MULTIMODAL SCORE-FOLLOWING ENVIRONMENT Martin Ritter School of Music University of British Columbia Vancouver, Canada Keith Hamel School of Music University of British Columbia Vancouver, Canada Bob Pritchard School of Music University of British Columbia Vancouver, Canada ABSTRACT The Integrated Multimodal Score-following Environment (IMuSE) [5] is a software project aimed at developing a system for the creation, rehearsal and performance of score-based interactive computer music compositions. An enhanced version of the NoteAbilityPro [7] music notation software is the central controller in IMuSE. The score contains conventional notes and performance indications as well as discrete and continuous control messages that can be sent to other applications such as MaxMSP [15] or Pure data (Pd) [11] during performance. As well, multiple modes of score-following [2] can be used to synchronize the live performance to the score. Score-following strategies include pitch-tracking of monophonic instruments, pitchtracking and amplitude-tracking of polyphonic instruments and gesture-tracking of performers' hand movements. In this paper, we present an overview of the IMuSE system, with a focus on its abilities to monitor and coordinate multiple pitch and gesture trackers. As an example of gesture-tracking, we show how video analysis is used to follow piano hand movements. We discuss strategies for negotiating between multiple streams of score-following data and we assess the strengths, limitations and future research directions of the IMuSE system. 1. BACKGROUND Score-following has a relatively long history, with the first research being done in the early 1980s by Dannenberg [5, 6] and Vercoe [24, 25]. This research was aimed at automated accompaniment where the computer would generate midi or audio tracks that were synchronized to a live musical performance. In the following decade, Puckette [17, 18] developed the explode Max object that could be used for monophonic score-following of pitch-tracked or midi data. In the early 2000s, score-following systems such as suivi- [20, 21] and antescofo- [4] were developed using Hidden Markov Models [1]; these systems prove to be far more robust in performance situations where machine perception and performer mistakes need to be accommodated. More recently, audio analysis has been used for music retrieval systems such as Query-byHumming [12] and intelligent audio editors [7]. Our focus is on score-following in live interactive computer music compositions involving several performers and where complex audio interactions between the live instruments required. and the computer generated sounds are 2. INTRODUCTION One of the challenges of creating score-based interactive computer music compositions is that there is often no connection between the notated score and the programming environment used to produce the electroacoustic layers of the composition or to process the live instruments. As a result, composers must develop strategies for synchronizing the live and electroacoustic components of their composition using manual cues, foot pedals or rigid timing controls. As well, the fact that several disconnected software applications are being used in an interactive performance can make rehearsing difficult and the editing of the composition can be awkward since it often means altering several different software components. The Integrated Multimodal Score-following Environment (IMuSE) has been developed to alleviate some of the issues associated with creating, rehearsing and performing score-based interactive computer music compositions. While this system could be used in some improvisatory performances, it is primarily designed for compositions that have a well-defined score and a where a high degree of synchronization between the live performance and the electroacoustic processing is needed. 3. SYSTEM OVERVIEW The central software component in IMuSE is the notation program NoteAbilityPro. This application has been enhanced to allow MaxMSP messages, multichannel breakpoint functions (BPFs), extended notes, conTimbre [9] controls and antescofo- action messages to be added to the score and aligned to other score events. As well, NoteAbilityPro can be externally controlled through network UDP messages so that synchronization information from pitch-tracking and gesture-tracking of a live performance can be used to alter the playback of the score. In order to accommodate conventional playback methods, remote messaging and synchronization with the live performance, music staves in a NoteAbilityPro score can have different functions: they can contain music notation that is intended for performers to play, notes and sound files that are intended to be played as audio events directly from the score, messages that are intended to control external applications, or notation and continuous graphical data 185 2013 ICMC
Top of page Top of page