13
"hearing places you inside an event [and] hearing brings us into the living world.” Journal Issue 3, December 2014 Strategies for spatial music performance: the practicalities and aesthetics of responsive systems design Ricky Graham, Brian Bridges DOI: 10.5920/divp.2015.33 This article will explore practical and aesthetic questions concerning spatial music performance by interrogating new developments within an emerging hyperinstrumental practice. The performance system is based on an electric guitar with individuated audio outputs per string and multichannel loudspeaker array. A series of spatial music mapping strategies will explore in-kind relationships between a formal melodic syntax model and an ecological flocking simulator, exploiting broader notions of embodiment underpinning the metaphorical basis for the experience and understanding of musical structure. The extension and refinement of this system has been based on a combination of practice-led and theoretical developments. The resulting mapping strategies will forge new gestural narratives between physical and figurative gestural planes, culminating in a responsive, bodily based, and immersive spatial music performance practice. The operation of the performance system is discussed in relation to supporting audiovisual materials. Keywords spatial music, mapping, embodied cognition, hyperinstrument, gesture, aesthetics IInnttrroodduuccttiioonn:: ffrraammiinngg ssppaattiiaall mmaappppiinnggss iinn ppeerrffoorrmmaannccee aanndd ddeessiiggnn AA ssyysstteemm ffoorr ssppaattiiaall iiddeeaass ((tthheeoorryy aanndd oouuttlliinnee ooff ddeessiiggnn)) Spatial music, by definition, treats space as a central creative parameter of musical experience, as opposed to an ancillary context. However, in practice, the question of how space may be musically significant provokes two somewhat divergent tactics: one more formalistic in nature, the other more perceptually based. An intuitively attractive approach treats space as largely analogous to the most common formal parameters of historical composition, pitch and rhythm. For example, Stockhausen’s Gesang der Jünglinge (Stockhausen, 1956) sought to apply permutational pitch structures directly to spatial structure (Smalley, 2000, pp. 6–7) via specific speaker assignments, in tandem with degrees of reverberation (Bates, 2010, pp. 131–132; Moritz, 2002). However, in spite of such rigorous process-based formalism, perceptual grouping factors may tend to dominate, especially given the basic spatialisation technique of alternating outputs via a limited array of five speakers (Bates, 2010, pp. 131–132; Smalley, 2000, p. 6). For example, Stockhausen’s later spatial work, Gruppen (Stockhausen, 1957), for multiple orchestras, focused explicitly on space’s contribution to perceptual grouping – (cf Bregman, 1990, pp. 293–302) – rather than space as a vehicle for simple formalism (Stockhausen, quoted in Moritz, 2002, cited in Bates, 2010, p. 133). Stockhausen (1975) later reflected on the problems inherent in a rigid adherence to serial processes without taking perceptual and environmental factors into account, advocating for more relationally based and empirically grounded approaches; see Bates (2010, p. 136–137 for further commentary). Approaches such as the latter treat space more explicitly as a framing device. As discussed in Bates (ibid., pp. 115–126), this type of approach became important in the twentieth century for enhancing the perception of complex musical materials in the work of composers such as Charles Ives and Henry Brant. More recently, this perspective has been theorised by Emmerson (1994; 2007, pp. 97–102) as a relational space frame typology for musical activity within different spheres of a sound environment (either real or virtual, created or evoked by electronic or digital processing).Smalley’s (2007) account of space in acousmatic music (and environmental experience) also approaches the issue from the perspective of framing. In a similar fashion, Sterne (2012, p. 9) describes a set of common defining tropes within sound studies, such as “hearing is spherical, vision is directional ... hearing immerses its subject ... hearing places you inside an event [and] hearing brings us into the living world.” Here, spatial concepts contribute to the definition of Issue 3, December 2014 Spatial Sound: Creative Practice in Electroacoustic Music by Prof. Eric Lyon Multichannel sound and spatial sound creation at Sporobole: A short account of live performance, studio design, outdoor multichannel audio, and visiting artists by Philippe-Aubert Gauthier Strategies for spatial music performance: the practicalities and aesthetics of responsive systems design by Ricky Graham, Brian Bridges An alternative approach to 3D audio recording and reproduction by Augustine Leudar Aural Territories: how phenomenology taught me how to compose electroacoustic spatial music by Frederico Macedo New developments for spatial music in the context of the ZKM Klangdom: A review of technologies and recent productions by Ludger Brümmer, Götz Dipper, David Wagner, Holger Stenschke, Jochen Arne Otto Exploded sounds: spatialised partials in two recent multi- channel installations by Nye Parry Audium – sound-sculptured space by Stan Shaff Home About Journal Resources CeReNeM Editorial Board Submissions Contact Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/... 1 of 13 2/12/15, 10:47 AM

Rgbb Divergence Issue3 2014

Embed Size (px)

DESCRIPTION

Strategies for spatial music performance: the practicalities and aesthetics ofresponsive systems design

Citation preview

Page 1: Rgbb Divergence Issue3 2014

"hearing places youinside an event [and]

hearing brings us intothe living world.”

JournalIssue 3, December 2014

Strategies for spatial music performance: the practicalities and aesthetics ofresponsive systems design

Ricky Graham, Brian Bridges

DOI: 10.5920/divp.2015.33

This article will explore practical and aesthetic questions concerning spatial music performance by

interrogating new developments within an emerging hyperinstrumental practice. The performance

system is based on an electric guitar with individuated audio outputs per string and multichannel

loudspeaker array. A series of spatial music mapping strategies will explore in-kind relationships

between a formal melodic syntax model and an ecological flocking simulator, exploiting broader

notions of embodiment underpinning the metaphorical basis for the experience and understanding

of musical structure. The extension and refinement of this system has been based on a combination

of practice-led and theoretical developments. The resulting mapping strategies will forge

new gestural narratives between physical and figurative gestural planes, culminating in a

responsive, bodily based, and immersive spatial music performance practice. The operation of the

performance system is discussed in relation to supporting audiovisual materials.

Keywords

spatial music, mapping, embodied cognition, hyperinstrument, gesture, aesthetics

IInnttrroodduuccttiioonn:: ffrraammiinngg ssppaattiiaall mmaappppiinnggss iinn ppeerrffoorrmmaannccee aanndd ddeessiiggnn

AA ssyysstteemm ffoorr ssppaattiiaall iiddeeaass ((tthheeoorryy aanndd oouuttlliinnee ooff ddeessiiggnn))

Spatial music, by definition, treats space as a central creative parameter of musical experience, as

opposed to an ancillary context. However, in practice, the question of how space may be musically

significant provokes two somewhat divergent tactics: one more formalistic in nature, the other more

perceptually based. An intuitively attractive approach treats space as largely analogous to the most

common formal parameters of historical composition, pitch and rhythm. For example, Stockhausen’s

Gesang der Jünglinge (Stockhausen, 1956) sought to apply permutational pitch structures directly to

spatial structure (Smalley, 2000, pp. 6–7) via specific speaker assignments, in tandem with degrees of

reverberation (Bates, 2010, pp. 131–132; Moritz, 2002). However, in spite of such rigorous process-based

formalism, perceptual grouping factors may tend to dominate, especially given the basic spatialisation

technique of alternating outputs via a limited array of five speakers (Bates, 2010, pp. 131–132; Smalley,

2000, p. 6). For example, Stockhausen’s later spatial work, Gruppen (Stockhausen, 1957), for multiple

orchestras, focused explicitly on space’s contribution to perceptual grouping – (cf Bregman, 1990, pp.

293–302) – rather than space as a vehicle for simple formalism (Stockhausen, quoted in Moritz, 2002,

cited in Bates, 2010, p. 133). Stockhausen (1975) later reflected on the problems inherent in a rigid

adherence to serial processes without taking perceptual and environmental factors into account,

advocating for more relationally based and empirically grounded approaches; see Bates (2010, p.

136–137 for further commentary).

Approaches such as the latter treat space more explicitly as a framing device. As discussed in Bates

(ibid., pp. 115–126), this type of approach became important in the twentieth century for enhancing the

perception of complex musical materials in the work of composers such as Charles Ives and Henry

Brant. More recently, this perspective has been theorised by Emmerson (1994; 2007, pp. 97–102) as a

relational space frame typology for musical activity within different spheres of a sound environment

(either real or virtual, created or evoked by electronic or digital processing).Smalley’s (2007) account of

space in acousmatic music (and environmental experience) also

approaches the issue from the perspective of framing. In a similar

fashion, Sterne (2012, p. 9) describes a set of common defining tropes

within sound studies, such as “hearing is spherical, vision is directional

... hearing immerses its subject ... hearing places you inside an event

[and] hearing brings us into the living world.” Here, spatial concepts contribute to the definition of

Issue 3, December 2014Spatial Sound: Creative Practice

in Electroacoustic Music

by Prof. Eric Lyon

Multichannel sound and spatial

sound creation at Sporobole: A

short account of live

performance, studio design,

outdoor multichannel audio, and

visiting artists

by Philippe-Aubert Gauthier

Strategies for spatial music

performance: the practicalities

and aesthetics of responsive

systems design

by Ricky Graham, Brian Bridges

An alternative approach to 3D

audio recording and

reproduction

by Augustine Leudar

Aural Territories: how

phenomenology taught me how

to compose electroacoustic

spatial music

by Frederico Macedo

New developments for spatial

music in the context of the ZKM

Klangdom: A review of

technologies and recent

productions

by Ludger Brümmer, Götz

Dipper, David Wagner, Holger

Stenschke, Jochen Arne Otto

Exploded sounds: spatialised

partials in two recent multi-

channel installations

by Nye Parry

Audium – sound-sculptured

space

by Stan Shaff

Home About Journal Resources CeReNeM Editorial Board Submissions Contact

Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/...

1 of 13 2/12/15, 10:47 AM

Page 2: Rgbb Divergence Issue3 2014

hearing via concepts of framing, immersion, and relationships within a wider sonic environment. Our

own approach follows Emmerson’s (1994, 2007); the delineation of different spheres of performative

activity and the sonic responses: local (singular or foregrounded events/streams, connected with

performer activity) and field (the wider environment) frames; see figure 1.

FFiigguurree 11 Local and field space frames, after Emmerson (2007, p. 98)

We believe that Emmerson’s ideas provide an intuitively accessible means of organising spatial control

in systems design and creative practice. Furthermore, due to their environmental logic, we believe that

they can be easily extended via a conceptual framework of spatial relations derived from current

theories of embodied cognition (Brower, 2000, 2008; Johnson, 1987; Johnson, 2007; Lakoff, 1987;

Solomon, 2007). These theories of embodiment provide a foundation from which to explore the

integration of performance gesture, control structures, and the performance space within a compatible

unifying framework (a shared gestural typology for space). As such, each stage of our performance

system, from performance gesture, via control mapping, to output, is treated in terms of an

environmental spatial logic, enhancing its potential for iterative development and extensibility as new

processes and controls are added.

DDeessiiggnn rraattiioonnaallee,, ccoonnttrrooll ssttrruuccttuurreess,, aanndd ssppaattiiaall ffrraammiinngg

Our spatial music performance system (or hyperinstrument) is based on an electric guitar with an

individuated audio pickup for multichannel audio (one channel per string) output (Graham, 2012, pp.

102–109). Aside from the enhanced pickup systems, physical affordances were essentially unmodified:

there were no “bolt-ons” of additional control surfaces or sensors (Graham, 2012; Graham and Bridges,

2013, 2014) on the guitar itself; see also Levitin (2002), which informed this perspective.. The central

motivation behind our approach is the intention to create a system that makes accessible connections

between familiar performance gestures from the guitar (specifically, detected notes, note groupings, and

note articulations) and the system’s spatialised output; see figure 2.

FFiigguurree 22 Outline of spatial music system, with parsing and spatialisation of multichannel audio feed

Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/...

2 of 13 2/12/15, 10:47 AM

Page 3: Rgbb Divergence Issue3 2014

As a result, the multichannel audio feed becomes our primary source of control data, specifically

focusing on the extraction and organisation of pitched (monophonic, note-based materials) materials.

Note event information is parsed from the audio feed, including pitch class, spectral content, and note

inter-onset (time between each note attack) data for each voice. This data is then integrated to provide

macrostructures, most importantly pitch contours, which provide an intuitive initial vehicle for spatial

mappings.

[VViiddeeoo eexxaammppllee 11 Basic parametric control, 6’20” http://www.youtube.com/watch?v=TFyO1xa4l14]

The note data is extracted in Pure Data (Pd) and applied to the control of first-order ambisonic

spatialisation azimuth angle (direction) and distance parameters. The system’s pitch tracking obtains

twelve pitch class divisions (pc0–pc11) in relation to a user-defined tonal centre (pc0) using the

time-domain, “Specially Normalized AutoCorrelation” or SNAC-based [helmholtz~] object for Pd (Vetter,

2011). The first version of the system controlled azimuth direction based on a simple cyclical mapping,

with pc0–pc11 mapped to 0–360 degrees. The distance parameters were mapped via a cognitive tonal

model from Lerdahl (2001; Lerdahl and Krumhansl, 2007), discussed further below, mapping tonally

central materials to a central (local) position and tonally peripheral materials to peripheral (field)

positions. In tandem with a frontal perspective from stage-based amplification, the system facilitates a

relational dialogue between local and field via the input tonal materials. More conventional guitar

voicings will reinforce events that are localised to the local/stage frames. Less conventional voicings will

activate a more pronounced off-centre spatial response (and more obvious/extended system responses),

activating the field frame via a spatial array surrounding the audience (see figure 3). This outlines a

basic relational dynamic to our use of space. Emmerson’s local/field distinctions and relationships are

thus articulated through: (1) the local of the more conventional guitar materials via on-stage monitoring

and (2) the field response, when the system’s diffused/spatialised responses predominate.

[VViiddeeoo eexxaammppllee 22 Basic parametric control 0’06” http://www.youtube.com/watch?v=TFyO1xa4l14]

FFiigguurree 33 Local and field frames as defined by the system’s responses

Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/...

3 of 13 2/12/15, 10:47 AM

Page 4: Rgbb Divergence Issue3 2014

EEmmbbooddiieedd ssppaacceess aanndd aanniimmaatteedd rreessppoonnsseess

EEmmbbooddiieedd ppiittcchh ssppaaccee aanndd ssppaattiiaall mmaappppiinnggss

Our approach seeks to match input performance gestures to spatial gestures in a manner that facilitates

accessibility for the performer and clarity of results for the audience. Thus, we adapted Emmerson’s

(2007) space frame typology as a means to provide a clear and intuitive way to present localisable and

non-localisable musicals materials within a physical performance space. We initially sought to collate

note event data (pitch classes) into a continuous macro-level melodic contour stream (0 to 1), which was

then used to control the ambisonic spatialisation of each audio channel, reifying tonal structures

through the relative spatial placement of the guitar’s individual voices. This approach established a

somewhat intuitive narrative between real-time instrumental voicings and the spatial locations and

trajectories of each audio output per string. Our model for the tonal-spatial mapping was then

developed to accommodate the tonal pitch space theories of Lerdahl (2001; Lerdahl and Krumhansl,

2007), themselves based upon the cognitive studies of Krumhansl (1990). These theories provided a set of

spatial dynamics that embodied cyclical positions (mapped to ambisonic azimuth angle) along with a

tonal distance factor relative to tonal centre (mapped to ambisonic distance). As such, they provided a

compatible, isomorphic base for mapping tonal structures to spatial forms.

Figure 4 depicts Lerdahl’s basic space, which arranges tonal materials into functional groupings:

octave/root, triadic, diatonic, and chromatic spaces. These functional groupings within the basic space

model formed the central part of our centre–periphery (local/field) spatial music mapping strategy.

[VViiddeeoo eexxaammppllee 33 Early spatial music performance tests 2’12” http://www.youtube.com

/watch?v=8YBLM2Ja6Uo]

Root/octave and triadic materials provide a grounding central dynamic, with diatonic and chromatic

materials activating the periphery. Such a spatial-relational mapping might be considered as a nascent

embodied cognition perspective on Lerdahl’s model, creating an embodied tonal space (depicted in the

second part of figure 4). This applies the embodied image schema theories of Lakoff and Johnson

(Johnson, 1987; Lakoff, 1987; Lakoff and Johnson, 1980) to various components of the basic space model.

FFiigguurree 44 From a cognitively based model of tonal pitch space, after Lerdahl (2001), to an embodied tonal

pitch space (highlighting verticality, cyclical, and centre–periphery or point-to-diffusion/dispersion

schemas).

Image schemas are theorised as common patterns of sensorimotor activity, “imported” into higher-level

cognitive functioning as some of the basic components of thought. More complex abstract models can be

conceptualised as combinations of these embodied image schemas. Theoretical work on describing

musical structures using these models has been carried out by Brower (2000; 2008) and Johnson (2007,

pp. 235–262). Solomon (2007, pp. 291–301) discusses these models specifically in the context of

spatialisation and spatial gestures in music; see also (Erickson, 1975, pp. 141–145) for some early

theorising about sound via spatial frameworks. Wilkie et al. (2010) provide evidence to support the

claim that musicians conceptualise musical structure in such a fashion. Some of the most musically

significant image schemas are depicted in figure 5.

FFiigguurree 55 Some common image schemas which have particular relevance for music, after Brower (2008,

p. 10)

Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/...

4 of 13 2/12/15, 10:47 AM

Page 5: Rgbb Divergence Issue3 2014

For our purposes, the advantage of an embodied perspective on Lerdahl’s tonal model is that it

highlights the relational dynamics of this formal model’s spatial structure (eg stable/unstable,

centre/periphery, point-to-diffused). This perspective on Lerdahl’s model therefore already embodies

ready-made mapping potential, as image schemas are inherently spatial-relational in origin, as they are

based on common movements in an environment. Furthermore, such an embodied model has potential

for extensibility whilst maintaining coherence. Other control modalities and intermediary control

mappings for spatial music may be easily incorporated into this system, with coherence contributed to

as long as they are structured around a compatible embodied/ecological base. This approach therefore

provides a framework for the integration of a variety of control approaches beyond our initial design.

In terms of the embodied components we identify within Lerdahl’s model, Brower’s theories are

particularly significant for the present work, drawing attention to verticality (grounded/stable to

air/unstable), cycle and centre–periphery schemas present within tonality (Brower, 2000, pp. 335–336;

2008, p. 15). The tonal hierarchy “cone” may be modelled as a combination of a verticality schema with

multiple cycle schemas comprising the different functional levels. The embodied coherence of these

components is further reinforced in our work by referencing centre-periphery schema (Brower, 2000, p.

318; Johnson, 1987; Lakoff, 1987) through spatialisation. One of the benefits of Emmerson’s (2007) space

frame typology is that it can be viewed as incorporating models of spatial containers (Brower, 2000, pp.

328, 336; Johnson, 1987; Lakoff, 1987; Solomon, pp. 291–296) – image schema terminology which relates

to framing and grouping. As applied in the present model (see figure 6), it provides us with tonal space

frames that conform to the previously presented local field dynamic.

FFiigguurree 66 Tonal space frames: movement of spatialised voices from local/localised to diffuse/field

positions due to functional positions in Lerdahl’s tonal hierarchy (eg triadic, diatonic, chromatic)

AAnniimmaatteedd ppiittcchh ssppaaccee mmaappppiinnggss

In addition to our embodied perspective on the (fixed) basic space model from Lerdahl, the next stage of

the system’s development saw the investigation of mappings that incorporated Lerdahl’s dynamic

models of tonal syntax behaviours (Lerdahl, 2001; Lerdahl and Krumhansl, 2007). We applied these

dynamic models of tonal attraction and repulsion to the animation of our tone space mappings,

incorporating the boids flocking algorithm (Reynolds, 1987; Singer, 1997) for the control of our

spatialised voices. Although this algorithm has previously been applied to spatialisation control (Bates,

2010), our innovation (Graham, 2012; Graham and Bridges, 2013; 2014) lies in its integration with

Lerdahl’s dynamic tonal models. Lerdahl provides a series of models influenced by “traditional” formal

cognitive models of tonal perception (Krumhansl, 1990). While these dynamic models are formalistic,

they arguably incorporate some clear embodied concepts. Lerdahl treats musical syntax dynamics as

analogous to forces, including a model for tonal attraction, which is based on gravitational attraction via

Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/...

5 of 13 2/12/15, 10:47 AM

Page 6: Rgbb Divergence Issue3 2014

an inverse-square law (Lerdahl, 2001; Lerdahl and Krumhansl, 2007). We adapted this embodied

perspective on spatial structures by mapping dynamic melodic syntax data (such as attraction and

inertia values) from Lerdahl’s tonal models to in-kind parameters (namely, attraction and inertia) of the

boids flocking algorithm (Graham, 2012; Graham and Bridges, 2013; 2014). This controlled dynamic

spatialisation effects relative to a specified central point within the ambisonic spatialisation field. In this

setup, each boid controls the movement of a single voice from the guitar’s multichannel audio output,

with strength of tonal attraction via the Lerdahl dynamic model reflected in the overall flock’s degree of

centricity and attraction; see figure 7 and video example 4.

[VViiddeeoo eexxaammppllee 44 Initial iteration of melodic model http://www.youtube.com/watch?v=r3W2G_QTsiw]

FFiigguurree 77 Boids mappings resolve tonal centre materials to spatial centre (providing “spatial closure” to

accompany tonal closure) in the presence of a tonic resolution

These embodied force structures can be seen as conforming to inertia-to-attraction force dynamics,

diffusion/dispersion-to-point image schemas, and spatialisation/diffusion based on local/field dynamics.

On a basic level, the flocking behaviours essentially treat melodic forces as physical or behavioural

forces. The mapping of melodic syntax to the flocking behaviour of the boids is outlined in figure 8.

FFiigguurree 88 Detail of melodic syntax model and boids mappings

MMeellooddiicc ssyynnttaaxx aanndd mmiisscc..

ccoommppuuttaattiioonnss ((nnoottee--

ttoo--nnoottee bbaasseess))

MMaappppeedd ttoo xx fflloocckkiinngg fflliigghhtt ppaatttteerrnn ““sstteeeerriinngg bbeehhaavviioouurrss””

Melodic attraction/tension Data was mapped to the centre parameter of each boid, reflecting the relationships

between stability and tension tone positions in a pitch class profile (default: Ionian

(major), conforming to Lerdahl’s basic space structure; other modes may still be

used with this profile, differences in pitch are articulated via centre–periphery

mappings).

Implicative denial Data was mapped to control the inertia of each boid, reflecting the denied

attractional potential as differences in behavioural flight patterns.

Ratios of asymmetrical

attraction

Data was mapped directly to the attract parameter, reflecting asymmetrical tonal

attractions relative to a sequential tone position in a pitch class profile. High values

will cause the boids to flock in attraction to a (tonic) point within a physical

performance space.

Tendency Data was mapped to control the speed matching behaviours of neighbouring boids.

High values produce matched speeds in flight patterns in an attempt to reflect

expectancy schemas through constancy (eg direction of melodic motion and tonal

attraction relationships).

Average note onset time

(ms)

Data was mapped to control the speed of the flight behaviours, establishing a

narrative between instrumental phrasing and the speed of the dynamic trajectory

adopted by the boids.

Pitch class distanceData was mapped to control the acceleration parameter of each boid, reifying the

notion that a listener may perceive smaller pitch class distances as occurring over a

shorter time period and larger pitch class distances over a longer time period; see

(Snyder, 2000, p. 12) for discussions of music and memory/timescales.

Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/...

6 of 13 2/12/15, 10:47 AM

Page 7: Rgbb Divergence Issue3 2014

an embodied spatialmodel and sharedgestural typology

Key cases are illustrated in figure 9; see also video example 5. Further details of our application of the

originating Lerdahl model can be found in Graham (2012, p. 130).

[VViiddeeoo eexxaammppllee 55 Initial iteration of melodic model, 13’16” http://www.youtube.com

/watch?v=r3W2G_QTsiw]

FFiigguurree 99 Dynamic boids ecological tonal-space for basic space (Ionian-Major) mappings; see also figure

4, for Lerdahl’s basic space model

Our investigation of Lerdahl’s tonal models has centred on the mapping of the resulting melodic syntax

values to in-kind embodied image schemas and force-based metaphors (boids). As a result, abstract tonal

structures are reified through dynamic spatialisation processes. In summary, the adapted models have

informed the design and integration of performance controls for an accessible and dynamic spatial

music performance system, maintaining coherence through a unifying embodied-relational framework.

Furthermore, our application of Lerdahl’s force-based models to ecological behaviours provides an

extension of the centre–periphery schemas via dynamic structures, which are compatible with Johnson’s

(2007, pp. 235–262) theory of embodied force metaphors in music cognition, specifically his moving

music and music as moving force metaphors.

GGeessttuurraall aaffffoorrddaanncceess aanndd eexxtteennddeedd eemmbbooddiieedd mmaappppiinnggss ffoorr rreeiinnffoorrcciinngg ssppaattiiaall

ssttrruuccttuurreess

AAnncciillllaarryy//aaccccoommppaanniimmeenntt ggeessttuurreess aanndd eexxtteennddeedd mmaappppiinnggss

The extension and refinement of this system has been based on a combination of practice-led and

theoretical developments. Our initial work provided an embodied spatial model and shared gestural

typology for developing connections between tonal and

physical/acoustic spaces. The result provided a model that could

facilitate the integration of more direct gesture tracking into this

system’s control structures. As such, connections between tonal

structures and embodied spatial domains could be further explored through the extraction of

larger-scale non-sounded performer bodily movements: ancillary/accompaniment gestures (Cadoz and

Wanderley, 2000). These movements, such as a change in central performer position (torso) or a

movement of the guitar’s body or neck, may be conscious or unconscious parts of the performer’s

creative practice. Such cases may be thought of as providing embodied accompaniments to resulting

musical structures.

For this iteration, physical gestural data was obtained using a combination of the Xbox Kinect sensor,

parsed via the Synapse application (Challinor, 2011), which provides values for velocity and acceleration

in addition to coordinate sets for skeletal points in a three-dimensional Cartesian space. Although these

are additional input modalities, they do not entail a laborious process of learning new instrumental

affordances and techniques (as might be the case with the provisional of additional “bolt-on” controls).

Rather, these types of gestures are already broadly accessible and familiar as by-products of established

performance practices. Furthermore, the control structures act either as moderators or reinforcements

of the established tonal-spatial control mappings.

[VViiddeeoo eexxaammppllee 66 Developing mapping strategies for spatial music performance: mapping skeletal data

using Xbox Kinect and Pure Data, 0’32” http://youtu.be/rA5ED9kEmo4]

Figure 10 shows this iteration of the system. In this version, the physical movement of the performer

Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/...

7 of 13 2/12/15, 10:47 AM

Page 8: Rgbb Divergence Issue3 2014

provides position and acceleration data for various points of the body and guitar, which can be

integrated with the rest of the system’s treatment of musical (specifically tonal) motion via force-based

and spatial analogues. For example, in addition to boids acceleration parameters being controlled by

computations of pitch class distance, direct bodily motion (as acceleration) can be applied to moderate

the tonal acceleration value.

FFiigguurree 1100 Integration of motion tracking for body movement with the system’s other physical motion

and force/musical motion and force mappings

EExxpplloorriinngg eemmbbooddiimmeenntt:: ddyynnaammiicc ssppaattiiaall mmaappppiinnggss ooff sskkeelleettaall ddaattaa

A number of fortuitous by-products of the joint-tracking process aided this wider exploration of

embodied controls. Firstly, the guitar’s neck was reliably treated as an extended limb (see figure 11).

Secondly, the nature of the tracking process implied that the guitarist’s picking hand would only be

tracked by the system when making larger accompaniment gestures rather than more typical

note-articulation gestures (picking/plucking, etc). As a result of these affordances of the technology, we

are able to access two distinct gestural-spatial ranges: (1) small-scale physical gestures for note

articulation and (2) more expansive bodily accompaniment/ancillary gestures. This combination

facilitates the treatment of these gestures both in terms of clear delineation of function (ie separate

functional mappings) whilst maintaining the overall holistic coherence of the integrated force–motion

metaphorical mappings.

FFiigguurree 1111 Skeletal tracking data from Synapse showing the treatment of the guitar neck as an extended

limb and superimposed with key control mappings

In video examples 7 and 8, the performer’s torso position sets the central attraction point for flocking

behaviours, allowing the performer to explore the notion of embodying an abstract musical concept

(tonal attraction/centricity).

[VViiddeeoo eexxaammppllee 77 Developing mapping strategies for spatial music performance: mapping skeletal data

Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/...

8 of 13 2/12/15, 10:47 AM

Page 9: Rgbb Divergence Issue3 2014

using Xbox Kinect and Pure Data http://youtu.be/rA5ED9kEmo4]

[VViiddeeoo eexxaammppllee 88 Mapping strategies for embodied metaphors – improvised music examples

http://youtu.be/6KgAULXmbzQ]

The velocity of the left “hand” (headstock of guitar) controls the flocking speed, directly linking the

motion of the guitar headstock to the rate of the flocking behaviours within the speaker array (see figure

11). Intuitively, the performer can easily calibrate axial positioning of the left hand (fretting hand) with

melodic tonal syntax when improvising, thus providing structurally informed spatial accompaniments

via left “hand” (neck and headstock) motion. The velocity of the right hand (picking hand) was mapped

to the following granular parameters: feedback, buffer position, grain reversal, and time variation. The

velocity of this hand also controlled the avoidance and acceleration flocking parameters, allowing for

the direct correlation between detached bodily movements outside of common instrumental practice to

be linked to more aggressive sonic transformations (see figure 11). As such, more expansive physical

gestures can be seen as spatial-performative correlates of more obvious and dynamic signal processing.

RReeiinnffoorrcciinngg ssppaaccee tthhrroouugghh eemmbbooddiieedd ssiiggnnaall pprroocceessss mmaappppiinnggss

In a similar fashion, some additional granular mappings are derived from our embodied version of the

Lerdahl basic tonal pitch space model (see figure 4). These mappings are designed to accentuate the

spatialisation effects previously noted. Conforming to the centre–periphery/point-source-to-diffuse

articulations, the effects of these mappings can be summarised as follows:

1. Stable positions within the tonal hierarchy/basic space produce centred granular images across

the horizontal plane, an application of high-pass filtering, and a decrease in grain gap size.

2. Less stable positions within the tonal hierarchy/basic space produce a wider granular image

across the horizontal plane, an application of low-pass filtering, and an increase in grain gap size.

The presence of more high frequency content in case (1) facilitates clearer spatial perception

(centre/point source/integration), while case (2) results in diffused/dispersed perspectives. See video

example 9.

[VViiddeeoo eexxaammppllee 99 Early spatial music performance tests, 2’11” http://www.youtube.com

/watch?v=8YBLM2Ja6Uo]

An additional example of an embodied granular mapping strategy with spatial implications can be

found in an additional note inter-onset time (rate) mapping from the note input. This mapping draws on

a metaphorical equivalence between this note rate and granular density and shorter grain durations.

Broadly speaking, an increase in note rate causes more obvious granular effects (via greater granular

density and shorter grain durations). In terms of audible spatial implications, the associated shorter

grain duration parametric mapping also implies an increased “presence” of the system’s higher

rate/tension response, as the shorter grains lead to increased noise and, hence, additional higher

frequency content (making it easier to localise materials). In metaphorical terms, increased inter-note

rate-effort events are mapped to greater activity density of the granulation process, leading to a greater

sense of presence and spatiotemporal coverage; see video example 10.

[VViiddeeoo eexxaammppllee 1100 Early spatial music performance tests, 5’30” http://www.youtube.com

/watch?v=8YBLM2Ja6Uo]

This particular mapping strategy reflects the theorised perspective provided by Johnson (2007, pp.

21–24) on the qualitative dimensions of movement; specifically, tension (embodying effort/amount of

activity, and connecting with rate for repeated short actions); see also (Graham and Bridges, 2013; 2014).

EEmmeerrggeenntt aaeesstthheettiiccss aanndd ffuuttuurree ddeevveellooppmmeenntt

MMoovviinngg mmuussiicc//mmoovviinngg ttiimmee mmaappppiinnggss

One fork of the performance system experimented with extended structural mappings influenced by

Johnson’s (2007, p. 248) moving music metaphor, whereby the listener’s metaphorical progression

through a piece of music is based on physical movements. Johnson’s connection of musical structures

with related spatial structures (eg paths of motion, cessation of motion, location of observer) is seen as

mapped to musical temporal structures and implications (respectively: gestural contours, points of

rest/stability/cadence and a sense of musical immediacy, ie “presence”/“the present”). This proposal of

cognitive connections between temporal structure and spatial structure suggests a provocative question

for designers of music performance systems: can spatialisation of materials be used to foreground

embodied/ecological metaphors underpinning resulting musical materials? Johnson’s equation of

“present” musical materials with the vantage point of a stationary observer suggests that spatial

movements relative to this observer may be used to highlight temporal structural relationships

embodied by musical materials. To examine this question, we must examine a number of motion-based

metaphors (figure 12):

Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/...

9 of 13 2/12/15, 10:47 AM

Page 10: Rgbb Divergence Issue3 2014

FFiigguurree 1122 Musical motion metaphors, adapted from Johnson (2007, p. 248)

Location of observer > Present musical event

Objects in front of observer > Future musical events

Objects behind observer > Past musical events

Path of motion > Musical passage

Temporary cessation of motion > Musical cadences/resolution

This model sees the tracking of new note articulations in order to spatialise each object from the front of

the array (future) via centre (present) to the rear of the speaker array (past), modelling a temporal

progression as the materials move past the listener. As notes are held, they are dynamically spatialised

towards the rear of the array and are diffused through the application of filtering and delay effects.

When an object reaches the rear of array, it remains (contained) in a static position until amplitude

levels fall below a predetermined threshold, allowing the event to decay. A centre–periphery tonal

schema is maintained through the use of a modification of the Lerdahl basic space model to provide

distance factors for the relative lateral (x-plane) spatialisation. The direction of spatialisation on the

horizontal axis is based on alternation of basic spaces around the centre point, thus being based on a

relational rather than absolute model. It was considered that the previous cyclical schema did not enjoy

a particular degree of cognitive salience and that a simplification on this basis would facilitate a more

dynamic front-to-rear relational mapping strategy. This version of the system is depicted in figure 13; see

also video examples 11 and 12.

[VViiddeeoo eexxaammppllee 1111 Sounds and schemas: mapping metaphors http://youtu.be/2KgYKpZYsWs]

[VViiddeeoo eexxaammppllee 1122 Sounds and schemas: moving music metaphor – improvised music examples

http://youtu.be/P-BBI_XRKCE]

One particular example of the interplay between horizontal and front-back perspectives in this mapping

can be found at 0’45” of video example 12. This illustrates how different positions within the tonal

hierarchy contribute to dynamic connections within and between individual instrumental legato

articulations. In this context, the re-articulation of a previously held note may lend a sense of cyclical

structural return. Ostinato structures become points of relative spatial stability.

FFiigguurree 1133 Spatial application of the moving music metaphor (Johnson, 2007)

As per the previous mapping strategies, more stable tonal materials, such as those in an octave/root

space, are placed at the centre of the field frame. Less stable tonal materials, such as the chromatic

space, are placed closer to the perimeter. Interpolation between axial points produced a favourable

aesthetic-structural response, particularly during glissando string events. Rate-effort mappings (note

inter-onset time and right hand velocity) were implemented to control the prominence (or feedback) of

granular signal processes, providing an extra degree of performative gestural control for effects with

spatial implications. These sets of mapping strategies provide the performer with greater control over

the framing of the system’s tonal-spatial materials.

CCeennttrree––ppeerriipphheerryy sscchheemmaass aanndd ppootteennttiiaallss

The current developments of the system imply certain emergent aesthetic preferences and priorities

that are common to each of the iterations and variants of the performance system. Space is treated as a

framing parameter, an aspect of parametric organisation that supports and facilitates the perception of

Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/...

10 of 13 2/12/15, 10:47 AM

Page 11: Rgbb Divergence Issue3 2014

other musical structuring principles. All versions of the performance system incorporate centre–

periphery relational structures (with accompanying diffuse/enveloped-to-point-source articulations of

sound materials). The use of a hierarchical tonal model as a control structure for distance parameters

provides an embodied modelling of competing musical currents. Furthermore, as dissonant materials

are diffused towards the periphery of the array, apparent sensory dissonance effects (ie beating,

roughness effects) may be reduced due to enhanced spatially based auditory stream segregation

(Bregman, 1990, pp. 293–302); see also (Graham, 2010), which discusses the application of this principle

in electric guitar performance via the use of multichannel audio feeds. Hence, in the present system’s

treatment, dissonant tonal materials may be viewed as spatially contained via the framing of our

centre–periphery processes.

In addition, the performer’s ability to change the centre location based on larger-scale bodily

movements (relative torso position) maintains this relational clarity whilst providing an opportunity for

the “central” spatial responses of the system to be dynamically controlled. However, this aspect of the

system’s control may require judicious engagement on the part of the performer: sudden larger-scale

bodily movements (rather than more progressive changes in the form of gestural accompaniments) may

provide more of a jarring of perspective than a useful expansion of spatial development. With that said,

a performer who is particularly aware of their relative positions within a performance space (and the

relevant application tracking ranges) might derive some benefit from the exploration of this modality.

CCoonncclluussiioonn

We have outlined an approach to the creation of spatial music via systems design and

hyperinstrumental performance practice. We consider this design approach to be broadly consistent

with perceptually informed spatial music practices and with the theories of Emmerson (2007), whereby

the final system iteration is used to contribute to the perceptual delineation of emergent musical

materials. Furthermore, our performance system’s application of a parsed audio feed to the control of

spatialisation via a tonal model from Lerdahl (2001) can be seen as grounding more abstract and formal

tonal concepts within an embodied performance environment. We believe that this approach provides

an intuitive means of connecting tonal structures to centre–periphery relational dynamics in a spatial

music performance practice. Further investigations led to an exploration into embodied/ecological

potentials within Lerdahl’s dynamic model of tonal syntaxes (Lerdahl, 2001; Lerdahl and Krumhansl,

2007), resulting in the treatment of Lerdahl’s musical forces and movement dynamics as control

structures for animated spatial mappings using the boids flocking algorithm (Bates, 2010; Reynolds,

1987). These animated mappings are used to apply melodic behaviours and related motion dynamics to

in-kind parameters in the boids algorithm, which provides a more sophisticated environmentally based

metaphorical mapping for the control of individual spatialised voices in real-time music performance.

The advantage of using embodied models as a means of contributing to rich yet coherent performance

system responses is highlighted further through the integration of a motion tracking component to the

system. The specification of additional controls and mappings via a unifying embodied framework

allows for additional parameters to be added whilst maintaining maximal consistency and coherence,

and hence accessibility for the performer. We consider these ecological/embodied models to be ripe for

wider application in spatial music systems design and creative practices due to these integrating

potentials. A variety of extensions to the signal processing side also have mappings that can be seen as

consistent with embodied/ecological relations. For example, various types of rate mappings are widely

applied to the density and dynamism of granular processing effects. Furthermore, some of the system’s

audio effects chains are designed to accentuate spatial centre–periphery dynamics, impacting the

localisation abilities of the listener. Finally, a more speculative mapping based on Johnson’s moving

music and moving time metaphors was discussed, approaching spatialisation from the perspective of

highlighting temporal progression as a set of tonal materials that flow past the listener on a front–back

axis, relative to the envelope profile of each instrumental note event. One of the greatest strengths of the

resulting system lies in its ability to integrate a wide variety of controls and creative outputs within a

coherent but extensible framework. Overall, these types of spatialisation processes place the performer

at the centre of an embodied creative space, establishing unique narratives between the image schemas

underpinning instrumental theory and physical technique. As such, we hope that the theoretical

principles underpinning some of our designs will suggest creative mapping possibilities to other

practitioners.

NNootteess:

RReeffeerreenncceess:

Bates, E. (2010) The Composition and Performance of Spatial Music. PhD,

University of Dublin, Trinity College, Dublin.

Bregman, A.S. (1990) Auditory Scene Analysis: The Perceptual Organization of Sound. Cambridge, Mass.:

MIT Press.

Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/...

11 of 13 2/12/15, 10:47 AM

Page 12: Rgbb Divergence Issue3 2014

Brower, C. (2000) A Cognitive Theory of Musical Meaning. Journal of Music Theory, 44(2), pp. 323–379.

DOI: 10.1215/00222909-44-2-323

Brower, C. (2008) Paradoxes of Pitch Space. Music Analysis, 27, pp. 51–106.

DOI: 10.1111/j.1468-2249.2008.00268.x

Cadoz, C. and Wanderley, M. (2000) Gesture-Music. In Wanderley, M. and Battier, M. (eds). Trends in

Gestural Control of Music. Paris: IRCAM. pp. 72–94. Available at: http://www.music.mcgill.ca

/~mwanderley/Trends/Trends_in_Gestural_Control_of_M usic/ [Accessed 30/8/2014]

Challinor, R. (2011) Synapse. Software application. Available at: http://synapsekinect.tumblr.com/

[Accessed 6/7/2014]

Emmerson, S. (1994) Local/Field: towards a typology of live electroacoustic music. Proceedings of the

International Computer Music Conference 1994. San Francisco: International Computer Music

Association. pp. 31–34.

Emmerson, S. (2007) Living Electronic Music. Aldershot: Ashgate.

Erickson, R. (1975) Sound Structure in Music. Berkeley: University of California Press.

Graham, R. (2010) The Effects of Polyphonic Technology on Electric Guitar Performance. Postgraduate

Conference of the Society for Musicology in Ireland, Dublin Institute of Technology, January 2010.

Available online at: https://www.academia.edu/1140419

/The_Effects_of_Polyphonic_Technology_on_Electric_Guitar_Performance [Accessed 30/8/2014]

Graham, R. (2012) Expansion of Electronic Guitar Performance Practice through the

Application and Development of Interactive Digital Music Systems. PhD.

University of Ulster, Northern Ireland.

Graham, R. and Bridges, B. (2013) Mapping and Meaning: embodied metaphors and non-localised

structures in performance system design. In: Re-New 2013 Conference Proceedings. Available online at:

http://www.l--l.dk/downloadables/re-new_2013_conference_proceeding.pdf [Accessed 30/8/2014]

Graham, R. and Bridges B. (2014) Gesture and Embodied Metaphor in Spatial Music Performance

Systems Design. In: Caramiaux, B., Tahiroğlu, K., Fiebrink, R. and Tanaka, A. (eds) Proceedings of the

International Conference on New Interfaces for Musical Expression 2014, Goldsmiths, University of

London, July 2014, pp. 581–584. Available online at: http://nime2014.org/proceedings/NIME2014-

Proceedings.pdf [Accessed 14/7/2014]

Johnson, M. (1987) The Body in the Mind: The Bodily Basis of Meaning, Imagination and Reason.

Chicago: University of Chicago Press.

Johnson, M. (2007) The Meaning of the Body: Aesthetics of Human Understanding. Chicago: University of

Chicago Press.

Krumhansl, C. (1990) Cognitive Foundations of Musical Pitch. Oxford: Oxford University Press.

Lakoff, G. (1987) Women, Fire, and Dangerous Things: What Categories Reveal

About the Mind. Chicago: University of Chicago Press.

Lakoff, G. and Johnson, M. (1980) Metaphors we Live By. Chicago: University of Chicago Press.

Lerdahl, F. (2001) Tonal Pitch-Space. Oxford: Oxford University Press.

Lerdahl, F., Krumhansl, C. (2007) Modeling Tonal Tension. Music Perception, 24(4), pp. 329–366.

DOI: 10.1525/MP.2007.24.4.329

Levitin, D. J. (2002) Control parameters for musical instruments: a foundation for new mappings of

gesture to sound. Organised Sound, 7(2), pp. 171–189. DOI: 10.1017/S135577180200208X

Moritz, A. (2002) Stockhausen: Essays on the Works. [Online] Available at: http://home.earthlink.net

/~almoritz/ [Accessed 14/7/2014]

Reynolds, C. (1987) Flocks, Herds and Schools: A distributed behavioral model. SIGGRAPH Comput.

Graph. 21(4), pp. 25–34.

Singer, E. (1997) Boids for Max. Software library. Available at: http://ericsinger.com/cyclopsmax.html

[Accessed 6/7/2014]

Smalley, J. (2000) Gesang der Jünglinge: History and Analysis. In: Masterpieces of 20th-Century

Electronic Music: A Multimedia Perspective [Online]. New York: Columbia University Computer Music

Center.

Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/...

12 of 13 2/12/15, 10:47 AM

Page 13: Rgbb Divergence Issue3 2014

Available online at: www.music.columbia.edu/masterpieces/notes/stockhausen [Accessed 14/7/2014]

Smalley, D. (2007) Space Form and the Acoustic Image. Organised Sound, 12(1), pp. 35–58.

DOI: 10.1017/S1355771807001665

Solomon, J. (2007) Spatialization in Music: The analysis and interpretation of Spatial Gestures. PhD.

University of Georgia, Athens, Georgia.

Sterne, J. (2012) Sonic Imaginations. In: Sterne, J. (ed.) The Sound Studies Reader. Abingdon, Oxford:

Routledge, pp. 1–18.

Stockhausen, K. (1956) Gesang der Jünglinge. Electroacoustic composition. Available on: Stockhausen

Edition 3. [Audio CD]. Kürten, Germany: Stockhausen Foundation for Music.

Stockhausen, K. (1957) Gruppen. [Composition]. Available on Stockhausen Edition 5. [Audio CD]. Kürten,

Germany: Stockhausen Foundation for Music.

Stockhausen, K. (1975) Music in Space. (Trans. R. Koenig). die Reihe, 5, pp. 67–82.

Wilkie, K., Holland, S., Mulholland, P. (2010) What Can the Language of Musicians Tell Us about Music

Interaction Design? Computer Music Journal, 34(40), pp. 34–48. DOI: 10.1162/COMJ_a_00024

AAuutthhoorr: Ricky Graham, Brian Bridges

AAbboouutt TThhee AAuutthhoorr:

Richard Graham is a guitarist and computer musician based in the United States. Graham has performed

across the US, Asia, UK, and Europe, including festivals and conferences such as Celtronic and the

International Symposium on Electronic Art. He has composed music for British and US television,

recorded live sessions for BBC radio, and has authored music for the popular video game, Rock Band.

Ricky was an artist-in-residence at STEIM in 2010, where he developed the first iteration of his

performance system for multichannel guitar. He received his Ph.D. in Music Technology from the

University of Ulster in 2012 and he is now an Assistant Professor of Music and Technology at Stevens

Institute of Technology in New Jersey. His most recent paper on performance systems was presented at

NIME 2014 and his most recent musical work, “Nascent”, was released on Fluttery Records in 2012.

Brian Bridges is a composer and music technology researcher from Dublin, Ireland. He is currently

based at the University of Ulster, Northern Ireland, where he has been Lecturer in Creative Arts/Creative

Technology since 2008. His research interests lie in the connection between theories of auditory

perception and cognition and creative practices and systems designs. His creative work spans the fields

of sound installation and audiovisual practices and electroacoustic and instrumental music. He is a

founder–member of the Dublin–based Spatial Music Collective and his compositions have been

programmed at festivals in Europe, the Americas and Asia. Brian is a graduate of Trinity College Dublin

(MPhil. in Music and Media Technologies) and the National University of Ireland, Maynooth (PhD on

microtonal music) and he has also undertaken private studies in the US with Glenn Branca and Tony

Conrad. www.brianbridges.net

Posted in: Issue 3, December 2014

Article view count: 284

Actions: E-mail | Permalink |

Divergence Press is an online, open access, peer-reviewed journal and resource hub forcontemporary music and is part of University of Huddersfield Press. A project of the Centre forResearch in New Music, University of Huddersfield. ISSN 2052-3467

Terms Of Use Privacy Statement Acknowledgements Copyright 2015 by University of Huddersfield Login

Strategies for spatial music performance: the practicalities and... http://divergencepress.com/Journal/JournalIssue/tabid/85/ID/...

13 of 13 2/12/15, 10:47 AM