Show simple item record

dc.contributor.advisorJoseph A. Paradiso.en_US
dc.contributor.authorFeldmeier, Mark Christopher, 1974-en_US
dc.contributor.otherMassachusetts Institute of Technology. Dept. of Architecture. Program In Media Arts and Sciencesen_US
dc.date.accessioned2006-07-31T15:11:33Z
dc.date.available2006-07-31T15:11:33Z
dc.date.copyright2003en_US
dc.date.issued2003en_US
dc.identifier.urihttp://hdl.handle.net/1721.1/33547
dc.descriptionThesis (S.M.)--Massachusetts Institute of Technology, School of Architecture and Planning, Program in Media Arts and Sciences, 2003.en_US
dc.descriptionIncludes bibliographical references (p. 113-114).en_US
dc.description.abstractOne of the difficulties in interactive music and entertainment is creating environments that reflect and react to the collective activity of groups with tens, hundreds, or even thousands of participants. Generating content on this scale involves many challenges. For example, how is the individual granted low latency control and a sense of causality, while still allowing for information retrieval from all participants so that the environment responds to the behavior of the entire group? These issues are particularly pertinent in the area of interactive dance. To address these issues, a low-cost, wireless motion sensor has been developed. The sensor is inexpensive enough to be considered disposable, allowing it to be given away to participants at large dance events, enabling the dancers to participate concurrently in a realtime, interactive musical performance. The sensors are either worn or held by participants and transmit a short RF pulse when accelerated past a certain threshold. The RF pulses are received by a base station and analyzed to detect rhythmic features and estimate the general activity level of the group. These data are then used to generate music that can either lead or follow the participants' actions, thereby tightening the feedback loop between music and dancer. Multiple tests of the system have been conducted, with groups ranging from fifteen to 200 participants. Results of these tests show the viability of the sensors as a large group interaction tool. Participants found the interface intuitive to use, effectively controlling such aspects of the music as style, tempo, voicing, and filter parameters. These tests also demonstrate the system's ability to detect both the activity level and dominant tempo of the participants' motions, and give considerable insight into methods of mapping these data to musical parameters that give participants direct feedback as to their current state. Furthermore, it is shown that participants, if given this direct feedback, will synchronize their actions and increase in activity level, creating a mutually coherent and pleasing outcome.en_US
dc.description.statementofresponsibilityby Mark Christopher Feldmeier.en_US
dc.format.extent114 p.en_US
dc.format.extent5413773 bytes
dc.format.extent5413583 bytes
dc.format.mimetypeapplication/pdf
dc.format.mimetypeapplication/pdf
dc.language.isoengen_US
dc.publisherMassachusetts Institute of Technologyen_US
dc.rightsM.I.T. theses are protected by copyright. They may be viewed from this source for any purpose, but reproduction or distribution in any format is prohibited without written permission. See provided URL for inquiries about permission.en_US
dc.rights.urihttp://dspace.mit.edu/handle/1721.1/7582
dc.subjectArchitecture. Program In Media Arts and Sciencesen_US
dc.titleLarge group musical interaction using disposable wireless motion sensorsen_US
dc.typeThesisen_US
dc.description.degreeS.M.en_US
dc.contributor.departmentProgram in Media Arts and Sciences (Massachusetts Institute of Technology)
dc.identifier.oclc52641276en_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record