This thesis presents the design, implementation and findings of a Virtual Reality Musical Instrument (VRMI). The project was done under the direction of the Sound and Physical Interaction (SOPI) research group. The project was made following an iterative design methodology and the metaphors and design patterns used in Ubiquitous Music Systems.
In contrast with the fast adoption of Virtual Reality as a platform for new entertainment productions, it is noticeable that the area of new interfaces for musical expression (NIME) has been disbelieving towards this technology. At the same time, previous projects under the category of VRMI have made a clear distinction between the instrument, an external 3D model, and the user. Thereby, this thesis presents a project that focuses on how VR can enhance individual musical interaction? In order to do so, this project is directed to blurry the lines between performer, instrument and environment by creating immersion through 3D audio, audiovisual feedback, bodily and spatial interaction, the performer and the system's autonomous responses. As a final result, this thesis reaches to provide the NIME community with a purposeful use of Virtual Reality as an interactive musical platform.