HOW WOULD THE WORLD BE LIKE IF ONE COULD SEE THROUGH THE EYES OF ANOTHER? WOULD IT HELP US TO UNDERSTAND EACH OTHER? WOULD IT HELP US TO UNDERSTAND OURSELVES?
THE MACHINE TO BE ANOTHER is an Open Source Art investigation on the relation of Identity and Empathy that has been developed on a basis of low budget experiments of Embodiment and Virtual Body Extension.
Designed as an interactive performance installation, the ‘Machine’ offers users the possibility of interacting with a piece of another person’s life story by seeing themselves in the body of this person and listening to his/her thoughts inside their mind.
The performer is someone interested in sharing a story about his/her existence. This role can be assumed by an actor interpreting a real situation, or rather it may be taken by any person (e.g. from the public) who is interested in sharing some episode about his or her life. In any case, the stories that are told by the performer will be experienced by another person, say the user.
The project is currently developed by BeAnotherLab and is looking to understand empathy, identity, body agency, and subjectiveness from an interdisciplinary perspective.
Our approach consists in designing interactive experiments in collaboration with ‘real people’ (volunteers, such as artists, activists, and the audiences of the cultural centers). With a defined individual or group, we conduct informal interviews and debates in order to create performances addressing issues like peaceful co-living, prejudice of other kinds of ideological conflicts. The experiments are followed by discussions about each one’s own perception of identity after experiencing to see the world through the eyes of another one (in case of the users) and sharing a life story (in case of the performers).
Our main interest through this approach is to use the ‘Machine’ as a tool to help promoting empathy among individuals of different social, cultural and ideological contexts. However, we are also open to new points of views and interests that might be offered by the collaborators.
HOW DOES IT WORK?
This interactive performance-installation is built in 2 identical spaces: one for the user and another for the performer.
Through immersive goggles (head mounted displays), the user sees a video with the eyes’ perspective of a different person (the performer), who follows the former’s movements. So, it all happens in a way in which the user kind of ‘controls’ the performer‘s movements.
The user can move and interact with objects inside a room, while listening to the performer‘s thoughts through a set of headphones. That generates the perception of someone speaking inside their mind.
The video transmitted to the user’s goggles is generated by a camera attached to the perfomer, that records his/her point of view in real time. The performer also wears a microphone used to tell a short story about his/her existence. This non lineal narrative is connected to the objects in the room (a photo of someone, a childhood toy, a pack of cigarettes, a mirror, etc), making them interactive: when the user interact with one object, the performer starts to speak about his experience and memories related to this object.
In general terms, the system merges technology with other variables: the performance, an interactive narrative (related to objects disposed at the same position in two identical spaces), the experiment’s assistants (with whom they can explore the touch sense), as well as sensorial/motor/physical stimuli disposed in the space, with which the user can interact. As for the objects, we always make use of mirrors and other different things that the user can either throw or feel (like a glass of water or a bunch of flowers).
The technological part consists of low budget material arranged in a way in which everything works perfectly. We use camera mapping of the user’s space, and digital compasses to track the user‘s head and body direction. As for the performer, we use a subjective camera controlled by servomotors, which is synchronized with the head position. Besides HMDs, Arduinos and RF transmitters to make it as wireless as possible.
:: 2 head mounted displays (our latest version runs with Oculus Rift)
:: wireless Earphone for the user
:: 1 wide angle webcam (for the performer). Note that we are developing with stereoscopy.
:: 1 computer.
:: Programed in: OpenFrameworks, Pure Data, and OSC mobile apps (TouchOSC and Control).
?badge=0″" width="”500″" height="”366″" frameborder="”0″" 8="webkitAllowFullScreen" 9="mozallowfullscreen" 10="allowFullScreen" scrolling="no" class="iframe-class">