back printable version

The Use of Bio-interfaces in Interactive Multimedia Works: Two Examples


This research is about the relationship between the human body and digital devices, particularly the interaction between real-time media (audio and visual) and the bio-data from performers or from an audience. This research belongs to the field of biofeedback, based on research pursued by the psychologist Neal E. Miller in the 60s.

As a result of this research, the paper introduces two interactive performances of my authorship using bio-interfaces that have been produced in the last six years. The results herewith are used for artistic purposes and not to reduce stress, as it is the case in biofeedback methods such as Miller’s.


Live electronics, Butoh, performance, biofeedback, interactive sound and video.

1. Biofeedback

This method was developed during the 60s by the psychologist Neal E. Miller, which consists of the process of measuring physiological parameters from a subject (for example, the heartbeat, the brainwaves or the breathing), sending this data to a computer to be analysed and afterwards translating these parameters to sound and video, feeding them back to the subject. This treatment technique was originally created to improve health, helping tense and anxious people to learn how to alter these functions by relaxing.

Figure 1. Visual representation of biofeedback1

The two bio-interfaces introduced in this paper are the EMG (electromyogram) interface used for the Butoh Performance-Installation: Seed/Tree and the EEG (electroencephalogram) interface used for the performance INsideOUT. Both performances were programmed in real time media with the software MAX/MSP-Jitter.

2. SEED /TREE (2005) Installation/Butoh Performance/Live Electronics

This project was created during an “artist in residence” program at the ZKM (Centre for Media Art in Karlsruhe, Germany).

Butoh is a modern expressive dance-form created in Japan in the 60’s by Tatsumi Hijikata and Kazuo Ohno. It is traditionally performed in white-body make-up and the movements are very slow and expressive. The movements in this dance form come from the inner world, they must emerge from within and not be imposed from without; Butoh is not a representative dance. During a performance, the Butoh dancer is in a state of ‘hyper-presence’, he is aware of everything around him and within his own body.

This installation-performance consists on a forest environment created by some panels projected by haptic images from tree cortex and human skin. By haptic images I mean: HAPTIC from (the Greek word: HAPT?S - tactile): the sensation “to feel”: the feeling that one can ‘touch’ with the eyes. There are three Butoh dancers in the space performing the process of a seed growing to become a tree. Each performer moves in his/her own way representing the same subject a-synchronically. In the performance, the inner impulse of every dancer emerges on the stage; the dancers have each their own movements, which are the product of their own imagination.

The installation runs for three hours. The dancers develop the main subject in twenty minutes, then they lie for ten minutes on the floor and afterwards they repeat the process from the beginning. During this time, the performers have the necessary time to experience their own imaginary world combined with the outer space created by the sound and video projection. There is a continuous feedback between the dancers and the media; the translation of emotional physiological parameters to sound and video, however, gives feedback not only to the dancers but also to the spectators.

Feelings, associations, mental images and spontaneous impulses are the starting point for the creation of stories and choreography. In Seed/Tree the dancers produce and transform the sound and the audience control the video projections. The results are instantaneous creations, expressions of the moment, with image, movement and music forming living signs in space.

There are two types of interactivity in this performance. The first one is the interaction between dance and sound: the performers have microphones and EMG electrodes attached to their bodies. The breathing and the heartbeat of two of the performers produce sounds that are continuously modified by the muscular tension of a third dancer. The second type of interactivity is that between the installation space and the visitors. During the performance, visitors can walk freely around the virtual forest. There is a video observing the installation space and human presence influences the video projections; these interactions create subtle differences of the video on the panels. The installation space itself becomes aware of the visitors and reacts according to their movements. In this way the visitor is invited to be part of the environment.

Figure 2. Seed/Tree at the ZKM Centre Karlsruhe (Germany)©, 2004

For Seed/Tree I used a wireless EMG (electromyogram) interface developed by Frieder Weiss, as shown in Fig. 3.

Figure 3. EMG Interface by Frieder Weiss

This interface has three pairs of electrodes, which are attached to three different muscles. This interface measures the muscle tension, while the program sends the values as a continuous OSC packet that is received through an OSC-route object in the MAX program; then, each value is used to trigger different sound effects in MAX/MSP.

Figure 4. Seed/Tree: EMG electrode attached to the Butoh dancer.

Figure 5. The Frieder Weiss program sending the values as a continuous OSC packet

Figure 6. Seed/Tree: the MAX/MSP patch by Claudia Robles.

A video excerpt of Seed/Tree can be downloaded at:

3. INsideOUT (2009) Performance/Live Electronics

This project was created during an “artist in residence” program at the KHM (Academy of Media Arts in Cologne, Germany). It is about the materialization of the performer’s thoughts and feelings on the stage. ‘The stage is a place for the appearance of the invisible. Yasu Ohashi says: the actors aim at our senses, our body and our unconscious and not at our intellect. Their gestures try to envision THE INVISIBLE WORLD’. (Haerdter and Kawai, 1998: 25)

The performer interacts with the sound and images using an EEG (electroencephalogram) interface, which measures the performer’s brain activity. Those sounds and images – some already stored in the computer and some produced live- are continuously modified by the values from two electrode combinations via MAX/MSP-Jitter. Hence, the performer determines how those combinations will be revealed to the audience. Images are projected to a screen and also onto the performer, while sounds are projected in surround.

Figure 7. INsideOUT:  MAX/MSP-Jitter patches by Claudia Robles

The German psychiatric doctor Hans Berger was the first to produce electroencephalograms of human subjects and also the first to observe the characteristic alpha-rhythm. The EEG project was meant at the beginning as the expression of the self, about turning the subject’s imagination from the inside to the outside.

For this performance, I used an open source EEG interface from Olimex, which measures the brain activity and consists of two assembled boards: one analogue and the other digital.

Figure 8. The analogue and digital boards from Olimex.2

Each board has two electrode combinations (or two EEG channels). It is possible to connect up to three boards (i.e. six EEG channels). I use only two for this piece: the frontal and the occipital channels. The rubber cap and the contact electrodes of the interface are those typically used in medical applications.

Figure 9. INsideOUT at the University of Miami’s CAS Art Gallery (USA) ©, 2009
(Photo by Javier A. Garavaglia)

To adapt this interface I received technical support by Lasse Scherffig and Martin Nawrath at Lab3-KHM. They modified the open EEG device by replacing the Atmel microcontroller with one running the Arduino firmware and changing the quartz clock accordingly to 16 MHz.

Lasse Scherffig wrote a program with the software Processing, which reads the values of both channels from the open EEG via a serial communication. The modified open EEG sends ASCII-formatted data representing the voltages of both channels (at a frequency of 100 Hz).  In Processing, a Fast Fourier Transform (FFT) is applied to that data and extracts the bins for the frequencies 0-50 Hz. From these, the median of the frequencies for the alpha channel (8-13 Hz) is extracted, smoothed using a low-pass filter and transmitted via OSC, which is received once again by the OSC-route object in the MAX program. This frequency band between 8-13 Hz correspond to the Alpha frequency range that is accentuated during relaxation. A small part of the Beta range is taken, which corresponds to an alert or an anxiety state.

For the performance of INsideOUT I have tried to train my brain in order to control the media combinations on the stage, putting in evidence different emotional and mental states. However, this conscious control is not completely attained due to the enormous and uncontrollable stream of feelings that generally appear surprisingly under such circumstances. My aim herewith is to be able to control and to be aware of the body and its functions and at the same time, to allow for an empty space that could be populated by the invisible or the imperceptible.

Figure 10. INsideOUT at the SIGGRAPHAsia2009 Yokohama (Japan) ©, 2009
(Photo by M.Goldowski)

A video excerpt of INsideOUT can be downloaded at:

This research and its artistic results aim to raise awareness of the human body and its functions (e.g. muscle tension, breathing, etc) as a means to manipulating the media by controlling those functions consciously on the stage. This shall create an environment that invites the visitor to perceive the body in a different way and to reflect on the relation human/machine.


  1. Taken from Wikipedia:
  2. Figure from:


  1. Artaud, Antonin (1938). ‘Le Théâtre et son double’. Gallimard, Paris.
  2. Deleuze and Guattari (1973). ‘L’anti – Oedipe. Capitalisme et Schizophrénie’. Minuit, Paris.
  3. Deleuze and Guattari (1980). ‘Capitalisme et Schizophrénie 2. Mille Plateaux’. Minuit, Paris.
  4. Diebner, Druckery and Weibel Peter (2001). ‘Sciences of the interface’. Tuebingen, Genista.
  5. Ferreira-Lopes, Paulo (2004). ‘Study of real time interfaces and interaction models: applications in music composition’. PhD Thesis. University of Paris VIII.
  6. Haerdter, Michael and Kawai, Sumie (1998). ‘Rebellion des Körpers, Butoh, ein Tanz aus Japan’. Alexander Verlag, Berlin.
  7. Rohr, Anette (2003). ‘Design eines flexiblen, computerunterstutzten. Biofeedback-Systems’. Thesis, Medien und Informationswesen. Fachhochschule Offenburg.

Claudia Robles was born in Bogotá (Colombia). Currently, she lives in Cologne (Germany).  She finished studies in Fine Arts in 1990 at the University J. T. Lozano in Bogotá (Colombia).  She pursued postgraduate studies such as: Film Animation (1992-1993) at the CFP (Milan-Italy); MA in Visual Arts (1993-1995) at the École Supérieure d'Art Visuel (Geneva-Switzerland) and Sound Design and Electronic Composition at the Folkwang University Essen (Germany) from 2001 to 2003. She was artist in residence (2004-2006) at the ZKM Centre for Media Art in Karlsruhe (Germany). Her most relevant work presented there was the piece Seed/Tree (audiovisual Installation/Butoh performance with live-electronics). Her work is known worldwide, having participated in several group and solo exhibitions around the globe, for example at the Bauhaus-archiv Museum fuer Gestaltung in Berlin, Germany (2003); the European Capitals of Culture: Sibiu and Luxemburg (2007); Enter3 in Prague, Czech Republic (2007); the International Computer Music Conference ICMC in Copenhagen, Denmark (2007) and Montréal, Canada (2009); at the SIGGRAPH Asia 2009 in Yokohama, Japan (2009) and lately at the DRHA 2010, Brunel University in London, UK.