Here, we describe virtual-hand and virtual-face illusion paradigms that can be used to study body-related self-perception/-representation. They have already been used in various studies to demonstrate that, under specific conditions, a virtual hand or face can be incorporated into one's body representation, suggesting that body representations are rather flexible.
Studies investigating how people represent themselves and their own body often use variants of "ownership illusions", such as the traditional rubber-hand illusion or the more recently discovered enfacement illusion. However, these examples require rather artificial experimental setups, in which the artificial effector needs to be stroked in synchrony with the participants' real hand or face—a situation in which participants have no control over the stroking or the movements of their real or artificial effector. Here, we describe a technique to establish ownership illusions in a setup that is more realistic, more intuitive, and of presumably higher ecological validity. It allows creating the virtual-hand illusion by having participants control the movements of a virtual hand presented on a screen or in virtual space in front of them. If the virtual hand moves in synchrony with the participants' own real hand, they tend to perceive the virtual hand as part of their own body. The technique also creates the virtual-face illusion by having participants control the movements of a virtual face in front of them, again with the effect that they tend to perceive the face as their own if it moves in synchrony with their real face. Studying the circumstances that illusions of this sort can be created, increased, or reduced provides important information about how people create and maintain representations of themselves.
According to Western philosophy, the human self consists of two aspects1: For one, we perceive our own body and our activities in the here and now, which creates a phenomenal self-representation (often called the minimal self). For another, we create more enduring representations of ourselves by storing information about our personal history, integrating new information into the emerging self-concept, and present ourselves to our social environment accordingly, which amounts to the creation of a so-called narrative self. The minimal or phenomenal self has been argued to emerge from two sources of information. One is top-down information about longer-lasting aspects of our body, such as information about the effectors we own or the shape of our face. The other is bottom-up information provided by self-perception in the current situation.
Investigations of the latter were strongly inspired by a clever study of Botvinick and Cohen2. These authors presented human participants with a rubber hand lying in front of them, close to one of their real hands, which however was hidden from view. When the real hand and the rubber hand were stroked in synchrony, so to create intermodal synchronous input, participants tended to perceive the rubber hand as part of their own body-the rubber-hand illusion. Further studies revealed that perceived ownership even went so far that participants would start sweating and trying to withdraw their real hand when the rubber hand was being attacked by a knife or otherwise being "hurt"3.
While Botvinick and Cohen have interpreted their findings to demonstrate that self-perception arises from the processing of bottom-up information, other authors have argued that the rubber-hand illusion results from the interaction between intermodal synchrony of input, a bottom-up source of information, and stored representations of one's own hands, a top-down source of information4. The idea is that the stimulus synchrony creates the impression that the real and the rubber hand are one and the same thing, and given that the rubber hand looks like a real hand, this impression is considered reality.
Later research by Kalckert and Ehrsson5 added a visuo-motor component to the rubber hand paradigm, which allows for the investigation of both perceived ownership (the impression that the artificial effector belongs to one's own body) and perceived agency (the impression that one is producing observed movements oneself). Participants were able to move the index finger of the rubber hand up and down by moving their own index finger, and the synchrony between real and rubber hand finger movements, the mode of movement (passive vs. active mode), and the positioning of the rubber hand (incongruous vs. congruous with regard to the participant's hand) were manipulated. The findings were taken to provide support for the notion that agency and ownership are functionally distinct cognitive phenomena: while synchrony of movement abolished both sense of ownership and agency, mode of movement only affected agency, and congruency of the rubber hand position had an effect on ownership only. The latter two result were replicated in a follow-up study in which the distance between real and rubber hand in the vertical plane varied6: ownership for the rubber hand decreased as its position increasingly mismatched the participant's real hand. However, agency was not affected by misplacements of the rubber hand in any condition.
However, recent research using virtual-reality techniques, which provide the participant with active control over the artificial effector, suggests that the role of the top-down part and the distinction between ownership and agency may have been overestimated7,8. These techniques have replaced the rubber hand by a virtual hand presented to participants on a screen in front of them or by means of virtual-reality glasses9. Participants commonly wear a dataglove that translates the movements of the participant's real hand into the movements of the virtual hand, either synchronously or asynchronously (e.g., with a noticeable delay). Similar to the rubber-hand illusion, synchronous translation strongly increases the participant's impression that the virtual hand becomes part of his or her own body10.
Employing virtual-reality techniques to create the rubber-hand illusion has several advantages over both the traditional rubber-hand paradigm and the combination of the rubber-hand paradigm with the visuo-motor components11. Moving one's hand and seeing an effector moving in synchrony with it creates a much more natural situation than facing a rubber hand and being stroked by an experimenter. Moreover, the virtual manipulation provides the experimenter with much more experimental flexibility and much more control over the perceptual relation between perceiving and moving one's real hand and one's perception of the event created by the artificial effector. In particular, using virtual techniques facilitates the manipulation of factors that are likely to influence perceived ownership and agency. For instance, the shape of the virtual hand can be modified much easier and faster than the shape of a rubber hand, and the movements of the virtual hand can be of any kind and for instance involve biologically impossible movements. Among other things, this facilitates exploring the limits of the illusion, as the artificial effector need not look like a hand but may be replaced by any kind of static or dynamic event. Of both practical and theoretical interest, a virtual effector is arguably a lot more immersive and feels a lot more real than a rubber hand, which is likely to reduce the necessity to invoke top-down interpretations to make sense of the present situation.
Ownership illusions have, however, not been restricted to hands. Tsakiris12 was the first to use the stroking technique to create the impression of participants that a static face in a picture presented in front of them is their own. Sforza et al.13 have also found evidence for this phenomenon, to which they refer as enfacement: participants incorporated facial features of a partner when their own and their partner's face were touched in synchrony. The neural mechanism underlying the enfacement illusion has recently been investigated by various researchers; for a comprehensive commentary and interpretation of the findings see Bufalari et al.14. We have recently turned the regular enfacement illusion design into a virtual-reality version (the virtual-face illusion), in which participants are controlling the movements of a virtual face in front of them by moving their own head15.
Here, we describe two experiments that used the virtual-hand illusion7 and the virtual-face illusion15 paradigms, respectively, to investigate self-representation. The virtual-hand experiment included three, completely crossed experimental factors: (a) the synchrony between (felt) real-hand and (seen) virtual-effector movements, which was either close to zero to induce ownership and agency or three seconds as a control condition; (b) the appearance of the virtual effector, which looked either like a human hand or like a rectangle (so to test the effect of real-virtual effector similarity on the ownership illusion); and (c) the opportunity to control the behavior of the virtual effector, which was either nonexistent in a passive condition or direct in an active condition. The virtual-face experiment included two, completely crossed experimental factors: (a) the synchrony between real-face and virtual-face movements, which was either close to zero to induce ownership and agency or three seconds as a control condition; and (b) the facial expression of the virtual face, which was either neutral or showing a smile, to test whether positive mood would lift the mood of the participant and improve his or her performance in a mood-sensitive creativity task.
All studies conformed to the ethical standards of the declaration of Helsinki and the protocols were approved by the Leiden University Human research ethics committee. Each condition tested about 20 participants.
1. Virtual-hand Illusion
Figure 1: (A) Participants wore an orientation tracker and a dataglove on their right hand, and SCR remote transmitter on their left hand. (B) Setup of the virtual-hand illusion experiment. (C) Setup of the virtual-face illusion experiment. (D) A screenshot of the computer screen. Please click here to view a larger version of this figure.
2. Virtual-face Illusion
Virtual-hand Illusion
We ran several experiments using the virtual-hand illusion paradigm, to investigate how people represent their bodies, in this case their hands. The number of tested participants depended on the amount of conditions, usually around 20 participants for each condition. Here we provide relevant results for one of the most elaborated studies we conducted in our lab. We will restrict our discussion to the subjective data, the average of the Likert-scale responses to the four ownership questions (O1-O4) and the Likert-scale response to the agency question (A1).
In this study8, we systematically investigated the effects of synchrony (synchronous vs. asynchronous), appearance of the virtual effector (virtual hand vs. rectangle), and activity (passive vs. active) on the participants' sense of ownership and sense of agency (all conditions were tested within participants). The results were the same for ownership and agency. As indicated in Figure 2, perceived ownership and agency were stronger if real and virtual hand moved in synchrony [F(1,43) = 48.35; p < 0.001; and F(1,43) = 54.64; p < 0.001; for ownership and agency, respectively], if the virtual effector was a hand than if it was a rectangle [F(1,43)=14.85; p < 0.001; and F(1,43) = 6.94; p < 0.02], and if the participant was active rather than passive [F(1,43) = 9.32; p < 0.005; and F(1,43) = 79.60; p < 0.001]. The synchrony effect replicates the standard virtual-hand illusion.
Figure 2: Ownership and agency ratings as a function of synchrony, appearance of the virtual effector, and activity of the participant. Please click here to view a larger version of this figure.
Figure 3: Ownership and agency ratings as a function of synchrony and activity of the participant. Note that the synchrony effect is more pronounced for active participants. Please click here to view a larger version of this figure.
Even more interestingly, both ownership and agency showed a significant interaction between activity and synchrony [F(1,43) = 13.68; p = 0.001; and F(1,43) = 23.36; p < 0.001; see Figure 3], but not between appearance and synchrony. This pattern suggests that activity plays a more dominant role for the ownership and the illusion than appearance does, it even showed that the illusory ownership perception is stronger in virtual than traditional rubber hand illusion paradigm. According to Hommel22, objective agency (i.e., the degree to which an external event can objectively be controlled) contributes to both subjective ownership and subjective agency, which explains why in this experiment, active, synchronous control over the virtual effector increased both subjective ownership and subjective agency.
While appearance failed to interact with synchrony, suggesting that the ownership illusion does not rely on appearance, it did produce a main effect. This indicates that appearance does have an impact on perceived ownership. It makes sense to assume that people have general expectations about what external objects might or might not be a plausible part of their body, which supports ownership perception in general but does not moderate the effect of synchrony. We thus conclude that multiple sources of information contribute to the sense of subjective ownership: general top-down expectations and bottom-up synchrony information. The relationship between these two informational sources does not seem to be interactive but compensatory, so that general expectations may dominate in the absence of synchrony, and vice versa.
Virtual-face Illusion
In another study, we investigated how people represent their face. We were able to replicate the traditional enfacement illusion in a virtual environment, which we refer to as the virtual-face illusion12. We further investigated whether people adopt the mood expressed by a virtual face they identify with. There was one within-participant factor-synchrony (synchronous vs. asynchronous) and one between-participant factor-facial expression (happy vs. neutral). The IOS ratings before the induction phase were subtracted from the IOS ratings after the induction phase, also the Affect Grid ratings before the induction phase were subtracted from the Affect Grid ratings after the induction phase, and these rating changes were used as the IOS and affect grid results.
The analysis of the ownership scores (O1-4), the agency scores (A1-2), and the IOS scale18 changes all showed main effects of synchrony [F(1,58) = 38.24; p < 0.001; F(1,58) = 77.33; p < 0.001; and F(1,58) = 43.63; p < 0.001; respectively], showing that synchrony between one's own head movements and the movements of the virtual face increased perceived ownership and agency, and facilitated the integration of the other's face into one's own self (see Figure 4). Synchrony also improved mood, as indicated by a synchrony effect on the affect grid19 changes [F(1,58) = 7.99; p < 0.01].
Figure 4: Ownership and agency ratings, as well as IOS changes, as a function of synchrony. Note that positive IOS changes imply an increase of integration of the other into one's self. Please click here to view a larger version of this figure.
Figure 5: Affect grid changes (positive values imply positive-going affect) and flexibility scores in the AUT, as a function of synchrony and the expression of the virtual face. Note that the interactions between synchrony and expression are driven by more positive-going mood and particularly good flexibility performance for the combination of synchrony and happy virtual face. Please click here to view a larger version of this figure.
There were significant main effects of the facial expression on IOS changes, affect grid changes, and flexibility in the AUT20,21,23but more important was the fact that the affect grid changes and the flexibility score interacted with synchrony [F(1,58) = 4.40; p < 0.05; and F(1,58) = 4.98; p < 0.05; respectively]. As shown in Figure 5, participants reported improved mood and showed more creative behavior after enfacing (i.e., synchronously moving with) a happy face as compared to the conditions where they moved asynchronously with a happy face or synchronously with a neutral face.
F/P/PES | EFF | ACT | SYN | EFF*ACT | EFF*SYN | ACT*SYN | EFF*ACT*SYN |
O1 | 11.66 | 10.11 | 45.38 | 10.08 | |||
0.001 | 0.003 | <0.001 | 0.003 | ||||
0.21 | 0.19 | 0.51 | 0.19 | ||||
O2 | 5.37 | 47.65 | |||||
0.025 | <0.001 | ||||||
0.11 | 0.53 | ||||||
O3 | 10.75 | 41.30 | 9.81 | ||||
0.002 | <0.001 | 0.003 | |||||
0.20 | 0.49 | 0.19 | |||||
O4 | 12.86 | 17.17 | 15.12 | 10.60 | |||
0.001 | <0.001 | <0.001 | 0.002 | ||||
0.23 | 0.29 | 0.26 | 0.20 | ||||
O1-4 | 14.85 | 9.32 | 48.35 | 13.68 | |||
<0.001 | 0.004 | <0.001 | 0.001 | ||||
0.26 | 0.18 | 0.53 | 0.24 | ||||
A1 | 6.94 | 79.60 | 54.64 | 23.36 | |||
0.012 | <0.001 | <0.001 | <0.001 | ||||
0.14 | 0.65 | 0.56 | 0.37 |
Table 1: F, P and Partial Eta squared (PES) values for the effects of the questionnaire item ratings, with df = 43. Factors are EFF: virtual effector (virtual hand vs. rectangle); ACT: activity (active exploration vs. passive stimulation); and SYN: synchrony (synchronous vs. asynchronous). Only results for significant effects are shown.
M/SE | H-P-SY | H-P-AS | H-A-SY | H-A-AS | R-P-SY | R-P-AS | R-A-SY | R-A-AS |
O1-4 | 4.37 | 3.44 | 5.09 | 3.50 | 3.79 | 3.14 | 4.68 | 3.05 |
0.20 | 0.23 | 0.19 | 0.25 | 0.23 | 0.23 | 0.20 | 0.21 | |
A1 | 3.59 | 3.11 | 6.36 | 4.36 | 3.07 | 2.57 | 6.09 | 3.80 |
0.30 | 0.32 | 0.15 | 0.33 | 0.28 | 0.27 | 0.24 | 0.33 |
Table 2: Means (M) and standard errors (SE) for the ownership and agency ratings in all eight conditions. H: hand; R: rectangle; A: active; P: passive; SY: synchronous; AS: asynchronous.
F/P/PES | Facial expression | Synchrony | Facial expression* Synchrony |
Ownership (O1-4) | 38.24 | ||
<0.001 | |||
0.40 | |||
Agency (A1-2) | 77.33 | ||
<0.001 | |||
0.57 | |||
IOS Changes | 4.03 | 43.63 | |
0.049 | 0.001 | ||
0.07 | 0.43 | ||
Affect Grid Valence Changes | 6.06 | 7.99 | 4.40 |
0.017 | 0.007 | 0.041 | |
0.10 | 0.13 | 0.07 | |
AUT-Flexibility | 5.42 | 4.98 | |
0.024 | 0.03 | ||
0.09 | 0.08 | ||
AUT-Fluency | 7.89 | ||
0.007 | |||
0.12 |
Table 3: F, P and Partial Eta squared (PES) values for relevant dependent measures, with df = 58 for questionnaire and IOS results, and df = 56 for the valence dimension of the affect grid mood and AUT results. Only results for significant effects are shown.
M/SE | Neutral-SY | Neutral-AS | Happy-SY | Happy-AS |
Ownership (O1-4) | 2.88 | 2.03 | 3.38 | 2.36 |
0.27 | 0.16 | 0.23 | 0.22 | |
Agency (A1-2) | 5.90 | 4.25 | 6.16 | 4.08 |
0.20 | 0.25 | 0.13 | 0.32 | |
IOS Changes | 0.37 | -0.80 | 1.00 | -0.40 |
0.21 | 0.25 | 0.20 | 0.24 | |
Affect Grid Valence Changes | -1.07 | -1.33 | 0.60 | -1.20 |
0.42 | 0.33 | 0.39 | 0.31 | |
AUT-Flexibility | 5.87 | 6.07 | 7.43 | 6.10 |
0.31 | 0.37 | 0.29 | 0.39 | |
AUT-Fluency | 7.27 | 8.27 | 9.73 | 7.37 |
0.51 | 0.68 | 0.68 | 0.49 |
Table 4: Means (M) and standard errors (SE) for relevant dependent measures in the four conditions. Neutral: neutral facial expression; Happy: happy facial expression; SY: synchronous; AS: asynchronous.
In this article we described two detailed protocols for the virtual-hand and virtual-face illusion paradigms, in which our virtual-face study was the first to replicate the traditional stroking-induced face-ownership illusion in virtual reality, together with representative results from the two paradigms.
The significant synchrony effects indicate that we were successful in inducing illusory ownership for the virtual hand and the virtual face, similar to more traditional illusion paradigms. Being able to reproduce these effects by means of virtual reality techniques has considerable advantages11,24. Virtual-reality techniques are freeing the experimenter from the rather artificial and interruptive stroking procedure and opens new possibilities for experimental manipulations. For example, morphing virtual effectors allowed us to systematically manipulate the impact of the appearance of the virtual hand and the similarity between the virtual and the participant's real hand, or the facial expression of the virtual face. The impact of agency can also be systematically explored by varying the degree (e.g., immediacy) to which participants can control the movements of the artificial effector.
Another promising avenue for future virtual reality research are first person perspective (1PP) virtual reality experiences. 1PP experiences can create an immense sense of immersion and feeling of presence, on a completely different scale than a third person perspective virtual reality experience25,26,27,28. In 1PP experiences one can truly feel like one is the avatar, that one is literally embodying the avatar. This opens up possibilities for all kinds of manipulations such as detaching parts of a person's body28, elongating29, rescaling body parts30, or changing a person's skin color31,32.
As the present and many other findings demonstrate, controlling virtual events in a synchronous fashion strongly increases the perception of these events belonging to one's own body. For example, our findings from the hand study suggest that immediate control is an important cue for distinguishing between self-produced and other-produced events (i.e., personal agency) and between self-related and other-related events (i.e., body ownership). The findings presented here and elsewhere suggest that bottom-up information plays a decisive role in the emergence of phenomenal self-representation, even for body parts that are not as identity-related as one's own body part4.
The most critical part of the described protocols is the induction process, which introduces correlations between visual, tactile and motor (i.e., proprioceptive) information-these correlations allows the cognitive system to derive ownership and agency. As these correlations rely on the relative timing of the respective events, such as the delay between the participant's own movements and the movements of the artificial effector, it is crucial to keep processing delays (especially with regard to the translation of data from the dataglove to the motion of the virtual effector on the screen) to a minimum. With our experiment setup the maximum time delay is around 40 ms, which is hardly noticeable and does not hamper the perception of causality and agency. Shimada, Fukuda, and Hiraki33 have suggested that the critical time window for the occurrence of multisensory integration processes constituting the self-body representation is 300 ms, which means that longer delays are likely to reduce the perception of control over virtual events.
Another important aspect of the protocol is tight experimental control over the participant's hand or face movements, depending on the paradigm. During induction, active movements of the respective factor are essential, as the required intersensory correlations rely on active explorative movements on the side of the participant. It is thus important to encourage participants to move frequently and to engage in active exploration. In other phases of the experiment, movements can impair the measurement however. For instance, in the virtual-hand illusion paradigm, moving the left hand (from which SCR was recorded) is likely to render measurements of the SCR level noisy and unreliable.
A limitation of the virtual-hand illusion paradigm technique is that, for practical reasons, participants commonly wear a dataglove and orientation tracker during the entire experiment (so to minimize distraction). This may not be comfortable, which in turn may affect the mood or motivation of the participant. One possible solution for that problem would be the use of lighter equipment or custom-made wearables. Another limitation of our current virtual-face illusion paradigm technique is that the equipment only registers head movements but no changes in facial expression. Allowing participants to control the facial expressions of a virtual face is likely to contribute to ownership illusions, but this would require hardware and software that provides reliable detection and categorization of facial expressions in humans-which we do not yet have available in our lab. The use of for example real-time (facial) motion capture utilities would be of great benefit in overcoming these limitations and would allow us to increase sense of agency and ownership of avatars to significantly higher levels.
As suggested by the findings from our study8, people consider various sources of information and update their body representation continuously. They seem to use bottom-up information and top-down information in a compensatory fashion, in the sense that one source of information plays a stronger role in the absence of the other-similar to what has been assumed for the sense of agency34. This provides interesting avenues for future research, as it for instance suggests that ownership can be perceived even for artificial effectors in awkward postures, provided a sufficient degree of surface similarity, or vice versa (i.e., if the artificial effector perfectly aligns with the real effector but differs from it in terms of surface features). The available findings also suggest that the boundaries between self and others are rather plastic, so that features of another person or agent can be perceived as a feature of oneself, provided some degree of synchrony between one's own behavior and that of the other35,36.
The authors have nothing to disclose.
This work was funded by the Chinese Scholarship Council (CSC) to K. M., and an infrastructure grant of the Netherlands Research Organization (NWO) to B. H.
Vizard (Software controlling the virtual reality environment) | Worldviz | Vizard allows importing hand models and integrating the hand, dataglove and orientation tracker modules through self-written command scripts. These scripts can be run to control the presentation of the virtual hand in the virtual environment, the appearance of the hand and the way it moves; they also control vibrator activities. | |
Cybertouch (Dataglove) | CyberGlove Systems | Cybertouch | Participants wear this dataglove to control the movements of the virtual hand in the virtual environment. Measurement frequency = 100 Hz; Vibrator vibrational frequency = 0-125 Hz. |
Intersense (Orientation tracker) | Thales | InertiaCube3 | Participants wear the Intersense tracker to permit monitoring the orientation of their real hand (data that the used dataglove does not provide). Update rate = 180 Hz. |
Biopac system (Physiological measurement device) | Biopac | MP100 | The hardware to record skin conductance response. |
Acquisition unit (Physiological measurement device) | Biopac | BN-PPGED | The hardware to record skin conductance response. |
Remote transmitter (Physiological measurement device) | Biopac | BN-PPGED-T | Participants wear the remote transmitter on their left hand wrist; it sends signals to the Biopac acqusition unit. |
Electrode (Physiological measurement device) | Biopac | EL507 | Participants wear the electrode on their fingers; it picks up skin conductance signals. |
AcqKnowledge (Software controlling acquisition of physiological data) | Biopac | ACK100W, ACK100M | The software to record skin conductance responses. |
Box | Custom-made | Participants put their right hand into the box | |
Computer | Any standard PC + Screen (could be replaced by VR glasses/devive) | Necessary to present the virtual reality environment, including the virtual hand. | |
Cape | Custom-made | Participants wear this cape on their right shoulder so they cannot see their right hand and arm. | |
Kinect (Head position tracker) | Microsoft | Kinect tracks the X-Y position of the participant's head. Recording frame rate = 30 Hz. | |
FAAST (Head position tracker software) | MXR | FAAST 1.0 | Software controls Kinect and is used to track the position of the participant's head. |
Intersense (Head orientation tracker) | Thales | InertiaCube3 | Intersense tracks rotational orientation changes of the participant's head. Update rate = 180 Hz. |
Facegen (Face-model generator software) | Singular Inversions | FaceGen Modeller | Facegen allows creating various virtual faces by varying various parameters, such as male/female-ness or skin color. |
Cap | Any cap, e.g., baseball cap | The cap carries the Intersense orientation tracker. | |
Computer | Any standard PC + Screen | Necessary to present the virtual reality environment, including the virtual head. |