Mobile brain and body imaging (MoBI), entailing multimodal electrophysiological recordings in freely behaving individuals, is a novel type of brain imaging offering the possibilities to understand the brain processes as they spontaneously evolve during natural activities, such as walking or running (Gwin et al., 2010), cycling (Zink et al., 2016), driving (Protzak and Gramann, 2018), talking or free viewing of environments (Roberts et al., 2018; Soto et al., 2018). A wireless electroencephalographic (EEG) recording is an essential component in MoBI experiments to allow for unobstructed continuous recordings during natural cognition.
A feature which often distinguishes MoBI experiments from laboratory-based experiments is the participant’s posture with laboratory experiments being traditionally conducted in seated individuals. Standing is a seemingly simple activity, at times assumed equal to sitting recumbent or laying supine. However, maintaining a standing balanced posture requires the rapid processing and integration of continuous vestibular, somatosensory and visual information (Mergner and Rosemeier, 1998; Nashner, 1976), reviewed in Thibault and Raz (2016).
How body posture affects brain activity has been difficult to study in the past and as such, our scientific understanding is limited. The lack of insight could be partially attributable to the inherent postural requirements of modern brain imaging techniques. The ecological nuances associated with functional magnetic resonance imaging (fMRI) have previously been described (Raz et al., 2005). These revolve around the necessity of supine participants to be restrained and to remain still as stimuli suddenly appear on screen. Positron emission tomography (PET) with modified scanner gantries have been used to study effects of posture on brain activations (Harmon-Jones et al., 2011; Riskind and Gotay, 1982). However, the use of radioactive contrasts as well as the size of the scanner limit the possibility of PET to explore effects of posture on brain processing in freely behaving individuals. EEG recordings classically position sitting subjects alone in dimly lit rooms and viewing monitors, whereas magnetoencephalography (MEG) uses both sitting and supine participants. In all cases, body and head movements are often minimized via chin rests and head braces. The most glaring differences among these imaging modalities is the participant’s posture during recordings and the type of restrictions to body movements. Furthermore, the experimental tasks being performed are, at times, incongruent with the body position that subjects must assume; for instance, performing arithmetic calculations while laying supine (Lipnicki and Byrne, 2008) or performing a driving task inside the scanner environment (Schweizer et al., 2013).
Moreover, modifications to body posture have been shown not only to affect physiological processes such as blood pressure regulation and brain activations, but also behavior and cognition (reviewed in Thibault and Raz, 2016). Standing on both feet compared to supine posture has been shown to increase activation in the right primary visual cortex (Ouchi et al., 1999) and cerebellum (Ouchi et al., 1999, 2001). Upright stances also aid psychomotor performance (Caldwell et al., 2003; Lazarus and Folkman, 1984) and increase anticipatory anxiety during mental arithmetic tasks (Lipnicki and Byrne, 2005). Similarly, a sitting posture has been shown to improve intelligence and augment perception compared to a supine posture (Lundström et al., 2006) and adopting certain postures (i.e., slumping) can reduce emotional responses (Harmon-Jones and Peterson, 2009) and influence rhythmic cortical activity, behavior (Harmon-Jones et al., 2011) and cognitive conflict processing (Sun and Harmon-Jones, 2021). Even small body movements like the adoption of different facial expressions affect emotional judgments and the generation of memories (Laird and Lacasse, 2014). Other studies have shown that leaning direction plays an important role in affect perception (Harrigan and Rosenthal, 1983; Troncoso et al., 2024). These findings suggest that neural processing of emotional stimuli could vary resulting from changes in body positions, alluding to a close-knit relationship between posture and emotional processing. We have previously shown that wirelessly recorded face-sensitive visual evoked components in freely behaving humans differentiated faces displaying four different emotional expressions (Soto et al., 2018). Here, the rationale is to build upon these previous findings to examine the effect of body posture on face processing in the brain. Such effects have been sparsely reported in laboratory studies in the past (Batty and Taylor, 2003; Blau et al., 2007) and a number of laboratory studies failed to show effects of emotional expressions on the face-sensitive N170 component (Chai et al., 2012; Eimer et al., 2003; Hendriks et al., 2007). It is possible however, that an upright stance may increase general arousal and certain emotional processes.
In the present study, we implemented wireless EEG recordings to analyse effects of posture on the early visual evoked component and face-related evoked components. Similar to our previous study (Soto et al., 2018), participants viewed human faces displayed on posters while they were moving or standing freely, and their EEG and eye movements were recorded continuously. In a different session and using the same types of recordings, participants viewed posters with faces in a reclining-supine posture. Faces showed one of four emotional expressions: happy, fearful, disgusted or neutral. We hypothesized, in line with previous data showing increased visual cortex activation in upright stance compared to supine posture (Ouchi et al., 1999), that upright stance would reduce the amplitude of both the early visual EMRP component and the face-sensitive EMRP component due to visual processing demands employed during standing.
Materials and methods ParticipantsFifteen healthy volunteers 24.6 ± 3.2 years old (mean ± SD), were recruited for the study by online advertisements and word of mouth. Subjects that wore corrective eye-glasses were excluded from recruitment due to the glare artifact the glass generates in the wearable eye-tracking system. Subjects wearing contact lenses were not excluded from the recruitment as these lenses do not interfere with the eye-tracking system, but were required to wear them for both sessions. One participant from the sample was rejected prior to data processing due to an excessively noisy EEG signal during the reclining session which caused a loss of more than 20% stimuli for one experimental condition. Therefore the final sample was composed of 14 subjects (eight females) with an average age of 24.6 ± 3.3 years. Experimental subjects gave written informed consent prior to taking part in the study in agreement with the ethical approval obtained from the University of Liverpool Research Ethics Committee. Participants received £30 (£15 per session attended) as compensation for their travel expenses and time. All experimental procedures were conducted in two sessions and in accordance with the Declaration of Helsinki.
StimuliFace images expressing four emotional categories (fearful, disgusted, happy and neutral) were selected from the Karolinska Directed Emotional Faces (KDEF; Lundqvist et al., 1998) set and from the NimStim face inventory (Tottenham et al., 2009). The full stimulus set consisted of 180 emotional human faces, 45 images in each emotional category presented on 20 A0 poster-size sheets (841 × 1,189 mm). A total of 120 images were taken from the KDEF and 60 from the NimStim set. All 180 images used in the experiment were cropped to match in size and quality (300 DPI) using CorelDRAW software. Additionally, all images were matched for temperature and brightness and were always presented as standard portrait shots including the shoulders. A0 size panels were constructed containing nine different image stimuli presented in full color approximately 20 × 25 cm in size, with a white-on-black square fixation cross presented in the centre (14.3 × 14.3 cm) (Figure 1A). Twenty panels contained all emotional face categories pseudo-randomly distributed and maintained similar distances from the fixation cross in the centre. All panels were pasted onto Styrofoam sheets and attached to the walls using adhesive tape creating a mock gallery (Figure 1B).
Figure 1. Experimental setup of the mock gallery and reclining sessions. (A) Example of one poster panel (120 × 90 cms) presenting face images. (B) Schematic illustration of the hallway used to setup the mock gallery. Black lines indicate the position of each poster within the mock gallery. (C,D) One subject wearing the full equipment and positioned in the experimental setups for both the standing and reclining sessions. (1) Pupil eye tracker; (2) MOVE wireless EEG transmitter; (3) 64 channel EEG actiCap; (4) Eye tracking PC in the backpack. Face images were selected from the Karolinska Directed Emotional Faces (KDEF; Lundqvist et al., 1998) set and from the NimStim face inventory (Tottenham et al., 2009).
EEG recordingsEEG data was continuously recorded using a 64-channel wireless and portable EEG system (Brain Products, GmbH, Münich, Germany). Brain Products MOVE system (Brain Products, GmbH) employs a lightweight signal transmitter which participants carry on a belt (Figures 1C,D). Active shielding Ag/AgCl EEG electrodes were mounted on an electrode cap (actiCAP, Brain Products, GmbH) according to the international 10–20 electrode system, aligned to the midpoint between the anatomical landmarks of the nasion and inion, and left and right preauricular points. Electrode FCz was used as the system ground and electrodes were referenced to Cz. Electrolyte gel was applied to lower electrode-to-skin impedances to under 50 kΩ. EEG recordings were sampled at a rate of 1,000 Hz. EEG average reference was applied and the signals were digitized at 1 kHz with a BrainAmp DC amplifier linked to BrainVision Recorder program version 1.20.0601 running on Windows operating system. Data were filtered online using a 0.1–200 Hz bandpass filter.
Eye movement recordingA PUPIL (Kassner et al., 2014) binocular eye-tracking system was used to record the locations of the gaze position. PUPIL Capture software v 0.9.6 running on Ubuntu v 14.04.4 was used to generate the recordings. PUPIL eye tracker is an open source, high resolution wearable system (Figure 1C) that provides a lightweight solution for mobile gaze-tracking. Pupil locations were recorded using infra-red cameras sampling at 90 Hz with a resolution of 320 × 240 pixels. The real-world video streams were set at a sampling rate of 60 frames per second with a resolution of 600 × 800 pixels.
To calibrate the gaze locations, a manual marker 3D calibration protocol was used to generate a 9-point grid in the field of view of the participant. Calibration was repeated until gaze positions were accurate everywhere on the blank panel. Small calibration offsets occurred at times due to displacements in the wearable eye-tracker on the subject’s face. These offsets were adjusted offline using the manual gaze correction plug-in on PUPIL Player during manual tabulation of stimulus onset times. Eye-tracking data were processed using PUPIL Player v. 0.9.6 program. The ocular pupils of both eyes were located based on a centre-surround detection algorithm (Świrski et al., 2012). Exported raw gaze data is time-locked to the processing computer’s internal clock, giving millisecond precision to the eye measurements. Additionally, all recorded frames contained an accurate time stamp based on the PC processor real-time clock. Eye-tracking video files were visually inspected and stimuli onsets were manually tabulated. Each stimulus was logged on a picture-by-picture basis with stimulus onset defined as the first instance in which the gaze position landed on an image. The real times corresponding to the tabulated frames were used to import stimulus onset latencies onto the raw EEG data. Eye-tracking and EEG data streams were aligned at the start of each recording using a custom-built light-emitting trigger box, which flashed a light into the eye-tracking camera and synchronously registered a transistor logic pulse in the EEG data recording.
ProcedureVolunteers attended two testing days, a standing session where stimuli were viewed while upright, and a reclining session where they sat in a recumbent position on a lounge chair. In both sessions, instructions were delivered and equipment was set up in a designated lab space following identical procedures. The order of conditions was counterbalanced across subjects with half of the participants taking part in the reclining task first, and the remaining participants taking part in the standing task first. Experimental sessions were separated by a minimum of 1 week.
Standing sessionParticipants were fit with the EEG cap and electrode-to-skin impedances were lowered below 50 kΩ before commencing the task as well as during the break between blocks. The mobile EEG system was connected and wireless signals were visually inspected on a standing participant. Next, eye-tracking glasses were placed on the participant over the EEG cap.
A mobile base unit was assembled using a rolling trolley where the wireless signal receiver, EEG amplifier and recording computer were placed. The base unit was positioned by the experimenter, keeping a distance of no more than 7 m from the participant in order to maintain optimal signal quality. The eye-trackers were calibrated on a standing participant against a blank white panel at a distance of 1 m from the centre of the panel to match the viewing conditions during the mock gallery. Gaze-tracking was optimized using a 3D calibration routine with manual markers. The laptop registering the eye-tracking was placed in a backpack and carried by the participant for the duration of the task (Figures 1C,D). EEG cables running from the electrodes to the lightweight transmitter were also placed in the backpack to reduce cable sway artifacts (Gramann et al., 2010; Gwin et al., 2010).
Two hallways within the Eleanor Rathbone Building of the University of Liverpool were used to create a naturalistic picture gallery where the standing session took place (Figure 1B). Participants were free to view images in any order and to navigate the gallery however they chose. All subjects were asked to stand still and view each image for a minimum of a few seconds before moving onto the next image. To do so, subjects were instructed to stand at a set distance of 1 m and remain still. They first looked at the centre fixation cross before viewing each image, and were requested to return their gaze to the fixation cross before moving to the next picture. Participants only continued onto a subsequent panel after viewing all images. To maintain task engagement and to avoid drowsiness, participants were asked to select their most and least preferred face from each panel and to mark the selected pictures on a small paper-printed version of each panel. The full experiment consisted of two blocks in the mock gallery. In each block, subjects viewed 10 panels with nine images on each one. In total, participants viewed 180 different images in the experiment. On average, each of the two gallery blocks lasted approximately 15 min.
Reclining sessionThe wireless EEG was placed on the participant in the same manner as in the standing session. For the reclining task, the eye-trackers were calibrated on a blank white panel with the participant lying on the lounge chair. The blank panel was hung at a distance of approximately 1 m with 140° angle to match viewing conditions across both sessions. The inclination angle of the lounge chair was lifted to approximately 40° (Figure 1D) bringing the fixation cross in the centre to the participant’s eyeline. In this session, subjects laid on a lounge chair that provided full body support as high as the neck. A foam neck support was used to prevent electrodes on the back of the scalp smearing against the chair and provided head support. An identical 3D calibration routine using manual markers was used to optimize the gaze-tracking for the reclined condition. The eye-tracking computer was placed on the same mobile base unit where the EEG recording computer rested in this session.
Participants were instructed to view the images displayed on each panel. The experimenters hung each panel as they were completed by the participants. Subjects were instructed to relax and view each face for a couple of seconds before returning to the fixation cross and moving onto the next image. As in the standing session, no restrictions were placed on eye-movements or blinks. The same selection task was used here to maintain engagement and avoid drowsiness. Participants viewed the same 180 different images in the experiment divided into two blocks with breaks between them. Each block in the reclined session also lasted approximately 15 min.
In both sessions, electrode impedances and gaze tracking calibration were checked in the break between blocks and corrected if required. Once the viewing task was completed, the EEG cap and the eye-tracking glasses were removed. Participants were then required to rate how much they liked and if they would approach the images that had been previously seen. Ratings were performed using three visual analogue scales (VAS) sized 10 cm and anchored on each extreme. The scales evaluated the general likability of each face image as well as the level of arousal and emotional valence. Subjects were instructed to use the middle of the scale for expressions deemed neutral. (i.e., “0: Do not like” up to “100: Like very much”; “0: Unarousing” up to “100: Very Arousing” and “0: Very negative” up to “100: Very positive”). Pictures and rating scales were presented on an LCD screen using Cogent program v. 1.32 (Welcome Department of Imaging Neuroscience, United Kingdom) running on MATLAB v. R2014a (MathWorks, Inc., USA). Participants rated the same faces after each session.
EEG analysis Eye movement related potentialsEEG data were pre-processed using the Brain Electrical Source Analysis program (BESA v.6.0, MEGIS Software GmbH, Munich, Germany). Data were initially referenced using the common averaging method (Lehmann, 1987) and digital filters were applied to remove frequencies lower than 0.5 Hz and higher than 35 Hz from the EEG data to minimize electromagnetic interference (EMI) from electrical appliances. Ocular artifacts including blinks as well as vertical and horizontal eye-movement artifacts were removed using pattern matching averaged artifact topographies (Berg and Scherg, 1994; Ille et al., 2002). This method is based on a principal component analysis and creating individualized artifact topographies to maximally match and resolve eye-movement artifacts in EEG data. Additionally, EEG data was visually inspected and corrected for the presence of remaining artifacts. Trials were excluded if artifacts were present in either eye-tracking or EEG data. If participants skipped an image, failed to fixate, or gaze tracking was lost during fixation, the trial was discarded from any further analysis. To ensure consistency across our conditions, subjects with more than 20% rejected epochs for one experimental condition were eliminated from the final analysis.
Individual stimulus onset times were detected by visual inspection of the eye-tracking data. For all stimuli the first instance where a participant’s gaze made contact with a stimulus image was registered as visual onset time for event marking. Event markers were then inserted into EEG data by synchronizing the time axes of the EEG and eye-tracking system in MATLAB (The MathWorks Inc., 2018). EEG data were epoched to range from −200 ms to 600 ms relative to the first contact of the gaze with any part of a picture in each of 180 pictures. This time point effectively corresponded to part of the saccade which brought the gaze onto a particular face or object in a picture. Artifact-corrected EEG signals were then exported into EEGLAB v.14.1.1, an open-source environment for processing EEG data (Delorme and Makeig, 2003). During averaging, the mean EEG activity in the baseline interval ranging from −200 ms to −100 ms was removed from each data point as it represented a more stable baseline relative to the −100 ms to 0 ms in free viewing conditions. Post-saccadic EMRPs were computed from all trials falling into eight different conditions including postural conditions (reclining and standing) and emotional facial expressions (afraid, disgusted, neutral and happy). The sampling rate of the eye-tracking system was previously calculated offline (average sampling rate: 41.1 Hz). The sampling rate was chosen based on 15-min pilot experiments to secure a continuous stream of eye-tracking data which was often discontinuous at higher sampling rates. However, due to the relatively low sampling rate, further analysis of the eye movements was not performed and eye-tracking data was used exclusively to detect stimuli onset times. EEG epochs were generated as cuts into the world video scene in a similar manner as the method described in our previous experiment (Soto et al., 2018). Artifact-corrected EEG data was then exported to EEGLAB for group-level analysis.
Data analysis and independent component clusteringIndividual subject data was entered into the EEGLAB STUDY structure (Delorme et al., 2011). Independent component analysis (ICA) was performed within EEGLAB for all subjects using an extended Infomax ICA algorithm to parse EEG by defining maximally statistically independent components (IC) (Dimigen, 2020; Makeig et al., 1996) using the default training parameters on EEGLAB. This method allows an accurate detection and removal of extra-cerebral sources of noise. ICA was performed on concatenated epochs for each subject individually. Event-related potentials (ERPs) and IC scalp maps were then computed across experimental conditions and used to construct a pre-clustering array to group ICs into 9 clusters. The DIPFIT2 routine from EEGLAB was used to model each IC as an equivalent current dipole by using a standardized boundary element (Oostenveld and Oostendorp, 2002). If the residual variance of the single-dipole exceeded 30%, or if the location landed outside the head, the IC was removed prior to clustering. The clustering of ICs was performed across all 14 subjects based on similarities in scalp topographies, equivalent dipole locations and ERP waveforms by means of EEGLAB’s k-means clustering routine (Lee et al., 2015). Three clusters of interest were identified and selected based on IC activity, average dipole locations and the presence of ICs in at least nine subjects. Furthermore, the two peak features in the global field power were used to identify component clusters for analysis. Cluster 2 presented 33 ICs across 13 subjects and accounted for the N170 component. Cluster 8 (16 ICs), clustered from nine subjects, presented a small peak in activity at around 90 ms and accounted for the vertex positive potential (VPP) peaking in medial scalp locations and a larger secondary peak at around 170 ms. Cluster 6 was composed of 18 ICs present in 10 subjects and represented the major contributor to the lambda component (Kazai and Yagi, 2003; Vigon et al., 2002). Statistical testing was performed on the grand averaged IC cluster activity across all experimental conditions for these three components.
Statistical analysisThe source waveforms representing the averaged clustered IC activity in each of the fitted dipoles were analysed using a 2 × 4 ANOVA for repeated measures ANOVA (Reclining vs. standing, four emotional face conditions). Specifically, a permutation-based repeated-measures ANOVA utilizing 5,000 permutations was employed (Maris and Oostenveld, 2007). Statistical testing was performed on the averaged source activity across the intervals of interest in SPSS v.22 (IBM Corp., NY, USA). We focused the analysis on 20 ms time intervals (−10 to 10 ms) relative to the peak in IC activity for each selected cluster. Post-hoc paired t-tests were performed when necessary and considered significant at p ≤ 0.05 and Bonferroni corrections for multiple comparisons were always used.
Results Face ratingsFigure 2 shows the average rating scores for each visual analogue scale across all four experimental conditions. To test the effects of body posture on the overt rating behavior, each of the three individual visual analogue scale ratings were submitted to separate 2 × 4 ANOVAs for repeated measures. Across all rating scales, we found no effects of posture observed on overt rating behavior between standing and reclining conditions [arousal scale: F(1, 14) = 0.009; p = 0.927; emotion detection scale: F(1, 14) = 1.112; p = 0.31; and pleasantness scale: F(1, 14) = 0.251; p = 0.624]. A significant effect of emotional condition on arousal level was also found in this analysis [F(3, 42) = 100.62, p < 0.001] as well as for the emotion [F(3, 42) = 69.008, p < 0.001] and pleasantness scale [F(3, 42) = 84.883, p < 0.001].
Figure 2. Grand average ratings for each visual analogue scale. Light color bars represent the standing condition while dark color bars represent the reclining condition. Error bars represent 1 standard deviation. Asterisks represent significance at p < 0.05.
Pairwise t-tests were used to test differences among all pairs of emotional categories. Paired comparisons confirmed that arousal ratings were significantly different across emotional expressions specifically; disgusted expressions rated as less arousing than afraid facial expressions [t(14) = −6.1; p = 0.003] and more arousing than neutral faces [t(14) = −2.9; p = 0.012]. Unexpectedly, afraid and neutral facial expressions were rated as equally arousing by our participants [t(14) = 0.104; p = 0.92]. Similarly, the emotional scale ratings showed happy faces (80.7 ± 7.8, mean ± SD) were consistently rated as highly positive while disgusted (35.2 ± 6.6), and afraid (39.1 ± 5.7) were rated on the negative end of the scale. Neutral facial expressions, however, were perceived to be slightly negative (41.1 ± 6.8). As expected, this effect was additionally encountered in the pleasantness ratings where faces reflecting happy emotional expressions were perceived as more pleasant than disgusted [t(14) = −12.6, p < 0.001], afraid [t(14) = 14.1; p < 0.001] or neutral faces [t(14) = 10.4, p < 0.001]. Again, neutral faces were rated equally pleasant as disgusted faces [t(14) = −2.1; p > 0.05] and afraid expressions [t(14) = 0.012, p > 0.05] irrespective of the postural position.
Eye movement related potentialsEMRPs were constructed from −200 ms to 600 ms relative to the first instance of contact of the subject’s gaze with an image (Figure 3). Given participants remained still to view images the EEG data showed minimal neck muscle or head movement artifacts, which have been shown to affect EEG data during walking or running (Gwin et al., 2010).
Figure 3. Grand average EMRP for the reclining and standing conditions. (A) Butterfly plots of the average EMRP waveforms between 200 ms and 600 ms. Peak latencies and scalp topographies overlaid on the volume rendering of the human head at three latency points in reclining (−25, 90, and 170 ms) and standing conditions (−24, 91, and 170 ms). EMRP components are highlighted with arrows. (B) Global field power of the EMRP for the reclining and standing session.
Independent component clustersIC clusters were selected on the basis of their maximum explained variance being >70% and that clusters were composed of IC from a minimum of seven subjects. Additionally, the spatial–temporal characteristics of the scalp maps and dipole locations were used to select ICs.
Three IC clusters were selected based on the peak in their averaged waveforms and scalp distributions (Figure 4). The mean IC dipole for each IC cluster of interest was located in, or near to the posterior cingulate cortex (cluster 2), the left thalamus (cluster 6) and the left angular gyrus (cluster 8). The average waveform activations of each of these IC clusters were statistically tested using a 2 × 4 (body posture × emotional condition) ANOVA for repeated measures.
Figure 4. Independent components composing the three IC clusters of interest. (A) Overlaid scalp plots of each cluster of interest (IC cluster 1, IC cluster 8, and IC cluster 6). Scalp plot polarities presented here are arbitrary. The shaded time interval where statistical testing was performed. (B) Average IC cluster locations and projection lines on a three-slice MRI template (red sphere represents the averaged dipole location for cluster 2, blue sphere represents the averaged dipole location for cluster 6 and green represents the averaged dipole location for cluster 8).
Cluster 2 presents as a tightly grouped cluster of 33 ICs in occipito-temporal portions of the brain present in 13 subjects. However, we must interpret averaged cluster localizations findings with caution given that dipole localization in EEG, while informative, lacks the spatial resolution to definitively determine the precise neural generators of the observed activity.
Figure 5A illustrates the averaged IC activity across experimental conditions as well as the location of individual ICs composing clusters (Figure 4B). Due to the peak latency and estimated average location of the cluster, we have associated it to the N170 face-sensitive component. The average IC dipole location of this cluster was estimated to be located medially in the posterior cingulate cortex behind the splenium of the corpus callosum (Talairach coordinates x, y, and z: 9, −48, 23; Brodmann Area 23). The mean IC dipole accounted for 86.72% of the total explained variance (13.28% RV ± 8.8; x̄± SD). Across the different experimental conditions, cluster 2 showed peaks in mean activity occurring between 150 ms and 230 ms with an average peak at 189 ms. The ANOVA performed over the 180 ms to 200 ms time interval reflected no effects of body position on cluster 2 [F(1, 13) = 0.033; p > 0.05]. No effects of emotional condition were present in the data [F(3, 39) = 0.776; p > 0.05] nor was there an observable interaction effect [F(3, 39) = 0.235; p > 0.05].
Figure 5. Average IC cluster activations for the N170 and VPP components. (A) IC activity accounting for the N170 between −200 ms and 600 ms. (B) IC activity accounting for the VPP between −200 ms and 600 ms. Red lines represent the standing condition and black lines represent the reclining condition. Shaded areas represent the time intervals where statistical testing was performed.
Cluster 8 was composed of 16 components present in nine subjects. The scalp distributions presented over the occipito-parietal regions of the scalp with a first small deflection occurring at roughly 100 ms and a main, larger peak at a latency of about 200 ms (Figure 5B). This cluster modeled the VPP which, given their co-occurrence and shared generators, has been theorized to represent the same brain mechanism as the N170 component (Joyce and Rossion, 2005). The averaged dipole explained 81.17% of the residual variance (19.83% ± 5.94; x̄ ± SD) and was located to the left angular gyrus (Brodmann Area 39, Talairach coordinates: −26, −62, 17). Statistical comparisons were performed in the time interval between 90 ms and 110 ms as well as between 190 ms and 210 ms. Within the first time interval, no significant effect of posture on evoked source activity of cluster 8 [F(1, 13) = 2.603, p > 0.1] was found in the latency period surrounding the peak. Additionally, no emotional effect [F(3, 39) = 0.198, p > 0.5] nor was there any effect of interactions between posture and emotional categories [F(3, 39) = 0.99, p > 0.1]. A similar pattern of results are in the interval ranging from 190 ms to 210 ms with no effects of posture [F(1, 14) = 0.007, p > 0.5] or emotional category [F(3, 39) = 1.426, p > 0.1] nor were there any interaction effects present [F(3, 39) = 1.677, p > 0.05] in the data.
Cluster 6 was comprised of 18 ICs from 10 of the subjects which were active within similar latencies as the lambda potential (Figure 6). The scalp distribution of the component shows a clear occipital distribution. The averaged IC dipole was fit to the left thalamus (Talairach coordinates −10, −6, 9) and the mean dipole accounted for 84.6% of the explained variance (16.4% ± 9.63). Cluster 6 activations showed significant effects of body posture on IC activity [F(1, 13) = 6.424; p < 0.05]. The averaged IC activity was larger when subjects reclined (−0.15 ± 0.621 μV) relative to when they stood (−0.085 ± 0.415 μV). There was no effect of emotion on the mean IC cluster activity [F(3, 39) = 0.475; p > 0.05] nor was an interaction effect found in this interval [F(3, 39) = 0.694; p > 0.05].
Figure 6. Mean IC cluster activation for cluster 6. (A) IC activity between −200 ms to 600 ms for each emotional condition. Red lines represent the standing condition and black lines represent the reclining condition. (B) Bar graphs represent the averaged IC activity over the shaded area between 110 and 130 ms.
DiscussionThe present study analysed EMRPs recorded wirelessly during viewing face photographs in an upright stance and a reclining posture. Results show that upright body postures reduced the lambda component, but no effect of body posture was found in the N170 component amplitude. None of the EMRP components differentiated emotional expressions of faces.
The lambda response has been identified primarily as a visual response during natural viewing (Thickbroom et al., 1991) and is regarded as the analogue of the P100 component elicited in laboratory-based environments (Kazai and Yagi, 2003). The lambda potential is similarly elicited by the afferent inflow beginning at onset of the visual stimulus (Yagi, 1981) and is thought to represent rapid indexing of visual stimuli and, like the P100 component, is also modulated by low-level visual features (Pourtois et al., 2005) and saccade size (Dimigen et al., 2011). Previous research has shown that leaning forward in a sitting position enhances the P100 component in response to sexually erotic images (Price et al., 2012), and shortens reaction times and increases left frontal cortical activity in response to appetitive food cues when compared to recumbent sitting positions (Harmon-Jones et al., 2011).
Our finding of a reduced lambda component during standing compared to reclining can be interpreted in the light of previous study demonstrating an increased visual cortex activation during standing compared to sitting (Ouchi et al., 1999). Such increased activation associated with standing might reduce the capacity of the visual system to process certain features of the environment which are not related to locomotion or maintaining an upright stance. It is therefore possible that reclining posture favored focused attention during free viewing of faces. It has been suggested that the P100 component also reflects discrete attentional mechanisms in which the enhancement of the P100 component would be invoked by the suppression of unattended stimuli (Clark and Hillyard, 1996). The focusing of visual attention has consistently yielded enhanced P1 amplitudes in laboratory-based testing in the past (Di Russo and Spinelli, 1999; Hillyard and Anllo-Vento, 1998). The P100 enhancement effect has been explained by the recruitment of additional extrastriate neurons when attention is focused on the image (Luck et al., 2000). In general, it is well established that visual evoked potentials are larger during focused attention (Van Voorhis and Hillyard, 1977) and previous ERP research has shown color, shapes and motion can modulate visual ERPs starting at about 100 ms (Anllo-Vento et al., 1998; Torriente et al., 1999). It is therefore likely that an increased lambda component in reclining posture is related to an increased capacity of the visual cortex to engage in a visual task which may be reduced during standing due to the increased posture-related activation of the visual cortex.
Another factor potentially contributing to an increased lambda component in reclining posture compared to standing may refer to the amount of low-frequency theta and alpha oscillations in different postures. Upright stance has been associated with increased high-frequency gamma oscillations but reduced frequency delta, theta and alpha oscillations (Chang et al., 2011; Thibault et al., 2014). The reduction of theta-and alpha-band oscillations during upright stance is relevant for interpretation of decreased lambda EMRP component during standing because the phase-locked evoked potentials can be viewed as a phase reset of underlying spontaneous oscillations (Brandt, 1997; Hanslmayr et al., 2007; Mazaheri and Picton, 2005). This aligns with previous studies showing supine postures reduce frontal alpha activity during rest (Price et al., 2012) and viewing of emotional images (Harmon-Jones and Peterson, 2009) Previous research into the energy distribution in the frequency domain has shown that the P100 component corresponds to increases in power in a wide range of frequency bands, primarily within the theta-and alpha-bands (Quiroga et al., 2001). This might not be the case for later cortical components, however. Recent ERP research has shown a supine body posture would reduce the amplitude of the error-related negativity (ERN) component due to a lowering in approach motivation (Sun and Harmon-Jones, 2021).
Our finding of a decreased lambda component during standing compared to reclining can, in an unknown manner, also be related to mechanical factors such as displacement of the brain inside the skull and changes in thickness of the cerebral spinal fluid surrounding the brain (Thibault and Raz, 2016). Changing from a prone to supine position has shown to generate decreases in cerebrospinal fluid thickness of approximately 30% which can in turn strongly influence occipital EEG signals (Rice et al., 2013). Further studies must address the potential effects of different propagation of electrical currents from the brain in different body postures. However, the lack of posture effects on the face-sensitive EMRP in the present study suggests that the mechanical and current propagation factors played a small role since all EMRP components would likely be affected.
To conclude, we utilized wireless EEG recordings to demonstrate that posture selectively reduces the amplitude of EMRPs. Specifically, upright stances reduced the amplitude of early lambda but not face-sensitive N170-like EMRPs, compared to reclining postures. As outlined, these findings are consistent with previous data showing increased visual cortex activation and increased attentional allocation during standing compared to sitting. The finding of a reduced lambda component in upright stance is important for the evaluation of subsequent MoBI studies involving visual processing in freely moving or standing individuals in natural settings, as early visual evoked components may be attenuated compared to laboratory-based recordings.
Finally, since participants were free to move naturally, their head movements were neither restricted nor controlled. This introduces the possibility that subtle differences in head position and movement across experimental conditions may have influenced the EEG signals. Such challenges are common in mobile EEG research and warrant further attention. Currently, only one project, ICMOBI (2024), is developing analysis tools to automatically distinguish brain and non-brain components in mobile EEG data. Future studies employing mobile EEG and eye-tracking will benefit from standardized preprocessing pipelines and automated artifact detection methods. While active electrodes significantly reduce the impact of electromagnetic interference (EMI), further advances in electrode shielding and robustness will also be necessary. Moreover, it is important to acknowledge the limitations of using EEG for dipole localization. Future research should adopt multimodal imaging approaches, such as mobile fNIRS-EEG studies, to more accurately resolve the neural sources of these components in the real world.
Data availability statementThe raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statementThe studies involving humans were approved by Research Integrity and Governance Committee Secretary, bWpiaWxsMkBsaXZlcnBvb2wuYWMudWs=. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.
Author contributionsVS: Investigation, Writing – original draft, Writing – review & editing. JT-C: Formal analysis, Investigation, Writing – review & editing. KK: Investigation, Writing – review & editing. HR: Investigation, Writing – review & editing. AB: Investigation, Writing – review & editing. DH: Investigation, Writing – review & editing. NF: Conceptualization, Investigation, Supervision, Writing – review & editing. TG: Conceptualization, Writing – review & editing. AS: Conceptualization, Data curation, Investigation, Writing – original draft, Writing – review & editing.
FundingThe author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was supported by the Engineering and Physical Science Research Council of the United Kingdom (Award reference: 1566731), ANID/Fondecyt (grant no. 11221227), and UNILEVER.
AcknowledgmentsWe are grateful to Julia Jones for assisting with organizational and technical aspects of the study.
Conflict of interestThe authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s noteAll claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Footnotes ReferencesAnllo-Vento, L., Luck, S. J., and Hillyard, S. A. (1998). Spatio-temporal dynamics of attention to color: evidence from human electrophysiology. Hum. Brain Mapp. 6, 216–238. doi: 10.1002/(SICI)1097-0193(1998)6:4<216::AID-HBM3>3.0.CO;2-6
PubMed Abstract | Crossref Full Text | Google Scholar
Batty, M., and Taylor, M. J. (2003). Early processing of the six basic facial emotional expressions. Cogn. Brain Res. 17, 613–620. doi: 10.1016/S0926-6410(03)00174-5
Crossref Full Text | Google Scholar
Berg, P., and Scherg, M. (1994). A multiple source approach to the correction of eye artifacts. Electroencephalogr. Clin. Neurophysiol. 90, 229–241. doi: 10.1016/0013-4694(94)90094-9
PubMed Abstract | Crossref Full Text | Google Scholar
Blau, V. C., Maurer, U., Tottenham, N., and McCandliss, B. D. (2007). The face-specific N170 component is modulated by emotional facial expression. BBF 3:7. doi: 10.1186/1744-9081-3-7
PubMed Abstract | Crossref Full Text | Google Scholar
Caldwell, J. A., Prazinko, B., and Caldwell, J. L. (2003). Body posture affects electroencephalographic activity and psychomotor vigilance task performance in sleep-deprived subjects. Clin. Neurophysiol. 114, 23–31. doi: 10.1016/S1388-2457(02)00283-3
PubMed Abstract | Crossref Full Text | Google Scholar
Chai, H., Chen, W. Z., Zhu, J., Xu, Y., Lou, L., Yang, T., et al. (2012). Processing of facial expressions of emotions in healthy volunteers: an exploration with event-related potentials and personality traits. Neurophysiol. Clin. 42, 369–375. doi: 10.1016/j.neucli.2012.04.087
PubMed Abstract | Crossref Full Text | Google Scholar
Chang, L. J., Lin, J. F., Lin, C. F., Wu, K. T., Wang, Y. M., and Kuo, C. D. (2011). Effect of body position on bilateral EEG alterations and their relationship with autonomic nervous modulation in normal subjects. Neurosci. Lett. 490, 96–100. doi: 10.1016/j.neulet.2010.12.034
PubMed Abstract | Crossref Full Text | Google Scholar
Clark, V. P., and Hillyard, S. A. (1996). Spatial selective attention affects early extrastriate but not striate components of the visual evoked potential. J. Cogn. Neurosci. 8, 387–402. doi: 10.1162/jocn.1996.8.5.387
PubMed Abstract | Crossref Full Text | Google Scholar
Delorme, A., and Makeig, S. (2003). EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 134, 9–21. doi: 10.1016/j.jneumeth.2003.10.009
PubMed Abstract | Crossref Full Text | Google Scholar
Delorme, A., Mullen, T., Kothe, C., Akalin Acar, Z., Bigdely-Shamlo, N., Vankov, A., et al. (2011). EEGLAB, SIFT, NFT, BCILAB, and ERICA: new tools for advanced EEG processing. Comput. Intell. Neurosci. 2011, 1–12. doi: 10.1155/2011/130714
留言 (0)