SciELO - Scientific Electronic Library Online

 
vol.29 issue1Social position and obesity in Chilean population: Psychological mediators and moderatorsPsychological aspects of soccer and futsal players: A systematic review author indexsubject indexarticles search
Home Pagealphabetic serial listing  

Services on Demand

Journal

Article

Indicators

Related links

  • On index processCited by Google
  • Have no similar articlesSimilars in SciELO
  • On index processSimilars in Google

Share


Suma Psicológica

Print version ISSN 0121-4381

Suma Psicol. vol.29 no.1 Bogotá Jan./June 2022  Epub Jan 20, 2022

https://doi.org/10.14349/sumapsi.2022.v29.n1.3 

ARTICLES

Silent speech of vowels in persons of different cognitive styles

Habla silenciosa de las vocales en personas de diferente estilo cognitivo

Omar López-Vargas1  * 
http://orcid.org/0000-0003-2298-6854

Luis Sarmiento-Vela1 
http://orcid.org/0000-0001-7928-4621

Jan Bacca-Rodríguez2 
http://orcid.org/0000-0003-2353-6994

Sergio Villamizar-Delgado2 

Jhon Sarmiento-Vela1 

1 Laboratory of Cognitive Science and Artificial Intelligence, School of Science and Technology, Universidad Pedagógica Nacional, Bogotá, Colombia

2 Department of Electrical and Electronics Engineering, School of Engineering, Universidad Nacional de Colombia, Bogotá, Colombia.


Abstract

Introduction:

This research measures the differences in silent speech of the vowels / a / - / u / in Spanish, in students with different cognitive styles in the Field Dependence - Independence (FDI) dimension.

Method:

Fifty-one (51) adults participated in the study. Electroencephalographic (EEG) signals were taken from 14 electrodes placed on the scalp in the language region located in the left hemisphere. Previously, the embedded figures test (EFT) was applied in order to classify them into dependent, intermediate and field independent persons. To analyse the EEG data, the signals were decomposed into intrinsic mode functions (IMF) and a mixed repeated measures analysis was performed.

Results:

It was found that the Power Spectral Density (PSD) in the vowels is independent of the cognitive style and its magnitude depends on the position of the electrodes.

Conclusions:

The results suggest that there are no significant differences in PSDs in the silent speech of vowels /a/-/u/ in persons of different cognitive styles. Significant differences were found in the PSDs according to the position of the 14 electrodes used. In our configuration, the silent speech of vowels can be studied using electrodes placed in premotor, motor and Wernicke areas.

Keywords: Voluntary signals; silent speech; cognitive style; EEG; vowels

Resumen

Introducción:

La investigación mide las diferencias en el habla silenciosa de las vocales /a/-/u/ en español, en estudiantes de diferente estilo cognitivo en la dimensión Dependencia - Independencia de campo (DIC).

Método:

En el estudio participaron 51 adultos. Se tomaron señales electroencefalográficas (EEG), a partir de 14 electrodos dispuestos sobre el cuero cabelludo de la región del lenguaje ubicada en el hemisferio izquierdo. Previamente les fue aplicado el test de figuras enmascaradas EFT con el fin de clasificarlos en personas dependientes, intermedios e independientes de campo. Para analizar los datos del EEG se descompusieron las señales en funciones de modo intrínseco (IMF) y se realizó un análisis mixto de medidas repetidas.

Resultados:

Se halló que la densidad espectral de potencia (PSD) en las vocales es independiente del estilo cognitivo y su magnitud depende de la posición de los electrodos.

Conclusión:

Los resultados sugieren que no existen diferencias significativas en los PSD en el habla silenciosa de las vocales /a/-/u/ en las personas de diferente estilo cognitivo. Se hallaron diferencias significativas en los PSD de acuerdo con la posición de los 14 electrodos utilizados. En nuestra configuración, el habla silenciosa de las vocales puede ser estudiada mediante electrodos situados en las áreas premotora, motora y de Wernicke.

Palabras clave: Señales voluntarias; habla silenciosa; estilo cognitivo; EEG; vocales

To study silent speech by means of electroencephalographic (EEG) signals, voluntary signals are used, which are produced autonomously by the person, a method that is a signal analysis alternative and is currently explored within the Brain-computer Interface (BCI) research area. EEG signals are easy to record. They have a high temporal resolution and are obtained non-invasively. Studies show that silent speech tasks, such as thinking or imagining vowels, syllables, or words mainly involve the Supplementary Motor Area (SMA) (DaSalla et al., 2009; Iqbal et al., 2015; Nguyen et al., 2018; Qureshi et al., 2018) and the Language Areas (Broca and Wernicke areas) located in the brain’s left hemisphere (Ikeda et al., 2014; Morooka et al., 2018; Sarmiento et al., 2021; Villamizar et al., 2021).

In this line of study, researchers seek to design and develop communication systems using BCIs, to favour the quality of life of persons with some type of language disability (Callan et al., 2000; D’Zmura et al., 2009; DaSalla et al., 2009; Fujimaki et al., 1994; Lee et al., 2021; Morooka et al., 2018; Qureshi et al., 2018). Thus, the use of EEG to study silent speech has gained momentum in recent years.

Research shows promising results in understanding and explaining vowel, syllable, and word decoding, among others (Cooney et al., 2018; González-Castañeda et al., 2017; Qureshi et al., 2018; Sarmiento et al., 2014; Yoshimura et al., 2016). Some studies take into account the potentials evoked in imagined silent speech tasks, which consist of internally simulating a motor movement without moving the body in any way (Fujimaki et al., 1994). In the case of speech imagery, the person has to imagine moving parts of the speech apparatus such as: tongue, lips, or lower jaw (Graimann et al., 2010). This task requires high concentration and training time from the person, therefore, fatigue in this type of tasks is frequent. In this line of work, evoked potentials are used for the processing of EEG signals, which include: Event Related Potential (ERP), Evoked Potential (P300), Movement Related Cortical Potential (MRCP), and Steady State Evoked Potentials (SSEP) (Rashid et al., 2020).

According to Martín et al. (2014), silent speech, imagined speech, covert speech, and inner speech are used in the same way, when a person thinks of a vowel, syllable, or word without the intentional movement of speech apparatus such as the lips or tongue. In this article we preferably use silent speech (Martin et al., 2014). In silent speech, EEG is used to decode the English vowels /a/ and /u/ (DaSalla et al., 2009; Iqbal et al., 2015), /a/, /i/, and /u/ (Nguyen et al., 2018) and /a/, /e/, /i/, /o/, /u/ (Ghosh et al., 2019) ; Korean vowels /a/, /e/, /i/, /o/, and /u/ (Min et al., 2016), syllables /ba/ and /ku/ (D’Zmura et al., 2009); words /go/, /back/, / left/, /right/, and /stop/ (Qureshi et al., 2018).

In contrast to this, other studies take into account the processing of voluntary signals for the study of silent speech, which is generated in the language area (Broca & Wernicke) (Ikeda et al., 2014; Morooka et al., 2018; Villamizar et al., 2021). Imagined speech is considered as the internal pronunciation of phonemes, words, or sentences, regardless of the movement of the phonatory apparatus and without any audible output (Cooney et al., 2020). Thinking of silent speech with electroencephalographic signals is characterised by the fact that persons do not require specific training processes and tasks are almost always performed using the native language. Therefore, it does not require high levels of training and attention which fatigue a person. (Fujimaki et al., 1994; Graimann et al., 2010). In this line of research, EEG is used to decode Spanish vowels /a/, /e/, /i/, /o/, /u/ (Sarmiento et al., 2014); Japanese vowels /a/, /i/ (Yoshimura et al., 2016). and /a/, /i/, /u/ (Ikeda et al., 2014). In the present research, this brain signal was selected because the person can generate this type of signal at will, unlike evoked potentials such as P300 or SSVEP, which depend on external stimuli.

Most studies developed based on EEG and BCIs, to identify silent speech processing, conventionally use different groups of brain rhythms in the frequency domain such as: delta, theta, alpha, beta, and gamma with linear and stationary characteristics, which require algorithms to identify based on brain signals (DaSalla et al., 2009; Matsumoto & Hori, 2014; Riaz et al., 2015; Sarmiento et al., 2014). However, in this line of research, recent studies focus on the use of methods which consider non-linear and non-stationary signals; inherent to electroencephalographic signals.

To study these signals, the following methods are used: empirical mode decomposition (EMD) (Hansen et al., 2019), multivariate empirical mode decomposition (MEMD), and recently, the method referred to as adaptive-projection intrinsically transformed MEMD (APIT-MEMD) (Hemakom et al., 2016; Villamizar et al., 2021; Yuan et al., 2018). Despite the studies conducted, there is still no consensus among the academic community regarding which brain rhythms are the most appropriate for identifying and processing silent speech or which signal processing methods are the most effective (Fujimaki et al., 1994; Morooka et al., 2018; Qureshi et al., 2018).

A clear understanding and explanation of the fundamental processes of silent speech processing has not yet been achieved. It is likely necessary to consider some of the differential characteristics of the subjects when they process information, which may be associated with cognitive style, and in this regard, studies show that there are individual differences in cognitive information processing (Evans et al., 2013; López-Vargas et al., 2020; Solórzano-Restrepo & López-Vargas, 2019).

Specifically, it is noteworthy, that probably, the most studied cognitive style is the one referred to as Field Dependence/Independence (FDI), proposed and developed by Witkin and his colleagues (Witkin, H. A. et al., 1977). FDI establishes a difference between persons with a tendency towards an analytical-type processing, regardless of contextual factors (field independent persons (FI), and those with a tendency to a global-type processing, highly influenced by the context (field dependent persons (FD).

FI persons follow an analytical information trend, fact which allows them to break down the information into its different components and restructure it according to their needs. In addition, they have strategies to organise, classify, and store information, and they resort to different clues in order to retrieve it later. For their part, FD persons are more sensitive to external signals and tend to receive the information as it is presented to them. In other words, they prefer externally structured information and address its global aspects (López et al., 2012; Valencia-Vallejo et al., 2019).

Within this area of research, electroencephalography is promising. For example, using EEG, it was found that FI persons exhibit less coherence between hemispheres, which indicates a greater hemispheric specialisation (Oltman et al., 1979; Zoccolotti, 1982). Also, when developing visual tasks and exercises related to auditory and somatosensory aspects, it was determined that FI persons show a greater neuronal activity for executive and inhibitory response processing. These studies support the idea that FI persons possess better inhibitory control (Imanaka et al., 2017; Jia et al., 2014).

Along these lines, few studies inquire into the possible relationships that may exist between the processing of voluntary signals in imagined speech and a persons’ stylistic characteristics when performing silent speech tasks using vowels. Understanding and explaining a persons’ individual differences, within the study of silent speech, may aid in the design of BCIs that favour the quality of life of people with some type of language disability. Consistent with these approaches, this study hereby poses the following research question:

Are there any significant differences in silent speech production using the vowels /a/-/u/ between persons with different cognitive styles in the Field Dependence/Independence (FDI) dimension?

This research proposes considering a subject’s cognitive style in the FDI dimension and using intrinsic mode functions (IMF) produced by APIT-MEMD to choose the signals related to silent speech using vowels. Also, the combination of the APIT- MEMD method with power spectral density (PSD) is used to analyse the data captured from EEG signals. The APIT-MEMD method allows for multivariate separation, over time, of brain signals that have non-linear and non-stationary characteristics.

In addition, the aim is to identify the best location for the electrode that enables generating higher energy levels in the PSDs of 14 electrodes placed on a neuroheadset and arranged in a matrix especially designed for the language region in silent speech tasks using vowels.

In this order of ideas, the objective of the research is based on determining whether the cognitive style in the FDI dimension of people, when performing silent vowel speech tasks, generates significant differences in the energy levels of the PSDs and also, to establish precisely the areas of the language region where these energy levels are maximal in order to take them into account when designing BCIs.

Method

Participants

This study involved 51 undergraduate and postgraduate students (21 women and 30 men), one group enrolled in a university in the city of Bogotá-Colombia and another group enrolled in a university in the city of Popayán-Colombia. Ages ranged from 18 to 41 years (M = 24.76, SD = 7.66). Participants did not exhibit any type of medical or neurological problem. All persons involved gave their written informed consent and the experiment was approved by the Ethics Committee of the Universidad Nacional de Colombia.

Experimental Protocol

The experiment was carried out in the Laboratory of Cognition and Intelligent Systems of a public university in the city of Bogota-Colombia, under controlled lighting conditions of 80 lm/m2 and minimum ambient sound (ASTM STC 63). First, the subjects were given an embedded figures test (EFT) to determine their cognitive style in the FDI dimension. Subsequently, each of them was asked to sit in a comfortable chair and was fitted with an EEG neuroheadset, and the 14 electrodes were placed on the scalp located over the Broca, Wernicke, and motor areas in the left hemisphere. Two reference electrodes were also placed on the frontal region. An abrasive gel was used to clean the scalp before placing each electrode.

The placement of the electrodes was done according to the neurological models on language by: Geschwind (1965) and Poeppel and Hickok (2004). To reference the neuroheadset on each subject’s head, the T3 and C3 positions were used according to the 10-20 system (Figure 1). Finally, a light source was placed at a distance of one meter from the person to signal the time to begin the task of thinking of a specific vowel with silent speech, and also, to end the activity. During the experimental phase, persons were asked to keep their eyes closed to reduce artifacts, such as blinking and eye movement, while developing the task.

Figure 1 Location of the neuroheadset, which contains 14 electrodes covering the left hemisphere (language, premotor, and motor areas) plus two reference electrodes on the frontal area. 

Signal Acquisition

Each person was told that as long as the light source was turned on, they were to think about the corresponding vowel continuously and with silent speech. They were also told that when the light source was turned off, they had to stop thinking about said vowel and go into a state of bodily relaxation. During the experiment, the light source was kept activated for four seconds and then, it was turned off for three seconds. The procedure was repeated 25 times for each of the two vowels. Between each silent speech task, the subjects rested for 5 minutes before proceeding to the next vowel change. The vowels were arranged in the following order: first /a/; and second /u/. Although some authors suggest that data collection in EEG signals should be taken randomly (Li et al., 2018), other researchers suggest performing the data collection of vowels or syllables with EEG signals in an orderly and deterministic way (Cooney et al., 2020; Pressel Coretto et al., 2017). The sampling intervals in silent speech were (385-896, 1281-1792, ...) (Figure 2).

Figure 2 Time intervals for the experimental process. During the intervals (385-896, 1281-1792, ...), the subject thought silently about the corresponding vowel. During the intervals (1-384, 897-1280, ...), the person was in a state of relaxation. 

The signals were recorded with a 14-channel EMOTIV EPOC+ amplifier, with a sampling frequency of 128 Hz, a 14-bit resolution with 1 LSB with 0.51 m V in monopolar configuration. The 14 electrodes of the EMOTIV EPOC+ device were arranged on the neuroheadset (E1, ..., E14) considering the name of each electrode of the EMOTIV EPOC+ device. For this experiment, the electrodes are numbered from E1 to E14 and the relationship with the name of the Emotiv electrodes is as follows: E1 (AF3), E2 (F7), E3 (F3), E4 (FC5), E5 (T7), E6 (P7), E7 (O1), E8 (O2), E9 (P8), E10 (T8), E11 (FC6), E12 (F4), E13 (F8), E14 (AF4). The two reference electrodes were placed on the subject’s forehead (Sarmiento et al., 2021; Villamizar et al., 2021) (Figure 1).

To export the data in MATLAB’s Simulink, Xcessity EpocSimulinkImporter acquisition software was used. Brain signal processing was performed with MATLAB R2016 software (The MathWorks, Inc., Natick, MA). Subsequently, data analysis was carried out using the Statistical Package for the Social Sciences (SPSS) Version 25 software.

To filter the EEG signals, the APIT-MEMD method has been selected, which is particularised by allowing the separation of multivariate, non-linear and non-stationary signals, into components called IMFs (Hemakom et al., 2016). The central concept is based on assuming that the data or signals are composed of simple intrinsic modes of oscillation that are characterised by having the same number of extremes and the same number of zero-crossings. These intrinsic modes of oscillation are called intrinsic modes of function (IMF). The APIT-MEMD algorithm has got the following steps: First, it calculates the covariance matrix and the eigenvalues of the multivariate signal; Second, it selects the first principal component related to the highest eigenvalues; Third, it develops a uniform Hammerseley sequence over an n-sphere; Later, it calculates the projection of the vectors in order to calculate new mean enveloping curves; Finally, with an iterative process, the designated MFIs are found (Hemakom et al., 2016).

For this research, brain signals were processed with the APIT-MEMD method, where p = 14 electrodes (E1, … E14) located over the left hemisphere. The results of this algorithm are multivariate IMFs related to p electrodes, v vowels with silent speech, and m multivariate levels.

In other words, power spectral density (PSD) is a statistical measure that quantifies the power of a signal based on a finite group of data. One outstanding application of PSD is the detection of signals immersed in noise (Sarmiento et al., 2014). Subsequently, the power spectral density (PSD) of each brain signal was determined, for which the periodogram (X p s, v (f)) was used, which determines the energy levels of each IMF using equation 1 (Li & Wong, 2013; Proakis & Manolakis, 2007).

Based on the above equation, S is the person, v is the vowel, p is the electrode, A is the width of the window, IMF is the multivariate function resulting from the APIT-MEMD, m is an IMF level, and n are the samples to be analysed. In this case, the number of IMFs analysed was 10 (Figure 3).

Figure 3 Shows the first 6 IMFs and their respective frequencies for the E3 electrode, with a sample of 1 second, for a FD subject (blue color), an INT subject (green color) and a FI subject (red color), in speech imagery tasks. The first column corresponds to the IMFs of the silent speech /a/. The second column corresponds to the Fourier transform of the previous IMFs. The third column corresponds to the IMFs of the silent speech /u/, and the fourth column corresponds to the IMFs Fourier transform of the silent speech /u/. In the Fourier transform for IMF 4 of /a/ and /u/, delta and theta waves are presented among 2 to 8 Hz. 

Subsequently, through a mixed repeated measures analysis, each IMF was analysed to determine the brain rhythms most related to silent speech and it was found that the multivariate IMF with the greatest significant differences corresponded to level 4. Finally, the PSD data were averaged among the participants to obtain the overall PSD average and thus proceed with the mixed repeated measures analysis (Figure 3).

Cognitive Style Test

The test used to determine cognitive style in the FDI dimension was Witkin’s Embedded Figures Test (GEFT). This perception test requires the person to locate a simple, previously seen, figure within another figure with a complex design. The test contains 18 complex figures. The task is to find the simple figures in a given amount time. The test lasts approximately 20 minutes. A version of the online instrument has been given to Colombian students (Hederich-Martínez et al., 2016). The test showed a Cronbach’s alpha of 0.847.

Results

The samples’ EFT average was 10.59; the standard deviation (SD = 3.656). Out of a maximum score of 18; the minimum value was 1 and the maximum value was 18 points. The subjects were grouped into FD, intermediate (INT), and FI, defining terciles for the total test score. Thus, three ranges of scores were identified, namely: (a) 17 FD persons (first tercile), (b) 21 INT persons (second tercile), and (c) 13 FI persons (third tercile).

A mixed repeated measures ANOVA analysis was used. The two intra-subject variables are: (1) Thinking of vowels with two values; /a/ - /u/ and (2) PSD recorded by EEG from 14 electrodes (E1, E2, …E14). The inter-subject variable was the cognitive style with three values; field dependent, intermediate, and independent. Table 1 shows a summary of the descriptive statistics of the PSDs recorded for each vowel from the 14 electrodes, considering cognitive style.

Table 1 Silent Speech PSD results for vowels /a/ - /u/ from each electrode: Mean scores and standard deviations in parentheses 

Mauchly’s test indicated that the sphericity assumption was not met. The data show that the main effect of PSDs on electrodes is: (X2(90) = 847.91, p < 0.05). Therefore, the degrees of freedom were corrected with Greenhouse-Geisser (ε = .17). The multivariate tests indicate that there are significant differences in silent speech using vowels /a/ - /u/ (Pillai’s Trace = 0.190, F (1,48) = 11.27, p = 0.002, η2 = 0.190) and electrode PSDs (Pillai’s Trace = 0.553, F (13,36) = 3.42, p = 0.002, η2 = 0.553). However, there are no significant double interactions between the intra-subject variables and between the intra-subject and inter-subject variables, suggesting that the PSDs recorded from the 14 electrodes depend on the silent speech of the vowels (Figure 4). In figure 4, estimated marginal means is presented for vowels /a/ and /u/. This figure shows that there are significant differences for the 14 EEG electrodes. Figure 5 and 6 shows a topographic map with the 14 electrodes, for a subject and a sample of silent speech vowels /a/ and /u/. Colors between 0 (blue) and 0.4 (yellow) indicate the value of PSD. (Figure 5 and 6).

Figure 4 PSD estimated marginal means for vowels /a/ - /u/ from each electrode 

Figure 5 Topographic map of PSD overall average, done with cubic splines, for all persons during silent speech tasks for the vowel /a/. In the figure, the yellow color’s intensity represents the highest PSD levels (μV2/Hz) and the blue color’s intensity represents the lowest PSD values. The location of the electrodes is denoted by E1, E2, ..., E14 

Figure 6 Topographic map of the PSD overall average, made with cubic splines for all persons during silent speech tasks for the vowel /u/. In the figure, the yellow color’s intensity represents the highest PSD levels (μV2/Hz) and the blue color’s intensity represents the lowest PSD values. The location of the electrodes is denoted by E1, E2, …, E14 

With regard to intra-subject contrast tests for the vowel variable /a/ - /u/, the data show that the average PSD values for vowel /a/ (M = 0.39) differ from the average PSD va-lues for vowel /u/ (M = 0.277). This contrast is significant (F (1,48) = 11.27, p = 0.002, η 2 = 0.190). Also, the Electrode variable shows significant differences (F (2.16,103.43) = 4.74, p = 0.009, η 2 = 0.090). For the Electrode variable, there are significant differences in the following contrasts: The first contrast (Level 2 vs. Level 3) is the average of the PSDs generated in electrode E2 (M = 0.35) with the PSD of electrode E3 (M = 0.32). This contrast is significant (F (1,48) = 13.92, p = 0.001, η 2 = 0.225). The second contrast (Level 4 vs. Level 5) shows an average PSD from electrode E4 (M = 0.32) with the PSD of electrode E5 (M = 0.34). This contrast was also significant (F (1,48) = 8.22, p =0.006, η2 = 0.146). The third contrast (Level 5 vs. Level 6) indicates an average PSD from electrode E5 (M = 0.34) with the PSD of electrode E6 (M = 0.31).

This contrast was significant (F (1,48) = 20.15, p <0.001, η 2 = 0.296). A fourth contrast (Level 9 vs. Level 10) indicates an average PSD from electrode E9 (M = 0.35) with the PSD of electrode E10 (M = 0.32). This contrast was also significant (F (1,48) = 36.29, p < 0.001, η 2 = 0.431). The fifth contrast (Level 10 vs. Level 11) indicates an average PSD from electrode E10 (M = 0.32) with the PSD of electrode E11 (M = 0.34). This contrast shows significant differences (F (1,48) = 11.58, p = 0.001, η 2 = 0.194). Finally, a sixth contrast (Level 12 vs. Level 13) indicates an average PSD from electrode E12 (M = 0.35) with the PSD of electrode E13 (M = 0.33). This contrast indicates significant differences (F (1,48) = 9.49, p = 0.003, η2 = 0.165).

Multiple comparisons according to Bonferroni indicate statistically significant differences (p < .05) between the PSDs of the following electrodes: (a) The E2 (M = 0.35) electrode is significantly larger than E3 (M = 0.32), E4 (M = 0.32), and E6 (M = 0.31), (b) E6 (M = 0.31) electrode, is significantly lower than E5 (M = 0.34), E8 (M = 0.35), E9 (M = 0.35), and E12(M = 0.35). (c) E7 (M = 0 .32) electrode, is significantly lower than E8 (M = 0.35), E9 (M = 0.35), and E12(M = 0.35), and finally, (d) E10 (M = 0.32) electrode, is significantly lower than E8 (M = 0.35), E9 (M = 0.35), and E12(M = 0.35).

Regarding the test of inter-subject effects, the data show that there are no significant differences in silent speech using vowels /a/ - /u/, in the PSDs generated from the different electrodes due to the subjects’ cognitive styles (F (2,48) = 0.75, p = 0.477, η2 = 0.030) (Figure 7).

Figure 7 Effect of cognitive style on PSD estimated marginal means from each electrode 

Discussion

The results show that PSDs in silent speech of the vowels /a/ and /u/ are not associated with a person’s cognitive style in the FDI dimension. The study’s analyses indicate that there are no significant differences between field independent, intermediate and dependent persons. Silent speech tasks using vowels /a/ - /u/, with eyes closed, likely differ from completing visual tasks and exercises related to somatosensory aspects, where FI persons were found to show increased neuronal activity for executive and inhibitory response processing (Imanaka et al., 2017; Jia et al., 2014).

The characteristics of a silent speech task with eyes closed, so as to reduce artifacts such as blinking and eye movement, ostensibly allowed field dependent persons to exhibit a greater attentional control. This preventing irrelevant or distracting aspects from arising within their perceptual field, which could alter their attention on thinking of silent speech using vowels /a/ - /u/, while completing the task. In this regard, field dependent, intermediate, and independent persons would have the same capacity for attention. Regarding field intermediate persons, it is noteworthy, that their relative functional mobility allows them to get a little closer to FI persons. However, it is not possible to assert that the former are under the same condition as the latter; but it is possible to assert that the former are closer to the level reached by the latter, as reported in other studies (Evans et al., 2013).

In this area of research, there are previous studies that addressed the classification of EEG signals through imagined speech, where a persons’ task was to imagine vocalising and pronouncing the vowels (DaSalla et al., 2009; Iqbal et al., 2015; Min et al., 2016; Nguyen et al., 2018). In these investigations, the signals were studied in the time domain, where advanced decoding algorithms are designed and validated to classify them. The present study is based on thinking of silent speech using vowels /a/ - /u/, where the object of study is the neuronal activity that takes place in the language region of the left hemisphere.

In this line of research, some studies have focused on the development of specialised algorithms for the classification of vowels (Sarmiento et al., 2014). Another study compared functional magnetic resonance imaging (fMRI) and EEG signals to identify the locations on the brain related to vowel production in silent speech (Yoshimura et al., 2016). Also, in the work by Ikeda et al. (2014), it was found that the study of Japanese vowels /a/, /i/, /u/, is related to the premotor, lower frontal gyrus, upper temporal gyrus, and motor areas. This research supplements these results insofar as the areas corresponding to electrodes 2,5,8, 9, and 12 are related to the areas found by Ikeda and collaborators.

For several years, motor imagery or movement processes have been the object of study by different researchers. In these studies, regions of the motor cortex (pre-central gyrus) and/or the somatosensory cortex (post-central gyrus) are activated, finding that motor imagination tasks are related to mu and Gamma waves (D’Zmura et al., 2009; DaSalla et al., 2009; Iqbal et al., 2015).

Within the studies of silent speech with non-invasive methods, it has been chosen to use delta, theta, alpha, beta, gamma and high gamma brain rhythms. For vowel recognition with silent speech, the following rhythms have tended to be used: theta, alpha and beta (Chi et al., 2011); delta, theta and alpha (Sarmiento et al., 2014); alpha, beta and gamma (Riaz et al., 2015); delta, theta, alpha, beta and gamma (Matsumoto & Hori, 2014) and delta, theta, alpha, beta, gamma and high gamma (Matsumoto & Hori, 2013). In the case of syllable recognition with silent speech, the following rhythms have been used: delta, theta and alpha (D’Zmura et al., 2009); theta, alpha and beta (Wang et al., 2013) and high gamma (Jahangiri & Sepulveda, 2019). In this sense, more research is needed to determine the frequency ranges for the different elements of language in speech imagery.

In contrast, this study found that the Delta and Theta waves related to IMF 4 are the most suitable rhythms for the study of silent speech using vowels /a/ - /u/ (Figure 3). These results require further study to be able to assert that these brain rhythms are the most relevant. In this vein, the development of research on the language region is a promising line of study to understand and explain EEG signals in thinking of silent speech tasks, whether using vowels, syllables, or words.

On the other hand, the study’s data show that multivariate IMFs, resulting from applying the APIT-MEMD method (Hemakom et al., 2016), evidence greater significant differences in different persons’ IMF 4, obtaining the best results in the PSD of this IMF. However, these results are not conclusive and, therefore, other studies on silent speech using different language units, such as syllables, consonants, or words, are required to corroborate that delta and theta brain rhythms are generalisable to other language units. In this regard, brain signal processing by means of the APIT-MEMD method together with the PSD, is a potential analysis alternative for identifying the signals related to thinking of silent speech.

The method used in this research considers non-linearity, non-stationarity, and existing correlations between the voluntary signals, which are more consistent with the nature of brain signals (Hemakom et al., 2016). By using these models, better results may likely be achieved when identifying and interpreting different brain processes. However, further in-depth studies are needed to contrast these results with methods based on linear and/or stationary systems, where different filters and transformations are used to identify brain rhythms (Delta, Theta, Alpha, Beta, and Gamma) while completing thinking of silent speech tasks.

Finally, when observing the different levels of PSDs from the 14 electrodes used, it is possible to identify that electrodes E2, E5, E8, E9, and E12 located in the premotor, motor, and Wernicke areas, exhibit significant differences that allow identifying vowels /a/ - /u/. Based on these results, it is possible to assert that identifying vowels with silent speech may be studied through the use of electrodes located in the premotor, motor, and Wernicke areas. These results are consistent with the spatial positions proposed in (Geschwind, 1965; Poeppel & Hickok, 2004) models. Hence, future studies on silent speech tasks should focus the spatial location of electrodes on the areas proposed above in order to enable the design of BCIs that relate specific tasks to the thinking of a specific vowel, consonant, or word.

Conclusions

It is possible to conclude that there are no significant differences in the PSDs, in the silent speech of vowels /a/ and /u/, due to the effect of the cognitive style in the FDI dimension of the participants. This result is promising in the design of BCIs, since it is an initial approach in establishing that people’s stylistic characteristics are not determining factors in the design and implementation of technological devices.

Also, the study shows that the APIT-MEMD method, together with the PSD, is a promising alternative for identifying voluntary signal processing related to the thinking of silent speech. Finally, to study silent speech using vowels /a/ - /u/, it is possible to suggest positioning the electrodes in the premotor, motor, and Wernicke’s areas.

Acknowledgements

The authors are grateful for the financial support to develop this research from the Administrative Department of Science, Technology and Innovation -MINCIENCIAS- Colombia, in accordance with funding contract RC 838/2017, code: 110177758402.

References

Callan, D. E., Callan, A. M., Honda, K., & Masaki, S. (2000). Single-sweep EEG analysis of neural processes underlying perception and production of vowels. Cognitive Brain Research, 10(1-2), 173-176. https://doi.org/10.1016/s0926-6410(00)00025-2Links ]

Chi, X., Hagedorn, J. B., Schoonover, D., & Zmura, M. D. (2011). EEG-Based discrimination of imagined speech phonemes. International Journal of Bioelectromagnetism, 13(4), 201-206. https://pdfs.semanticscholar.org/b74f/c325556d1a7b5eb05fe90cde1f0c891357a3.pdfLinks ]

Cooney, C., Folli, R., & Coyle, D. (2018). Neurolinguistics research advancing development of a direct-speech brain-computer interface. IScience, 8, 103-125. https://doi.org/10.1016/j. isci.2018.09.016Links ]

Cooney, C., Korik, A., Folli, R., & Coyle, D. (2020). Evaluation of hyperparameter optimization in machine and deep learning methods for decoding imagined speech EEG. Sensors (Switzerland), 20(16), 1-22. https://doi.org/10.3390/s20164629Links ]

D’Zmura, M., Deng, S., Lappas, T., Thorpe, S., & Srinivasan, R. (2009). Toward EEG sensing of imagined speech. Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioin-formatics), 5610 LNCS(PART 1 ), 40-48. https://doi.org/10.1007/978-3-642-02574-7_5Links ]

DaSalla, C. S., Kambara, H., Sato, M., & Koike, Y. (2009). Single-trial classification of vowel speech imagery using common spatial patterns. Neural Networks, 22(9), 1334-1339. https://doi.org/10.1016/j.neunet.2009.05.008Links ]

Evans, C., Richardson, J. T. E., & Waring, M. (2013). Field independence: Reviewing the evidence. British Journal of Educational Psychology, 83(2), 210-224. https://doi.org/10.1111/bjep.12015Links ]

Fujimaki, N., Takeuchi, F., Kobayashi, T., Kuriki, S., & Hasuo, S. (1994). Event-related potentials in silent speech. Brain Topography, 6(4), 259-267. https://doi.org/10.1007/BF01211171Links ]

Geschwind, N. (1965). Disconnection syndromes in animals and man. Brain, 88(2), 237-294. [ Links ]

Ghosh, R., Sinha, N., Biswas, S. K., & Phadikar, S. (2019). A modified grey wolf optimization based feature selection method from EEG for silent speech classification. Journal of Information and Optimization Sciences, 40(8), 1639-1652. https://doi.org/10.1080/02522667.2019.1703262Links ]

González-Castañeda, E. F., Torres-García, A. A., Reyes-García, C. A., & Villaseñor-Pineda, L. (2017). Sonification and textification: Proposing methods for classifying unspoken words from EEG signals. Biomedical Signal Processing and Control, 37, 82-91. https://doi.org/10.1016/j.bspc.2016.10.012Links ]

Graimann, B., Allison, B. Z., & Pfurtscheller, G. (2010). Brain-Computer interfaces: Revolutionizing human-computer interaction (1st ed.). Springer Verlag. [ Links ]

Hansen, S. T., Hemakom, A., Gylling Safeldt, M., Krohne, L. K., Madsen, K. H., Siebner, H. R., Mandic, D. P., & Hansen, L. K. (2019). Unmixing oscillatory brain activity by EEG source localization and empirical mode decomposition. Computational Intelligence and Neuroscience, 2019. https://doi.org/10.1155/2019/5618303Links ]

Hederich-Martínez, C., López-Vargas, O., & Camargo-Uribe, A. (2016). Effects of the use of a flexible metacognitive scaffolding on self-regulated learning during virtual education. International Journal of Technology Enhanced Learning, 8(3/4), 1. https://doi.org/10.1504/ijtel.2016.10002201Links ]

Hemakom, A., Goverdovsky, V., Looney, D., & Mandic, D. P. (2016). Adaptive-projection intrinsically transformed multivariate empirical mode decomposition in cooperative brain-computer interface applications. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 374(2065). https://doi.org/10.1098/rsta.2015.0199Links ]

Ikeda, S., Shibata, T., Nakano, N., Okada, R., Tsuyuguchi, N., Ikeda, K., & Kato, A. (2014). Neural decoding of single vowels during covert articulation using electrocorticography. Frontiers in Human Neuroscience, 8(MAR), 1-8. https://doi.org/10.3389/fnhum.2014.00125Links ]

Imanaka, M., Kakigi, R., & Nakata, H. (2017). The relationship between cognitive style and event-related potentials during auditory and somatosensory Go/No-go paradigms. NeuroReport, 28(13), 822-827. https://doi.org/10.1097/WNR.0000000000000833Links ]

Iqbal, S., Khan, Y. U., & Farooq, O. (2015). EEG based classification of imagined vowel sounds. 2015 International Conference on Computing for Sustainable Global Development, INDIA Com 2015, 1591-1594. [ Links ]

Jahangiri, A., & Sepulveda, F. (2019). Correction to: The Relative Contribution of High-Gamma Linguistic Processing Stages of Word Production, and Motor Imagery of Articulation in Class Separability of Covert Speech Tasks in EEG Data. Journal of Medical Systems, 43(8), 237. https://doi.org/10.1007/s10916-019-1379-1Links ]

Jia, S., Zhang, Q., & Li, S. (2014). Field dependence-independence modulates the efficiency of filtering out irrelevant information in a visual working memory task. Neuroscience, 278, 136-143. https://doi.org/10.1016/j.neuroscience.2014.07.075Links ]

Lee, W., Seong, J. J., Ozlu, B., Shim, B. S., Marakhimov, A., & Lee, S. (2021). Biosignal sensors and deep learning-based speech recognition: A review. Sensors (Switzerland), 21(4), 1-22. https://doi.org/10.3390/s21041399Links ]

Li, R., Johansen, J. S., Ahmed, H., Ilyevsky, T. V., Wilbur, R. B., Bharadwaj, H. M., & Siskind, J. M. (2018). Training on the test set? An analysis of Spampinato et al . ( 31). 1-18. http://arxiv.org/abs/1812.07697Links ]

Li, Y., & Wong, K. M. (2013). Riemannian distances for signal classification by power spectral density. IEEE Journal on Selected Topics in Signal Processing 7(4), 655-669. https://doi.org/10.1109/JSTSP.2013.2260320Links ]

López-Vargas, O., Ortiz-Vásquez, J., & Ibáñez-Ibáñez, J. (2020). Autoeficacia y logro de aprendizaje en estudiantes con diferente estilo cognitivo en un ambiente m-learning. Pensamiento Psicológico, 18(1), 71-85. https://doi.org/10.11144/Javerianacali.PPSI18-1.alaeLinks ]

López, O., Hederich, C., & Camargo, A. (2012). Logro de aprendizaje en ambientes hipermediales: Andamiaje autorregulador y estilo cognitivo. Revista Latinoamericana de Psicología, 44(2), 13-25. [ Links ]

Martin, S., Brunner, P., Holdgraf, C., Heinze, H. J., Crone, N. E., Rieger, J., Schalk, G., Knight, R. T., & Pasley, B. N. (2014). Decoding spectrotemporal features of overt and covert speech from the human cortex. Frontiers in Neuroengineering, 7(MAY), 1-15. https://doi.org/10.3389/fneng.2014.00014Links ]

Matsumoto, M., & Hori, J. (2013). Classification of silent speech using adaptive collection. Proceedings of the 2013 IEEE Symposium on Computational Intelligence in Rehabilitation and Assistive Technologies, CIRAT 2013 - 2013 IEEE Symposium Series on Computational Intelligence, SSCI 2013, 5-12. https://doi.org/10.1109/CIRAT.2013.6613816Links ]

Matsumoto, M., & Hori, J. (2014). Classification of silent speech using support vector machine and relevance vector machine. Applied Soft Computing Journal, 20, 95-102. https://doi.org/10.1016/j.asoc.2013.10.023Links ]

Min, B., Kim, J., Park, H. J., & Lee, B. (2016). Vowel Imagery Decoding toward Silent Speech BCI Using Extreme Learning Machine with Electroencephalogram. BioMed Research International. https://doi.org/10.1155/2016/2618265Links ]

Morooka, T., Ishizuka, K., & Kobayashi, N. (2018). Electroencephalographic Analysis of Auditory Imagination to Realize Silent Speech BCI. 2018 IEEE 7th Global Conference on Consumer Electronics, GCCE 2018, October, 683-686. [ Links ]

Nguyen, C. H., Karavas, G. K., & Artemiadis, P. (2018). Inferring imagined speech using EEG signals: a new approach using Riemannian manifold features. Journal of Neural Engineering, 15(1), 016002. https://doi.org/10.1088/1741-2552/aa8235Links ]

Oltman, P. K., Semple, C., & Goldstein, L. (1979). Cognitive style and interhemispheric differentiation in the EEG. Neuropsychologia, 17(6), 699702. [ Links ]

Poeppel, D., & Hickok, G. (2004). Towards a new functional anatomy of language. Cognition, 92(1-2), 112. https://doi.org/10.1016/j.cognition.2003.11.001Links ]

Pressel Coretto, G. A., Gareis, I. E., & Rufiner, H. L. (2017). Open access database of EEG signals recorded during imagined speech. 12th International Symposium on Medical Information Processing and Analysis, 10160, 1016002. https://doi.org/10.1117/12.2255697Links ]

Proakis, J. G., & Manolakis, D. G. (2007). Digital signal processing: Principles, algorithms, and applications (4th ed.). Pearson Education. [ Links ]

Qureshi, M. N. I., Min, B., Park, H. J., Cho, D., Choi, W., & Lee, B. (2018). Multiclass classification of word imagination speech with hybrid connectivity features. IEEE Transactions on Biomedical Engineering, 65(10), 2168-2177. https://doi.org/10.1109/TBME.2017.2786251Links ]

Rashid, M., Sulaiman, N., P. P. Abdul Majeed, A., Musa, R. M., Ahmad, A. F., Bari, B. S., & Khatun, S. (2020). Current status, challenges, and possible solutions of eeg-based brain-computer interface: A comprehensive review. Frontiers in Neurorobotics, 14(June), 1-35. https://doi.org/10.3389/fnbot.2020.00025Links ]

Riaz, A., Akhtar, S., Iftikhar, S., Khan, A. A., & Salman, A. (2015). Inter comparison of classification techniques for vowel speech imagery using EEG sensors. 2014 2nd International Conference on Systems and Informatics, ICSAI 2014, Icsai, 712-717. https://doi.org/10.1109/ICSAI.2014.7009378Links ]

Sarmiento, L. C., Lorenzana, P., Cortés, C. J., Arcos, W. J., Bacca, J. A., & Tovar, A. (2014). Brain computer interface (BCI) with EEG signals for automatic vowel recognition based on articulation mode. ISSNIP Biosignals and Biorobotics Conference, BRC. https://doi.org/10.1109/BRC.2014.6880997Links ]

Sarmiento, L. C., Villamizar, S., López, O., Collazos, A. C., Sarmiento, J., & Rodríguez, J. B. (2021). Recognition of eeg signals from imagined vowels using deep learning methods. Sensors, 21(19), 6503. https://doi.org/10.3390/s21196503Links ]

Solórzano-Restrepo, J., & López-Vargas, O. (2019). Efecto diferencial de un andamiaje metacognitivo en un ambiente e-learning sobre la carga cognitiva, el logro de aprendizaje y la habilidad metacognitiva. Suma Psicológica, 24(1), 33-50. https://doi.org/10.14349/sumapsi.2019.v26.n1.5Links ]

Valencia-Vallejo, N., López-Vargas, O., & Sanabria-Rodríguez, L. (2019). Effect of a metacognitive scaffolding on self-efficacy, metacognition, and achievement in e-learning environments. Knowledge Management and E-Learning, 11(1), 1-19. https://doi.org/10.34105/j.kmel.2019.11.001Links ]

Villamizar, S. I., Sarmiento, L. C., López, O., Caballero, J., & Bacca, J. (2021). EEG vowel silent speech signal discrimination based on apit-emd and svd. Lecture Notes in Electrical Engineering, 685 LNEE(Mi), 74-83. https://doi.org/10.1007/978-3-030-53021-1_8Links ]

Wang, L., Zhang, X., Zhong, X., & Zhang, Y. (2013). Analysis and classification of speech imagery EEG for BCI. Biomedical Signal Processing and Control, 8(6), 901-908. https://doi.org/10.1016/j.bspc.2013.07.011Links ]

Witkin, H. A., Moore, C. A., Goodenough, D. R., & Cox, P. W. (1977). Field-Dependent and Field-Independent Cognitive Styles and Their Educational Implications. Review of Educational Research, 47(1), 1-64. https://bit.ly/3q6bncGLinks ]

Yoshimura, N., Nishimoto, A., Belkacem, A. N., Shin, D., Kambara, H., Hanakawa, T., & Koike, Y. (2016). Decoding of covert vowel articulation using electroencephalography cortical currents. Frontiers in Neuroscience, 10(MAY), 1-15. https://doi.org/10.3389/fnins.2016.00175Links ]

Yuan, R., Lv, Y., & Song, G. (2018). Multi-fault diagnosis of rolling bearings via adaptive projection intrinsically transformed multivariate empirical mode decomposition and high order singular value decomposition. Sensors (Switzerland), 18(4), 1210. https://doi.org/10.3390/s18041210Links ]

Zoccolotti, P. (1982). Field dependence, laterality and the EEG: A reanalysis of O’Connor and Shaw Biological Psychology, 15, 203-207. [ Links ]

Received: August 29, 2021; Accepted: January 19, 2022

* Corresponding author: Omar López-Vargas, email: olopezv@pedagogica.edu.co

Creative Commons License This is an open-access article distributed under the terms of the Creative Commons Attribution License