Next Article in Journal
MBPPE: A Modular Batch Processing Platform for Electroencephalography
Next Article in Special Issue
Assessing Gender Bias in Auditory-Perceptual Ratings of Tracheoesophageal Speakers
Previous Article in Journal
GENICS: A Framework for Generating Attack Scenarios for Cybersecurity Exercises on Industrial Control Systems
Previous Article in Special Issue
Effect of Subglottic Stenosis on Expiratory Sound Using Direct Noise Calculation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning for Neuromuscular Control of Vocal Source for Voice Production

1
Utah Center for Vocology, University of Utah, Salt Lake City, UT 84112, USA
2
School of Biological Sciences, University of Utah, Salt Lake City, UT 84112, USA
*
Author to whom correspondence should be addressed.
Submission received: 28 October 2023 / Revised: 8 January 2024 / Accepted: 8 January 2024 / Published: 16 January 2024
(This article belongs to the Special Issue Computational Methods and Engineering Solutions to Voice III)

Abstract

:
A computational neuromuscular control system that generates lung pressure and three intrinsic laryngeal muscle activations (cricothyroid, thyroarytenoid, and lateral cricoarytenoid) to control the vocal source was developed. In the current study, LeTalker, a biophysical computational model of the vocal system was used as the physical plant. In the LeTalker, a three-mass vocal fold model was used to simulate self-sustained vocal fold oscillation. A constant /ə/ vowel was used for the vocal tract shape. The trachea was modeled after MRI measurements. The neuromuscular control system generates control parameters to achieve four acoustic targets (fundamental frequency, sound pressure level, normalized spectral centroid, and signal-to-noise ratio) and four somatosensory targets (vocal fold length, and longitudinal fiber stress in the three vocal fold layers). The deep-learning-based control system comprises one acoustic feedforward controller and two feedback (acoustic and somatosensory) controllers. Fifty thousand steady speech signals were generated using the LeTalker for training the control system. The results demonstrated that the control system was able to generate the lung pressure and the three muscle activations such that the four acoustic and four somatosensory targets were reached with high accuracy. After training, the motor command corrections from the feedback controllers were minimal compared to the feedforward controller except for thyroarytenoid muscle activation.

1. Introduction

Voice production is a highly complex, fine motor skill that requires the coordinated function of three major components: lungs, larynx, and both sub-glottal (trachea) and supraglottal (vocal tract) airways. Lung pressure drives the airflow in the trachea towards the larynx. The activation of the intrinsic laryngeal muscles defines the posturing (adduction and tension) of the vocal folds present in the larynx. The airflow causes the vocal folds to oscillate under certain pre-phonatory vocal fold posturing conditions, generating audible pulses of airflow into the vocal tract. The vocal tract filters these pulses and radiates the sound into the atmosphere [1]. The control of voice production involves the complex integration of multiple types of information by the brain including acoustic, auditory, somatosensory, proprioceptive, and motor representations [2]. The frontal, temporal, and parietal lobes of the cerebral cortex, along with the subcortical structures such as the cerebellum, basal ganglia, brain stem and their functional connections form the voice motor control system [3]. Approximately 100 different muscles act on the three subdivisions of the vocal system (the lungs, larynx, and airway) [4]. In this study, we investigated the role of auditory and somatosensory–motor control of the intrinsic laryngeal muscles (vocal source) that control the pre-phonatory vocal fold posturing. The airway (both subglottal and supraglottal) shape was assumed constant during phonation. Lung pressure was considered a variable of interest for aerodynamical control, but the lungs’ muscular control was not included.
Elaborating on the neural control of the vocal system, in the human brain, the spatial and temporal auditory information is converted to perceptual information by the peripheral and central auditory nuclei [5,6]. More accurately, the physical properties of the acoustic pressures, i.e., fundamental frequency, sound pressure level, harmonic and noise content are transformed to corresponding auditory perceptual properties, i.e., pitch, loudness, brightness, and roughness. The planning for producing a sound with desired characteristics and the required laryngeal muscle activations are performed in cortical regions of the brain. These direct pathways can be called feedforward pathways. The auditory feedback of the produced phonation integrates with the planned information to generate auditory error maps. These maps are transformed to motor information in the motor cortex. Similarly, the proprioceptive and somatosensory information from the larynx is fed back to the motor areas via the somatosensory cortex and cerebellum [7,8]. The somatosensory error maps are also transformed into corrective motor commands in the motor cortex. The combined motor commands from the direct and feedback pathways in the motor cortex, gated by the basal ganglia, are projected on to the phonatory neurons that directly control laryngeal muscles. A detailed map of the brain regions involved in planning of auditory and somatosensory targets, feedback and descending pathways can be found in Jurgens (2002) [9].
In humans, the learning of vocal communication is dependent on the feedback of auditory and somatosensory information during development. As speech motor planning is honed, auditory feedforward control dominates production [2]. Auditory feedback, however, is required to maintain the quality and precision of speech throughout adulthood [5,6]. Evidence based on compensation to perturbation in auditory feedback and jaw movement indicate a key role for somatosensory feedback in both the learning and maintenance of speech [7]. The auditory and somatosensory feedback is essential to understand if the desired outcome was achieved through the feedforward control, and if not, to generate corrective motor commands such that the intended outcome is attained [10].
The majority of approaches to vocal source simulation to date have been open-loop. Physical parameters are specified, and acoustic output is produced. There is no internal recall schema for preferred paths through a maze of physical parameters. All explorations begin from scratch. Currently, there are a limited number of studies addressing the neural control of vocal source and vocal tract [11,12,13,14]. In particular, the DIVA neural control model for the vocal tract was recently expanded to include the control of the body-cover vocal source model (named LaDIVA) [15]. The acoustic targets for LaDIVA are the fundamental frequency (fo) and sound pressure level (SPL), and the control parameters are subglottal pressure (Ps), and cricothyroid (CT) and thyroarytenoid (TA) muscle activations. Somatosensory targets are not included in the current version of LaDIVA.
The current study used a similar architecture to the DIVA model, i.e., one feedforward controller, one acoustic feedback controller, and one somatosensory feedback controller. However, the two differ in the composition of the controllers. The DIVA/LaDIVA model used linear Jacobian matrices to model the feedback controllers while the current study developed non-linear deep neural networks-based controllers, allowing more complex control mechanisms. Also, the current study included acoustic measures of voice quality (brightness and roughness) as acoustic targets, along with fo and SPL. Control parameters included lateral cricoarytenoid muscle activation along with the lung pressure (PL), CT and TA activations. In contrast to the LaDIVA model, four somatosensory targets (vocal fold length, and fiber stress in the three vocal fold layers) were also included.
Machine learning and artificial intelligence are gaining traction in voice and speech research. They are being used for the automatic detection of voice disorders [16,17,18], improving the computational efficiency of simulators [19], and estimation of voice control parameters from acoustic output signals [20,21]. Currently, machine learning is being used to detect disorders such as laryngeal cancer, dysphonia, vocal fold nodules, polyps, edema, vocal fold paralysis, and neuromuscular disorders from voice and speech samples. Additionally, computationally expensive components of the traditional voice and speech production models such as the flow and pressure calculations are being replaced by faster machine learning models [19]. Unsupervised learning techniques are being used for disorder detection [22], emotion recognition [23], and voice quality detection using voice and speech samples [24].
Reinforcement learning is another branch of machine learning where the training data are not needed ahead of time [25]. The model learns to control the physical plant by interacting with it in a trial-and-error manner. The reinforcement learning approach can be used to learn the control of the vocal system. These neural control systems try to mimic how the brain controls the vocal system. When used with voice simulators, such neural control systems can provide insights into neuromuscular disorders such as vocal tremor, Parkinson’s disease, and spasmodic dysphonia. Even though the DIVA model, and other neural controllers of the vocal system, including the control system developed in the current study do not use reinforcement learning in its true sense, they fall under its broader category.
The primary objective of the current study was to develop a deep learning-based neural control system for the vocal source with feedforward and feedback controllers that generates lung pressure and intrinsic laryngeal muscle activation levels to achieve acoustic targets of interest. The secondary objective was to understand the relative role of the feedback and feedforward controllers towards the total motor command.

2. Materials and Methods

2.1. Control System

The neural architecture designed in the current study to control the physical plant (LeTalker) is shown in Figure 1. Three controllers are used to control the LeTalker. The control begins with four auditory perceptual features, pitch, loudness, brightness, and roughness [26]. These perceptual features were expressed in terms of four acoustic physical quantities (fundamental frequency, sound pressure level, normalized spectral centroid, and signal-to-noise ratio, respectively) [27]. The pressure-to-acoustic converter neural network (PtoA NN), which was pretrained, estimates the corresponding acoustic features from the radiated pressure (Po) signal. The acoustic and somatosensory feedback controllers, respectively, receive the generated acoustic aspects of phonation and somatosensory states from the physical plant. The error between the targets and their corresponding predictions serves as input to the feedback controllers. The resulting motor commands from these negative feedback controllers minimize this error.
The feedback controllers, due to the nature of their inputs, suffer from delays in the feedback loop. Physiologically, this delay is compensated by auditory planning [28]. The feedforward controller represents the auditory planning in the higher-order brain regions and generates motor commands from auditory targets. In the current study, the feedforward controller was designed to take the acoustic targets as input and generate the feedforward motor commands. The output of the three controllers was added together to model the final phase of the sensory–motor transformation, i.e., produce motor commands. These commands were saturated between 0 and 1 before passing them as inputs to the physical plant.

2.2. LeTalker

A biophysical computational model (LeTalker) of the vocal system was used as the physical plant in the current study. The LeTalker takes lung pressure and three intrinsic laryngeal muscle activations (cricothyroid (CT), thyroarytenoid (TA), and lateral cricoarytenoid (LCA)) as inputs and generates the radiated pressure (Po) and the four somatosensory features as an output at every time sample. The four somatosensory features are the vocal fold length (Lg), and the fiber stress ( σ m u c ,   σ l i g ,   σ m u s ) in the three vocal fold layers (mucosa, vocal ligament, and TA muscle), respectively. Each phonation was 0.2 s long and simulated at a sampling rate of 44,100 Hz. A brief description of LeTalker’s dynamics is provided here; for a detailed explanation refer to [29,30,31].

2.2.1. Three-Mass Vocal Fold Model

Anatomically, each vocal fold is subdivided into five layers, as shown in Figure 2A. Computationally, the five layers are grouped either into two-layer schema (body (epithelium, superficial, and intermediate layers)-cover (deep layer and TA muscle)) or three-layer schema (mucosa (epithelium and superficial layers), vocal ligament (intermediate and deep layers), and TA muscle). In the LeTalker, the vocal fold dynamics were modeled using the two-layer schema and the three-layer schema was used to compute the fiber stress.
The three-mass model shown in Figure 2B is a lumped element approximation of the body-cover layer schema of the vocal folds. The two vocal folds were assumed to be symmetric and have the same movement. Male vocal fold dimensions of 1.6 cm length (L0), 0.8 cm thickness (T0), and 0.8 cm depth (D0) were considered. Two masses ( m u and m l ) were used to model the cover layer, and the third mass ( m b ) simulated the effect of the body layer. The two masses of the cover layer were connected using a coupling spring. The coupling spring ( k c ) simulates the shear forces in the cover layer. Each of the cover layer masses was connected to the body layer mass using nonlinear springs ( k u and k l ) and damping elements ( d u and d l ). The two springs simulate the cover layer’s stiffness and the effective coupling stiffness between the cover and body layers. The body mass was connected to the rigid thyroid cartilage boundary using a spring ( k b ), which represent the effective stiffness of the body, and a damping element ( d b ). The equations of motion for the three masses were given in [31].

2.2.2. Subglottal and Supraglottal Airway

The airway was modeled as a series of cylindrical sections. As the focus of the current study was on the neuromuscular control of the vocal source, the airway was kept constant without any articulation. The subglottal airway was modeled after MRI measurements [32] and consisted of 32 cylindrical sections. An /ə/ vowel shape with 44 cylindrical sections was used for the supraglottal airway [33]. Each subglottal and supraglottal section was 0.3968 cm long [34], half the distance the sound travels in 1/44,100 s. The area function values for both subglottal and supraglottal airways were given in [35]. Attenuation factors were used to model the airway losses [29]. The choice of the section length allowed the computation of forward and backward traveling waves at section boundaries [34]. Radiation impedance was modeled by an inertance (L) and resistance (R) in parallel as in [36].

2.2.3. Glottal Flow

The three-mass vocal fold model was connected to the subglottal and supraglottal airways. The glottal flow equation, u g which is the primary vocal source can then be derived using the incident (P+) and reflected (P) pressures at the glottis from the vocal tract. The computation of the glottal flow equation was described in [37]. Here, only the final equation was given:
u g = a g c k t a g A ± a g A 2 + 4 k t ρ c 2 P s + P e 1 / 2
where ρ = 0.00114   g / cm 3 is the air density, c is the speed of sound in air, a g is the glottal area, P s + is the forward traveling subglottal incident pressure on the glottis, and P e is the backward traveling supraglottal incident pressure on the glottis (see Figure 2B). If A s is the subglottal area, and A e is the supraglottal area, then the effective vocal tract area, A was given as
A = A s A e A s + A e
and k t as
k t = 1 2 a g / A e 1 a g / A e

2.2.4. Fiber Stress in the Vocal Fold Layers

The spring constants in the three-mass model can be augmented to account for anisotropic tension in the fibers present in the vocal fold layers. The body-cover representation of the vocal folds was divided into three layers (mucosa, vocal ligament, and thyroarytenoid muscle) to compute the fiber stress. The equations to compute the fiber stress in the three vocal fold layers were detailed in [31]. The passive stress in the three vocal fold layers was computed using the following equation
σ p = 0                         f o r   ϵ < ϵ 1 , σ 0 ϵ 1 ϵ ϵ 1                         f o r   ϵ 1 ϵ ϵ 2 , σ 0 ϵ 1 ϵ ϵ 1 + σ 2 e C ϵ ϵ 2 C ϵ ϵ 2 1             f o r   ϵ > ϵ 2 .
The mucosa and ligament layers have only passive stress, whereas the TA muscle has both active and passive stress. Therefore, the fiber stress, σ m u s in the TA muscle layer was given as
σ m u s = a T A σ a m max 0 , 1 b ϵ ϵ m 2 + σ p
where ϵ is the vocal fold strain, ϵ 1 is the strain where the linear portion of the passive stress goes to zero, ϵ 2 is the strain where the exponential part begins, σ 0 is the stress when ϵ = 0 , σ 2 is a scale factor for the exponential portion, a T A is the thyroarytenoid muscle activation level between 0 and 1, σ a m = 105   kPa is the maximum active stress in the TA muscle, ϵ m = 0.4 is the optimum sarcomere strain, and b = 1.07 is the empirically determined constant for each muscle. Table 1 provides the parameter values used in the current study to compute the fiber stress in the three vocal fold layers. Except where suggested, the parameters in Table 1 do not have any units.

2.2.5. Rules for Muscle Control of the Three-Mass Model

Vocal fold posturing is defined as the adduction, abduction, elongation, or shortening of the vocal folds resulting from the activation of intrinsic laryngeal muscles. In this section, the rules/equations that govern vocal fold posturing originally derived in [31] are presented. To reduce the number of control parameters, only three muscle activations were used: cricothyroid (CT), thyroarytenoid (TA), and lateral cricoarynteoid (LCA). The LCA muscle can approximate the role of the other two muscles (interarytenoid, and posterior cricoarytenoid).

Vocal Fold Elongation Rule

ϵ = G R T a C T a T A H a L C
where G = 0.2 is the gain of elongation, R T = 3.0 is the torque ratio, H = 0.2 is the adductory strain factor, and a C T ,   a T A ,   a L C are the CT, TA, and LCA muscle activation levels, respectively, between 0 and 1.

Dynamic Vocal Fold Length, Thickness, and Depth Rule

L e n g t h ,   L g = L 0 1 + ϵ
T h i c k n e s s ,   T = T 0 1 + 0.8 ϵ
D e p t h   o f   b o d y   l a y e r ,   D b = a T A D m u s + 0.5 D l i g 1 + 0.2 ϵ
D e p t h   o f   c o v e r   l a y e r ,   D c = D m u c + 0.5 D l i g 1 + 0.2 ϵ
where D m u c = 0.2   cm , D l i g = 0.2   cm , and D m u s = 0.4   cm are the depth of the mucosa, vocal ligament, and TA muscle layers, respectively.

Vocal Fold Adduction Rule

The adduction rule for the glottal half-width ξ 02 at the vocal processes was governed by the LCA muscle activation.
ξ 02 = 0.25 L 0 1 2 a L C

Vocal Fold Convergence Rule

The TA muscle governs the prephonatory convergence of the glottis.
ξ c = T 0.05 0.15 a T A
These rules, governed by the intrinsic laryngeal muscle activations, change the adduction and tension in the vocal folds. They, in turn, change the dynamics of vocal fold vibration, resulting in several acoustic characteristics in the sound. Figure 3 shows example radiated pressure (Po), glottal flow (ug), and glottal area signals (ag) generated by the LeTalker. The muscle activation plots of the acoustic and somatosensory features with respect to their dominant control parameters were given in [35].

2.3. Neural Network Architecture

The feedforward and feedback controllers, along with the PtoA NN, consisted of three parts: 1. an input layer with the same number of cells as the number of inputs; 2. a core with the hidden layers; and 3. a full connected output layer with the same number of cells as the number of output parameters. Figure 4 shows the structure of the neural networks.
The auditory and somatosensory feedback controllers receive error information, i.e., the difference between targets and corresponding predictions and transform them into lung pressure and muscle activations. Therefore, the input layer consists of four inputs, the core consists of three layers of 128 neurons each, and the output layer consists of four cells corresponding to the lung pressure and the three muscle activations. Similarly, the feedforward controller uses a four-cell input layer corresponding to the four acoustic planned targets and the output layer has four cells corresponding to the lung pressure and the three muscle states. The core of the acoustic feedforward controller consists of three layers of 256 neurons each. The number of layers and units used in the layers was empirically determined. All the layers of the feedback controllers use a hyperbolic tangent activation function with a −1 to 1 range. This allows for the feedback controllers to compensate for the error in motor commands in both positive and negative directions. All the layers of the feedforward controller use a sigmoid activation function allowing the control outputs to vary between their range of 0 and 1.
The LeTalker generates one radiated pressure value at every time sample. A buffer was maintained to store 1102 radiated pressure values (i.e., 25 ms of sound output at 44,100 Hz sampling rate). The buffer was updated at every time sample using a 1-sample sliding window of 25 ms. The PtoA NN takes this buffer of 1102 radiated pressure values as input and generates four acoustic predictions as output at every time sample. Therefore, the PtoA NN has 1102 neurons in the input layer, 4 neurons in the output layer, and 512 neurons in each of the hidden layers.

2.4. Training and Testing the Neural Networks

To train and test the control system, 50,000 radiated pressure signals with a length of 0.2 s each were generated by varying the lung pressure and the three muscle activations with Monte Carlo random sampling. The P L and a L C were generated using normal distributions, whereas a C T and a T A were generated using uniform distributions. The set of these 50,000 control parameters ( M T ) were given as input to LeTalker, and the corresponding steady-state somatosensory and acoustic features were computed. Among them, a set of 40,000 features were used as targets for training, and the remaining 10,000 were used as targets during testing of the controllers. The entire set of 40,000 acoustic and somatosensory targets were used simultaneously to train the controllers. The acoustic features were transformed into a normal distribution and the fiber stress were transformed into a logarithmic scale. The weights for all the three controllers were updated simultaneously at every time sample. The targets were kept constant for the 0.2 s of simulation without any shuffling. The cost function to update the weights was the mean-squared error between the targets and their predictions as the targets were all continuous variables:
c o s t = 1 N 1 N M T , S T , A T M P , S P , A P 2
where M T , S T , A T are the control, somatosensory, and acoustic targets, respectively, and M P , S P , A P are the current control, somatosensory, and acoustic predictions, respectively. An Adam optimizer with a learning rate of 5 × 10−4 was used for updating the weights. Since all the N = 40,000 targets were used simultaneously, only one run of 0.2 s was used to train the controllers. As a result, the weights were updated 8820 times (0.2 s of simulation at 44,100 Hz sampling rate).
Each of the 50,000 radiated pressure signals was divided into 50 segments, each with a length of 25 ms (1102 samples) to train the PtoA NN. This was accomplished by using a 25 ms sliding window with a step size of 3.5 ms. Thus, the 50,000 signals were split into 2.5 million segments. For each of the 2.5 million segments, corresponding acoustic features were computed. Among the 2.5 million segments, 1.5 million were used to train, and the remaining 1 million were used to test the PtoA NN. An Adam optimizer with a learning rate of 1 × 10−3 was used to update the PtoA NN weights. The training was run for 800 epochs. The cost function to train the PtoA NN was the mean-squared error between the targets and predictions, as given below:
c o s t = 1 N n = 1 N A T A P 2
where A T are the acoustic targets, A P are the acoustic predictions, and N is the number of training segments. The trained models were tested on the data that were not used for training using mean-squared error, mean absolute error, and error in % metrics.

3. Results

3.1. PtoA Neural Network

The change in cost function during training as a function of the number of epochs is shown in Figure 5. The training cost reduced to 2.07 × 10−4 in the normalized units by the end of the training with an accuracy of 0.96. On the test data, the cost was 4.03 × 10−4, with an accuracy of 0.94. Boxplots for the percentage error computed based on Equation (15) for all the four acoustic features from the test data are shown in Figure 6.
E r r o r   i n   % = A T A P A T × 100
For better clarity, the y-axis was limited to ± 10 % . However, the outliers extended up to −500% and +100%. A value was considered an outlier if it was away more than 1.5 times the interquartile range from the box’s edges. The percentage of outliers with respect to the total test segments for fo, SPL, NSC, and SNR were obtained as 6.9%, 2.7%, 3.8%, and 8.6%, respectively. The percentage error for all four acoustic features varied between ± 6 % , excluding the outliers (Figure 6). The median error was close to 0% for all the features except SPL, whose median error was at 1.7%. This indicates that the predicted value of SPL is often less than the target. The range of error was higher for NSC and SNR compared to that of fo and SPL. This was expected given the abstractness of brightness and roughness the NSC and SNR represent, respectively.
Figure 7 shows the boxplots for the difference between targets and predictions in the actual units of the acoustic features from the 1 million test segments. It can be observed that the error was between ± 10   Hz for fo, ± 1   dB for SPL, ± 1 for NSC, and ± 1 dB for SNR, excluding the outliers. The ranges for fo, SPL, NSC, and SNR across all the 2.5 millions segments were 350 Hz, 60 dB, 66 (no units), and 124 dB, respectively. This suggests that the prediction error obtained was only a tiny fraction of the target values for most cases, indicating that the PtoA NN can accurately predict the acoustic features for a wide variety of sound signals.

3.2. Control System

Figure 8 shows how the cost function given in Equation (13) was reduced as a function of the epoch (sample) number for the training data. The cost reduced to 0.0032 in normalized units at the end of the 8820 epochs. The error did not distribute equally across the three types of features/parameters. The individual cost values in normalized units for control, somatosensory, and acoustic targets were obtained as 0.0067, 0.0019, and 0.0011, respectively, at the end of the training. The error was highest for predicting muscle activations. Physiologically, a set of somatosensory and acoustic features could be achieved by many combinations of muscle activations, described as motor equivalence. Therefore, higher errors in muscle states does not necessarily imply error in reaching the targets but the presence of alternate activations that result in the same somatosensory and acoustic targets. A combined cost value of 0.0041 was obtained for the test data in normalized units at the end of 0.2 s of simulation. The individual cost values in normalized units for muscle, somatosensory, and acoustic features from the test data were obtained as 0.0076, 0.0021, and 0.0027, respectively, at the end of the 0.2 s of simulation. This indicates that the performance decreased on the test data, but only slightly. Here, as well, the major error occurred in reaching the muscle targets.
Figure 9 shows the example target and predicted radiated pressure (sound) signals generated by LeTalker for some of the test data cases. The first row shows the pressure signal with an fo in the speech range, the second row shows a signal with a higher fo, the third row shows the waveforms with higher SPL, the fourth row shows the waveforms with a lower spectral centroid, and the fifth row shows the waveforms with more noise in the signal. From the waveforms, it can be observed that the predicted signals reached the steady state between 40 and 80 ms, typically seen in LeTalker. The predicted radiated pressure signals obtained by the control of neural architecture were very similar to the targeted signals. However, there were slight differences between the targets and the predicted signals. The difference is apparent in the pitch and amplitude of the signals. For example, the fo is lower in the predicted signal compared to the targeted signal (second row), and the amplitude is lower in the predicted signal compared to the targeted signal (fourth row). The differences in other two features cannot be observed easily from the waveforms. Therefore, the targeted features and their corresponding predictions for these five examples are given in Table 2.
From Table 2, it can be observed that for the muscle parameters, a significant difference occurred in reaching LCA activation ( a L C ). There were slight differences in the prediction of a C T and a T A as well (first and fifth row). Lung pressure was predicted accurately for all five examples. With regard to somatosensory features, the differences are small for all the four features. For the acoustic features, there are significant differences in some cases for all the features. These results indicate that across all the test cases, the error would be higher in the prediction of a L C , and moderately high in the prediction of a C T , a T A and the four acoustic features.
Figure 10 shows the error in % for all the twelve features/parameters from the controller test data. Like the boxplots for PtoA NN results, the y-axis was limited to exclude the outliers for better visibility. As observed from the individual cost values, the error in % is higher in predicting muscle targets ( ± 80 % ) than somatosensory and acoustic targets ( ± 30 % ), excluding outliers. For the muscle parameters, the order in error is P L < a C T < a T A < a L C according to each box’s interquartile range and the extent of whiskers. Among the somatosensory features, the error in % is moderately high for σ l i g ( ± 30 % ), followed by Lg ( ± 21 % ), excluding the outliers. For the other two features ( σ m u c and σ m u s ), the error in % is low ( ± 11 % ) across all the test signals, excluding the outliers. For the acoustic features, the error in % is moderately high in the prediction of SNR ( ± 30 % ), and very low in the prediction of fo ( ± 7 % ), excluding the outliers. For the SPL and NSC, the error is relatively low ( ± 12 % ).
Figure 11 shows the difference between the targets and predictions in actual units for somatosensory and acoustic features. The control parameters were not included because they are non-dimensional, ranging between 0 and 1. The error is between ± 0.2 cm for L g , ± 100   kPa for σ m u c , and ± 20   kPa for both σ l i g and σ m u s . For the acoustic features, the error is between ± 15   Hz for f o , ± 2   dB for SPL, ± 3 for NSC, and ± 4   dB for SNR, excluding the outliers.
Figure 12 shows the boxplot of outputs generated by each controller for the test data. It can be observed that the acoustic feedforward controller (affc) is the dominant controller after training. The range of values generated by the affc is higher than the range of values generated by the feedback controllers except for TA muscle activation. Values generated by affc ranged between 0 and 1 except for a L C . The values produced by affc for a L C varied only in a small range around 0.5. Among the feedback controller outputs, only a T A varied significantly, suggesting that more than one controller was necessary to learn the TA muscle control’s complexity.

4. Discussion

The current study focused on developing a neural-network-based control system for the vocal source. One acoustics feedforward controller (affc) and two feedback controllers (sfbc and afbc) were used to generate muscle activations and lung pressure to control the vocal source. A three-mass-model-based vocal system called LeTalker was used as the physical plant. The LeTalker generates somatosensory features as output, but the acoustic features should be computed from the oral pressure signal. Therefore, in the current study, a neural network (PtoA NN) was trained to compute the four acoustic features using 25 ms of the oral pressure signal. The results indicated that the PtoA NN was able to predict the four acoustic features very accurately. The error was less than ± 6 % for all the four acoustic features, excluding the outliers. The number of outliers ranged between 2.7% and 8.6% for the four acoustic features. A significant number of outliers in the data is expected because the lung pressure and the three muscle activations were varied randomly to generate the data. Unusual input combinations would result in several configurations that does not result in natural phonation, making the predictions extremely difficult. However, varying the input parameters randomly was necessary to explore the entire territory and make the controllers robust. The results from the controller testing suggested that, for a given target, the control system produces parameters that result in sustained phonation that is within ± 15   Hz of the targeted fo and ± 2   dB of the targeted SPL, highlighting the stability of the control system. Previous studies found that accurate voice imitators also made errors as high as ± 15   Hz during pitch imitation tasks [38], indicating excellent performance by the developed control system.
In the current study, the weights of the three controllers were updated simultaneously using a single cost function without providing any a priori knowledge about the role to each controller. Such an approach facilitated the testing of the hypothesis that the feedforward controller dominates the control after training compared to the feedback controllers. The results obtained in the current study supported the hypothesis that the range of affc outputs is higher than the range of sfbc and afbc outputs. This finding was observed in other studies as well [3,39]. Due to the inherent delays in the feedback, after training, the feedforward controller dominates, and the contribution of the feedback controllers will be minimal unless there is external disturbance [40]. This also confirms the previous observations where adults could produce speech intelligibly even after hearing loss [41]. The variation in the outputs of the feedforward controller was significantly higher than the variation in the outputs of the feedback controllers except for the TA muscle activation. This is understandable because TA muscle control is more complex compared to the other control parameters. The TA muscle changes both adduction (thyromuscularis portion of TA muscle) as well as tension (thyrovocalis portion of TA muscle) in the vocal folds.
The example waveforms indicated that the controllers could generate lung pressure and muscle activations that produced a wide variety of phonations. However, there were some differences between the targeted and predicted waveforms in terms of pitch and amplitude. It could be because of the representation of the speech signal with only four acoustic features which may not be adequate. We will address this limitation in future studies by exploring muti-parameter representation of the timbre.
Across all the test cases, the error in % was highest in reaching muscle targets ( ± 80 % ). Major contributions came from a L C and a T A . The error in % for the other two control parameters was less than ± 40 % . The error in % was less than ± 30 % for both somatosensory and acoustic features, excluding the outliers. The higher error in the prediction of control targets could be due to the nonlinearity and motor equivalence present in LeTalker. The LeTalker includes non-linear source–filter interaction phenomenon [33] described in Equation (1), which can lead to the generation of same output for multiple combinations of muscle activations. The somatosensory and acoustic features were reached with little variation in a L C activation around 0.5. This could be due to the minimal dependence of the chosen acoustic and somatosensory features in the current study on a L C . The a L C is primarily an adductor muscle, and glottal width was not considered as one of the somatosensory features in the current study. The glottal width is two-dimensional [42] and varies along the length and thickness of the vocal fold surface. It is typically measured at the level of the vocal processes (posterior end of the vocal folds). In the LeTalker, the steady-state glottal width was zero at the vocal processes for most of the control parameter combinations, resulting in no pattern to learn with a change in control parameters. As noticed with individual waveforms, the difference in targeted and predicted acoustic and somatosensory features could be due to the use of only four acoustic and four somatosensory features. We will explore more accurate representation of the speech waveform using acoustic features, especially for timbre, and include more somatosensory and proprioceptive features for better representation of the vocal fold posturing.
Neural networks train well on normally distributed data [43,44]. However, varying all the control parameters with normal distributions in the current study eliminated data at extreme values (i.e., values closer to 0 and closer to 1) of the muscle activations. Randomly varying all the control parameters using uniform distribution produced data even at the extremes of muscle activations. However, in the current study, the neural networks performed poorly on uniformly distributed control parameters. Therefore, the data to train and test the control system were generated by varying PL and a L C using normal distribution, and a T A and a C T using uniform distribution as a compromise. In the future, we will build the control system to overcome this limitation. Some of the somatosensory and acoustic features generated using the above-mentioned control parameters have a non-Gaussian distribution. Therefore, the acoustic features were transformed into normal distribution using a quantile transformer and the three fiber stresses were converted into logarithmic scale before training the networks. Such a transformation enhanced the performance of the neural networks.
Since the targets were constant across time in the current setup, we used feedforward artificial neural networks to represent the controllers. However, we will use recurrent neural networks in the future to accommodate time-varying targets. Since the focus of the current study was the control of vocal source, we kept the airway constant. In the future, we will include the control of the vocal tract as well for more realistic applications. Also, in the current study, we used only the acoustic and somatosensory features as targets. In the future, we will include reflex proprioceptive loops as well for more accurate representation of the control system.
We believe that this work is a steppingstone to simulate and study various motor disorders related to voice. In disorders such as Parkinson’s disease, ataxia, and flaccid paralysis, the feedforward motor program is disrupted due to damage to downstream motor control processes. It alters the established relationships between the motor commands, sensory sequences, and movement outcomes [45]. Hearing loss, muscle atrophy, or injury to the larynx can alter the feedback mechanisms associated with voice production. The control system developed in the current study can simulate such voice and sensory disorders that disrupt feedforward and feedback mechanisms and provide insights into different strategies that can be applied to restore near normal voice production. Such strategies can guide surgical interventions as well as therapy.
The utility of this work reaches beyond phonosurgery and therapy. Health scientists may be able to develop neural stimulation strategies to address motor disorders [46], while engineers may use this information in the development of prosthetics, and talking and singing robots [47]. Understanding motor learning helps voice coaches design better instruction for skill acquisition [48]. Professional vocalists will be able to improve their practice skills, along with their performances. It is therefore expected that this study will have a broad impact on voice production in humans.

5. Conclusions

One acoustic feedforward controller and two feedback (somatosensory and acoustic) controllers were used to generate the lung pressure and three muscle activations to produce phonation with desired acoustic targets. After training the controllers, the feedforward controller dominated the feedback controllers in generating the control parameters despite explicitly not suggesting each controller’s role. The controllers automatically learned this behavior when all the controllers were trained simultaneously using a single cost function. Due to the motor equivalence and nonlinear nature of LeTalker, the error in reaching muscle targets ( ± 80 % ) was higher than the error in predicting the somatosensory and acoustic targets ( ± 30 % ). Among the somatosensory features, the error was higher in predicting σ l i g than the other three features. It was observed that for most of the test cases, for a given target, the controller produces lung pressure and muscle activations that result in phonation that is within ± 15   Hz of the targeted fo and ± 2   dB of the targeted SPL, indicating excellent performance.

Author Contributions

Conceptualization, A.P. and I.R.T.; methodology, A.P., R.K.A. and I.R.T.; software, A.P. and R.K.A.; validation, A.P.; formal analysis, A.P.; investigation, A.P.; resources, I.R.T.; data curation, A.P.; writing, A.P.; funding acquisition, I.R.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by NIH/NIDCD Grant No. R01DC017998. The content is solely the responsibility of the authors. It does not necessarily represent the official views of the National Institute of Deafness and Other Communication Disorders or National Institutes of Health.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data will be provided upon request by the corresponding author. The data are not publicly available due to the complex nature of the data.

Acknowledgments

We thank Brad Story for sharing the source code for LeTalker.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Titze, I.R. Principles of Voice Production; National Center for Voice and Speech: Iowa City, IA, USA, 2000. [Google Scholar]
  2. Guenther, F.H. Neural Control of Speech; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  3. Guenther, F.H.; Vladusich, T. A neural theory of speech acquisition and production. J. Neurolinguist. 2012, 25, 408–422. [Google Scholar] [CrossRef]
  4. Franken, M.K.; Eisner, F.; Acheson, D.J.; McQueen, J.M.; Hagoort, P.; Schoffelen, J.M. Self-monitoring in the cerebral cortex: Neural responses to small pitch shifts in auditory feedback during speech production. NeuroImage 2018, 179, 326–336. [Google Scholar] [CrossRef]
  5. Bizley, J.K.; Cohen, Y.E. The what, where and how of auditory-object perception. Nat. Rev. Neurosci. 2013, 14, 693–707. [Google Scholar] [CrossRef] [PubMed]
  6. Escabi, M.A.; Read, H.L. Representation of spectrotemporal sound information in the ascending auditory pathway. Biol. Cybern. 2003, 89, 350–362. [Google Scholar] [CrossRef] [PubMed]
  7. Tremblay, S.; Shiller, D.M.; Ostry, D.J. Somatosensory basis of speech production. Nature 2003, 423, 866–869. [Google Scholar] [CrossRef]
  8. Smith, D.J.; Stepp, C.; Guenther, F.H.; Kearney, E. Contributions of auditory and somatosensory feedback to vocal motor control. J. Speech Lang. Hear. Res. 2020, 63, 2039–2053. [Google Scholar] [CrossRef] [PubMed]
  9. Jurgens, U. Neural pathways underlying vocal control. Neurosci. Behav. Rev. 2002, 26, 235–258. [Google Scholar] [CrossRef]
  10. Ito, T.; Ohashi, H.; Gracco, V.L. Changes of orofacial somatosensory attenuation during speech production. Neurosci. Lett. 2020, 730, 135045. [Google Scholar] [CrossRef]
  11. Bailly, G. Learning to speak. Sensori-motor control of speech movements. Speech Commun. 1997, 22, 251–267. [Google Scholar] [CrossRef]
  12. Guenther, F.H.; Ghosh, S.S.; Tourville, J.A. Neural modeling and imaging of the cortical interactions underlying syllable production. Brain Lang. 2006, 96, 280–301. [Google Scholar] [CrossRef]
  13. Kroger, B.J.; Kannampuzha, J.; Rube, C.N. Towards neurocomputational model of speech production and perception. Speech Commun. 2009, 51, 793–809. [Google Scholar] [CrossRef]
  14. Markey, K.L. The Sensorimotor Foundations of Phonology: A Computational Model of Early Childhood Articulatory and Phonetic Development. Ph.D. Dissertation, University of Colorado, Boulder, CO, USA, 1994. [Google Scholar]
  15. Weerathunge, H.R.; Alzamendi, G.A.; Cler, G.J.; Guenther, F.H.; Stepp, C.E.; Zanartu, M. LaDIVA: A neurocomputational model providing laryngeal motor control for speech acquisition and production. PLoS Comput. Biol. 2022, 18, e1010159. [Google Scholar] [CrossRef] [PubMed]
  16. Al-Dhief, F.T.; Latiff, N.M.A.; Malik, N.N.N.A.; Salim, S.N.; Baki, M.M.; Albadr, M.A.A.; Mohammed, M.A. A survey of voice pathology surveillance systems based on Internet of Things and machine learning algorithms. IEEE Access 2020, 8, 64514–64533. [Google Scholar] [CrossRef]
  17. Verde, L.; De Pietro, G.; Sannino, G. Voice disorder identification by using machine learning techniques. IEEE Access 2018, 6, 16246–16255. [Google Scholar] [CrossRef]
  18. Hegde, S.; Shetty, S.; Rai, S.; Dodderi, T.A. A survey of machine learning approaches for automatic detection of voice disorders. J. Voice 2019, 33, e11–e947. [Google Scholar] [CrossRef] [PubMed]
  19. Zhang, Y.; Zheng, X.; Xue, Q. A deep neural network based glottal flow model for predicting fluid-structure interactions during voice production. Appl. Sci. 2020, 10, 705. [Google Scholar] [CrossRef]
  20. Zhang, Z. Voice feature selection to improve performance of machine learning models for voice production inversion. J. Voice 2021, 37, 479–485. [Google Scholar] [CrossRef]
  21. Zhang, Z. Estimation of vocal fold physiology from voice acoustics using machine learning. J. Acoust. Soc. Am. 2020, 147, EL264–EL270. [Google Scholar] [CrossRef]
  22. Rueda, A.; Krishnan, S. Clustering Parkinson’s and age-related voice impairment signal features for unsupervised learning. Adv. Data Sci. Adapt. Anal. 2018, 10, 1840007. [Google Scholar] [CrossRef]
  23. Zhang, Z.; Weninger, F.; Wollmer, M.; Schuller, B. Unsupervised learning in cross-corpus acoustic emotion recognition. In Proceedings of the 2011 IEEE Workshop on Automatic Speech Recognition & Understanding, Waikoloa, HI, USA, 11–15 December 2011; pp. 523–528. [Google Scholar]
  24. Kojima, T.; Fujimura, S.; Hasebe, K.; Okanoue, Y.; Shuya, O.; Yuki, R.; Shoji, K.; Hori, R.; Kishimoto, Y.; Omori, K. Objective assessment of pathological voice using artificial intelligence based on the BRBAS scale. J. Voice 2021. [Google Scholar] [CrossRef]
  25. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  26. Hickok, G.; Houde, J.F.; Rong, F. Sensorimotor integration in speech processing: Computational basis and neural organization. Neuron 2011, 69, 407–422. [Google Scholar] [CrossRef]
  27. Palaparthi, A.; Smith, S.; Titze, I.R. Mapping thyroarytenoid and cricothyroid activations to postural and acoustic features in a fiber-gel model of the vocal folds. Appl. Sci. 2019, 9, 4671. [Google Scholar] [CrossRef]
  28. Dastolfo-Hromack, C.; Walsh, E. Evaluation of neurogenic voice disorders. In Neurogenic and Neurodegenerative Diseases of the Larynx; Weissbrod, P., Francis, D., Eds.; Springer: Cham, Switzerland, 2020. [Google Scholar]
  29. Story, B.H.; Titze, I.R. Voice simulation with a body-cover model of the vocal folds. J. Acoust. Soc. Am. 1995, 97, 1249–1260. [Google Scholar] [CrossRef]
  30. Titze, I.R. Regulating glottal airflow in phonation: Application of the maximum power transfer theorem to a low dimensional phonation model. J. Acoust. Soc. Am. 2002, 111, 367–376. [Google Scholar] [CrossRef] [PubMed]
  31. Titze, I.R.; Story, B.H. Rules for controlling low-dimensional vocal fold models with muscle activation. J. Acoust. Soc. Am. 2002, 112, 1064–1076. [Google Scholar] [CrossRef] [PubMed]
  32. Story, B.H. Using imaging and modeling techniques to understand the relation between vocal tract shape to acoustic characteristics. In Proceedings of the Stockholm Music Acoustics Conference, Stockholm, Sweden, 14–15 June 2003. [Google Scholar]
  33. Palaparthi, A.; Titze, I.R. Analysis of glottal inverse filtering in the presence of source-filter interaction. Speech Commun. 2020, 123, 98–108. [Google Scholar] [CrossRef]
  34. Story, B.H. Physiologically-based Speech Simulation Using an Enhanced Wave-Reflection Model of the Vocal Tract. Ph.D. Dissertation, University of Iowa, Iowa City, IA, USA, 1995. [Google Scholar]
  35. Palaparthi, A. Computational Motor Learning and Control of the Vocal Source for Voice Production. Ph.D. Dissertation, The University of Utah, Salt Lake City, UT, USA, 2021. [Google Scholar]
  36. Titze, I.R.; Palaparthi, A. Radiation efficiency for long-range vocal communication in mammals and birds. J. Acoust. Soc. Am. 2018, 143, 2813–2824. [Google Scholar] [CrossRef] [PubMed]
  37. Titze, I.R. Parameterization of the glottal area, glottal flow, and vocal fold contact area. J. Acoust. Soc. Am. 1984, 75, 572–580. [Google Scholar] [CrossRef]
  38. Belyk, M.; Pfordresher, Q.; Liotti, M.; Brown, S. The neural basis of vocal pitch imitation in humans. J. Cogn. Neurosci. 2016, 28, 621–635. [Google Scholar] [CrossRef]
  39. Psaltis, D.; Sideris, A.; Yamamura, A.A. A multilayered neural network controller. IEEE Control Syst. Mag. 1988, 8, 17–21. [Google Scholar] [CrossRef]
  40. Franklin, D.W.; Wolpert, D.M. Computational mechanisms of sensorimotor control. Neuron 2011, 72, 425–442. [Google Scholar] [CrossRef] [PubMed]
  41. Haggins, M.B.; Carney, A.E.; Schulte, L. Physiological assessment of speech and voice production of adults with hearing loss. J. Speech Hear. Res. 1994, 37, 510–521. [Google Scholar] [CrossRef] [PubMed]
  42. Omori, K.; Slavit, D.H.; Kacker, A.; Blaugrund, S.M. Quantitative videostroboscopic measurement of glottal gap and vocal function: An analysis of thyroplasty type I. Ann. Otol. Rhinol. Laryngol. 1996, 105, 280–285. [Google Scholar] [CrossRef]
  43. Atlun, H.; Bilgil, A.; Fidan, B.C. Treatment of multi-dimensional data to enhance neural network estimators in regression problems. Expert Syst. Appl. 2007, 32, 599–605. [Google Scholar]
  44. Bowden, G.J.; Dandy, G.C.; Maier, H.R. Data transformation for neural network models in water resources applications. J. Hydroinform. 2003, 5, 245–258. [Google Scholar] [CrossRef]
  45. Houk, J.C.; Wise, S.P. Distributed modular architectures linking basal ganglia, cerebellum, and cerebral cortex. Their role in planning and controlling action. Cereb. Cortex 1995, 5, 95–110. [Google Scholar] [CrossRef]
  46. Ostry, D.; Gribble, P.L. Sensory plasticity in human motor learning. Trends Neurosci. 2016, 39, 114–123. [Google Scholar] [CrossRef]
  47. Nuchkrua, T.; Leephakpreeda, T. Novel compliant control of a pneumatic artificial muscle driven by hydrogen pressure under a varying environment. IEEE Trans. Ind. Electron. 2022, 69, 7120–7129. [Google Scholar] [CrossRef]
  48. Kannampalli, T.; Smyth, J.M.; Jones, S.; Payne, P.R.O.; Ma, J. Cognitive plausibility in voice-based AI health counselors. Npj Digit. Med. 2020, 3, 72. [Google Scholar] [CrossRef]
Figure 1. Neural Control System for the control of LeTalker.
Figure 1. Neural Control System for the control of LeTalker.
Applsci 14 00769 g001
Figure 2. (A) The layered structure of the right vocal fold, (B) three-mass model approximation of the body-cover schema of the vocal folds.
Figure 2. (A) The layered structure of the right vocal fold, (B) three-mass model approximation of the body-cover schema of the vocal folds.
Applsci 14 00769 g002
Figure 3. Example waveforms generated by LeTalker. (Top) radiated pressure, (Middle) glottal flow, and (Bottom) glottal area signals.
Figure 3. Example waveforms generated by LeTalker. (Top) radiated pressure, (Middle) glottal flow, and (Bottom) glottal area signals.
Applsci 14 00769 g003
Figure 4. Structure of the neural networks used in the architecture designed for the control of LeTalker.
Figure 4. Structure of the neural networks used in the architecture designed for the control of LeTalker.
Applsci 14 00769 g004
Figure 5. Cost function used to train PtoA NN as a function of epoch number.
Figure 5. Cost function used to train PtoA NN as a function of epoch number.
Applsci 14 00769 g005
Figure 6. Boxplots showing error in % for the four acoustic features from the PtoA NN test data.
Figure 6. Boxplots showing error in % for the four acoustic features from the PtoA NN test data.
Applsci 14 00769 g006
Figure 7. Difference between targets and predictions for the four acoustic features from PtoA NN test data in actual units.
Figure 7. Difference between targets and predictions for the four acoustic features from PtoA NN test data in actual units.
Applsci 14 00769 g007
Figure 8. Cost value used to train the control system as a function of sample (epoch) number.
Figure 8. Cost value used to train the control system as a function of sample (epoch) number.
Applsci 14 00769 g008
Figure 9. Comparison between targets (left) and LeTalker predictions (right) from the test set. (First row) A typical signal in the speech range. (Second row) A signal with high fo. (Third row) A signal with high SPL. (Fourth row) A signal with low spectral centroid. (Fifth row) A signal with higher noise.
Figure 9. Comparison between targets (left) and LeTalker predictions (right) from the test set. (First row) A typical signal in the speech range. (Second row) A signal with high fo. (Third row) A signal with high SPL. (Fourth row) A signal with low spectral centroid. (Fifth row) A signal with higher noise.
Applsci 14 00769 g009
Figure 10. Boxplots showing error in % for the twelve features/parameters from the control system test data.
Figure 10. Boxplots showing error in % for the twelve features/parameters from the control system test data.
Applsci 14 00769 g010
Figure 11. Difference between targets and predictions for (Top row) somatosensory features, and (Bottom row) acoustic features.
Figure 11. Difference between targets and predictions for (Top row) somatosensory features, and (Bottom row) acoustic features.
Applsci 14 00769 g011
Figure 12. Boxplots of each controller outputs across all the test data. Here, affc—acoustic feedforward controller, sfbc—somatosensory feedback controller, and afbc—acoustic feedback controller.
Figure 12. Boxplots of each controller outputs across all the test data. Here, affc—acoustic feedforward controller, sfbc—somatosensory feedback controller, and afbc—acoustic feedback controller.
Applsci 14 00769 g012
Table 1. Parameter values used to compute fiber stress in the three vocal fold layers.
Table 1. Parameter values used to compute fiber stress in the three vocal fold layers.
ParameterMucosaVocal LigamentTA Muscle
ϵ 1 −0.5−0.5−0.5
ϵ 2 −0.350−0.05
σ 0 kPa 0.50.41.0
σ 2 kPa 30.01.391.5
C 4.4176.5
Table 2. Control, somatosensory, and acoustic targets and their corresponding predictions for the example signals shown in Figure 9. The control parameters were normalized. Vocal fold length has units of cm, the three fiber stress values have units of dyn/cm2 converted into a logarithmic scale. For the acoustic features, fo has units of Hz, SPL is in dB, NSC has no units, and SNR has units of dB.
Table 2. Control, somatosensory, and acoustic targets and their corresponding predictions for the example signals shown in Figure 9. The control parameters were normalized. Vocal fold length has units of cm, the three fiber stress values have units of dyn/cm2 converted into a logarithmic scale. For the acoustic features, fo has units of Hz, SPL is in dB, NSC has no units, and SNR has units of dB.
Control ParametersSomatosensory FeaturesAcoustic Features
Row PL a C T a T A a L C Lg σ m u c σ l i g σ m u s foSPLNSCSNR
FirstTargets0.200.190.500.461.4812.78.1312.91299210.5−3.12
Predictions0.220.260.620.441.5112.98.1813.21289211.3−4.6
SecondTargets0.60.90.460.192.2615.816.513.53611045.3−0.29
Predictions0.630.950.450.412.2415.7516.413.53241014.253.1
ThirdTargets0.630.480.090.451.8814.512.411.812310613.9−2
Predictions0.650.440.110.471.8414.411.811.912710613.3−1.6
FourthTargets0.580.980.0070.612.3416.117.412.92691003.813.2
Predictions0.570.960.00.432.3816.217.813.0229953.510.9
FifthTargets0.630.010.0050.451.4712.78.19.47810526.2−7.3
Predictions0.680.040.170.491.4212.48.011.89610520.7−6.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Palaparthi, A.; Alluri, R.K.; Titze, I.R. Deep Learning for Neuromuscular Control of Vocal Source for Voice Production. Appl. Sci. 2024, 14, 769. https://0-doi-org.brum.beds.ac.uk/10.3390/app14020769

AMA Style

Palaparthi A, Alluri RK, Titze IR. Deep Learning for Neuromuscular Control of Vocal Source for Voice Production. Applied Sciences. 2024; 14(2):769. https://0-doi-org.brum.beds.ac.uk/10.3390/app14020769

Chicago/Turabian Style

Palaparthi, Anil, Rishi K. Alluri, and Ingo R. Titze. 2024. "Deep Learning for Neuromuscular Control of Vocal Source for Voice Production" Applied Sciences 14, no. 2: 769. https://0-doi-org.brum.beds.ac.uk/10.3390/app14020769

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop