#PAGE_PARAMS# #ADS_HEAD_SCRIPTS# #MICRODATA#

Fuzzy jump wavelet neural network based on rule induction for dynamic nonlinear system identification with real data applications


Authors: Mohsen Kharazihai Isfahani aff001;  Maryam Zekri aff001;  Hamid Reza Marateb aff002;  Miguel Angel Mañanas aff003
Authors place of work: Department of Electrical and Computer Engineering, Isfahan University of Technology, Isfahan, Iran aff001;  Biomedical Engineering Department, Engineering Faculty, University of Isfahan, Isfahan, Iran aff002;  Biomedical Engineering Research Centre (CREB), Automatic Control Department (ESAII) Universitat Politècnica de Catalunya-Barcelona Tech (UPC), Barcelona, Spain aff003;  Biomedical Research Networking Center in Bioengineering, Biomaterials, and Nanomedicine (CIBER-BBN), Spain aff004
Published in the journal: PLoS ONE 14(12)
Category: Research Article
doi: https://doi.org/10.1371/journal.pone.0224075

Summary

Aim

Fuzzy wavelet neural network (FWNN) has proven to be a promising strategy in the identification of nonlinear systems. The network considers both global and local properties, deals with imprecision present in sensory data, leading to desired precisions. In this paper, we proposed a new FWNN model nominated “Fuzzy Jump Wavelet Neural Network” (FJWNN) for identifying dynamic nonlinear-linear systems, especially in practical applications.

Methods

The proposed FJWNN is a fuzzy neural network model of the Takagi-Sugeno-Kang type whose consequent part of fuzzy rules is a linear combination of input regressors and dominant wavelet neurons as a sub-jump wavelet neural network. Each fuzzy rule can locally model both linear and nonlinear properties of a system. The linear relationship between the inputs and the output is learned by neurons with linear activation functions, whereas the nonlinear relationship is locally modeled by wavelet neurons. Orthogonal least square (OLS) method and genetic algorithm (GA) are respectively used to purify the wavelets for each sub-JWNN. In this paper, fuzzy rule induction improves the structure of the proposed model leading to less fuzzy rules, inputs of each fuzzy rule and model parameters. The real-world gas furnace and the real electromyographic (EMG) signal modeling problem are employed in our study. In the same vein, piecewise single variable function approximation, nonlinear dynamic system modeling, and Mackey–Glass time series prediction, ratify this method superiority. The proposed FJWNN model is compared with the state-of-the-art models based on some performance indices such as RMSE, RRSE, Rel ERR%, and VAF%.

Results

The proposed FJWNN model yielded the following results: RRSE (mean±std) of 10e-5±6e-5 for piecewise single-variable function approximation, RMSE (mean±std) of 2.6–4±2.6e-4 for the first nonlinear dynamic system modelling, RRSE (mean±std) of 1.59e-3±0.42e-3 for Mackey–Glass time series prediction, RMSE of 0.3421 for gas furnace modelling and VAF% (mean±std) of 98.24±0.71 for the EMG modelling of all trial signals, indicating a significant enhancement over previous methods.

Conclusions

The FJWNN demonstrated promising accuracy and generalization while moderating network complexity. This improvement is due to applying main useful wavelets in combination with linear regressors and using fuzzy rule induction. Compared to the state-of-the-art models, the proposed FJWNN yielded better performance and, therefore, can be considered a novel tool for nonlinear system identification.

Keywords:

Neurons – Neural networks – Nonlinear dynamics – Nonlinear systems – Torque – Electromyography – Approximation methods – Cell signaling structures

Introduction

System identification is a challenging work in the many fields of engineering, which is concerned with achieving the model of dynamic nonlinear or linear systems based on the input and output observations, especially from experimental data with prior knowledge or inadequate information [1]. In recent years, many studies have been conducted using the combination of computational intelligence methods for nonlinear dynamic system modeling, function approximation, and time-series prediction [29].

Neural networks (NN) and fuzzy systems as computational intelligence methods are suitable tools for modeling expert knowledge and dealing with uncertain nonlinear processes or time series [8]. Incorporating NNs, wavelets, and fuzzy inference systems offer sophisticated solutions, named fuzzy wavelet neural networks (FWNN) [39]. The FWNN employs the learning ability of neural networks, time-frequency localization property of wavelets, and approximate reasoning characteristics of fuzzy systems to present a practical model handling uncertainty and disturbances in real data for complex hybrid nonlinear-linear problems. Hence, FWNNs require lower training time and have fewer rules and higher efficiency than fuzzy systems or neural networks [10].

FWNN model is a traditional Takagi-Sugeno-Kang neuro-fuzzy system in which the consequent part of fuzzy rules is replaced by a wavelet neural network. The antecedent part of each fuzzy rule in the FWNN model divides the input space into local fuzzy regions, and its consequent part corresponds to a sub-wavelet neural network [11].

In [3], the nonlinear autoregressive moving average with exogenous inputs was identified by a dynamic time-delay fuzzy wavelet neural network. In [12], a fuzzy wavelet neural network was proposed for nonlinear function approximation. In that network, a sub-wavelet neural network consisting of single-scaling wavelets was introduced for the consequent part of each fuzzy rule. The adaptive type of the network was developed in [8] as a solution for controlling nonlinear affine systems. Moreover, to identify and regulate nonlinear systems, the summation form of multidimensional wavelet functions constituted the fuzzy rule consequent part of the FWNN [4]. In [7], FWNN model membership functions in the antecedent part were wavelets similar to the activation functions in the consequent part of fuzzy rules.

Despite various FWNN model advantageous, they cannot correctly deal with systems with both linear and nonlinear dynamics. Also, inappropriate regulation of wavelet parameters reduces the generalizability of the model [13]. Also as another challenge of the FWNN, it is not a simple work to extract effective fuzzy rules [14]. The mentioned challenges are significant problems, especially in practical applications. For example, FWNN application in blood glucose concentration prediction in [15] has not worked so well in comparison with other types of neural networks. On the other hand, wavelet-based models for dynamical systems lead to a large number of neurons and time delay lines. This issue increases the complexity of the network [16].

Therefore, methods such as C-means clustering are used to enhance the FWNN structure [14, 17]. For example, in Fuzzy wavelet polynomial neural networks described in [17], the dominant selected wavelets were classified, and then each class was placed in the then-part of a fuzzy rule. In that study, C-means clustering was implemented in the antecedent of the fuzzy rules instead of Gaussian membership functions. In a similar example in [14], a self-adapted fuzzy C-means clustering was used to determine the number of fuzzy rules of the FWNN model. While promoting the structure of fuzzy rules by applying C-means clustering is sensitive to noise, not comprehensive and only affects the number of fuzzy rules.

This paper presents a new FWNN model named as fuzzy jump wavelet neural network (FJWNN) for identification dynamic nonlinear-linear systems. The proposed FJWNN provides modeling by using input regressors and their wavelet transform in the consequent part of each fuzzy rule. Consequently, each fuzzy rule can locally model both linear and nonlinear properties of a system. The linear relationship between the inputs and the output is learned by neurons with linear activation functions whereas nonlinear relationship is locally modeled by wavelet neurons. The OLS and GA methods are respectively applied to extract dominant wavelets exerting the most significant effect on the output. In selecting dominant wavelets, it is supposed that wavelet neurons are along with the linear combination of input regressors. Then, to optimize the proposed FJWNN structure including the number of fuzzy rules, effective inputs of each fuzzy rule, and model parameters, fuzzy rule induction is used. The performance of FJWNN in experimental applications is illustrated by applying the real-world Box-Jenkins gas furnace system and the electromyographic (EMG) signal modeling problem. Also, well-known benchmarks in function approximation, identification of dynamic nonlinear systems and time series prediction are studied to identify the ability of the proposed FJWNN model, in comparison with the state-of-the-art models. The main features of the FJWNN model proposed in this work are:

  • In our approach, it is possible to handle real data problems of large dimensions because the effective procure of choosing wavelets used in OLS method is not very sensitive to the input dimension. It is worth noting that the role of GA combined with OLS is on selecting the most influential wavelets. While in most reported neuro-fuzzy models such as [9, 18, 19], GA is determined to train unknown parameters.

  • The proposed FJWNN utilizes input regressors, and their wavelet transforms in the consequent part of each fuzzy rule. Consequently, wavelets with different scale values combined with input regressors under fuzzy rules are fully used to capture different global or local behaviors of dynamic nonlinear-linear systems.

  • By applying fuzzy rule induction [20], it is possible to assign a weight to each fuzzy rule, which determines the importance of each fuzzy rule. Consequently, when defining a proper threshold, insignificant rule is removed and it leads to a more straightforward structure. Also, in comparison with [14, 17], in our research, fuzzy rule induction prunes unnecessary inputs from each of the fuzzy rules.

This paper is organized as follows. The materials and methods section introduces benchmarks and presents the FJWNN model. FJWNN modeling results and their interpretation in comparison with the state-of-the-art models are provided in the results and discussion section. The final section provides some concluding remarks.

Materials and methods

Materials

The study materials consist of simulated benchmark examples and experimental problems. Function approximation, identification of dynamic nonlinear systems and time series prediction as machine learning problems and Box-Jenkins gas furnace system and the EMG signal modeling problem as input-output measurements of real datasets are taken into account in this study to identify the ability of the proposed FJWNN model, compared to the state-of-the-art models.

Simulated datasets

Multiple simulated benchmarks including a piecewise single variable function, five nonlinear dynamic plants with various nonlinear structures, and the chaotic Mackey Glass time series (with different signal to noise ratio (SNR) and various chaotic degrees) are considered to verify the effectiveness of the proposed FJWNN model. Comparing simulation results in the following examples, the conditions are considered similar to their corresponding references.

Example 1—Function approximation

The piecewise single variable function, which has been studied in the literature frequently [5, 6, 12], is described as follows:


The training data is composed of 200 input-output pairs uniformly sampled in the region [–10, 10], as mentioned in [5, 6, 12].

Example 2—Dynamic nonlinear system identification

Illustrating the ability of proposed approach in identification of dynamic nonlinear systems, five systems are considered in the following. In these examples, multiple different nonlinearity structures are used.

This example (Example 2–1) is a nonlinear dynamic system defined, which has been studied in the literature frequently [4, 11, 2123], as follows:


The input training signal is an independent and identically distributed uniform sequence over [–2,2] for about half of the 900 time steps and a sinusoid given by 1.05sin(πk⁄45) for the remaining time. Also, the following input signal is used as the test input signal:


In the following plants [2427], different structures including nonlinearity in relative to input and its delays, output delays or both of them are considered (Examples 2–2, 2–3, 2–4 and 2–5) respectively





For all these four plants, training input signal was taken from a 1000 time step uniformly distributed random signal over the interval [-1,+1] and test input signal is a 600 time step sinusoidal signal as given by


Example 3—Predicting chaotic time series

This simulation example is the Mackey-Glass time series, which is considered as a prediction problem used in [2833] as a benchmark example. The Mackey–Glass system was introduced as a white blood cell production model [34]. This time series is obtained from the following delay differential equation:


with x(0) = 1.2, τ = 17, and x(t) = 0 for t < 0 and regardless of the noise. 1200 input-output data points were generated while data pairs with t = 124 to 1123 were chosen for system identification. The first 500 points were used for training, and the remaining data were used for testing. Similar to [2833], the following input regressors were selected to identify the output x(t).


In following steps, the Mackey-Glass time series for other different values of tau (τ = 13, τ = 30 and τ = 100) and also corrupted data by uniformly-distributed stationary additive noise (SNR = 0 dB, SNR = 10 dB and SNR = 20 dB) are considered.

The real dataset

Example 4—Real-world Box-Jenkins gas furnace system

The Box-Jenkins system [35] is a complicated nonlinear system. The benchmark dataset consists of 296 input-output measurements of a real-world gas furnace process. The measurements as a time series data include the gas flow rate u(k) and the CO2 concentration y(k). In [1], y (k-1), y(k-2), y(k-3), u(k), u(k-1), u(k-2) are chosen as the model inputs. The number of 296 samples is equally divided into two parts which the first part is used for training, and the second is used for testing.

Example 5—EMG signal modeling problem

In this example, the EMG signal modeling is considered. This application is a necessary step in the load-sharing problem, a suitable solution for movement analysis, described in [3638]. This problem estimates the individual mechanical contribution of the muscles acting on the same joint based on their electromyography (EMG) and the total torque [39]. Data related to the physiological processes during muscle contraction is provided in surface electromyography (sEMG) envelops. Since the force produced by a particular muscle cannot be measured, it is usually estimated from sEMG envelops [40, 41]. The problem is the estimation of the torque exerted on a joint using the EMG signals of the contracted muscles.

In this study, the dataset provided in [36] was used. The inclusion criteria of the participants were no sign of the previous neuromuscular disorders. Also, the sampling method was convenience non-random sampling. Five healthy males (age: 21.3 ± 2.8 years; height 174.3 ± 2.6 cm; body mass 71.0 ± 3.4 kg) performed three series of flexion-extension force ramps. Each series lasted 25 s, and contrariwise (with n = 30, 50, 70) involved four isometric extension (e) and flexion (f) ramps from n% eMVC to n% fMVC. Data used in this study was obtained from a previous study [36], where written informed consent in accordance with the declaration of Helsinki was confirmed by each participant and the experimental protocol was approved by the ethical committee of the Politecnico di Torino.

Isometric voluntary flexions (extensions with elbows flexed at 90°) stored sEMG signals from the Biceps Brachii (BB), Brachioradialis (BR), and lateral and medial heads of the Triceps Brachii (TBL and TBM). The signals were acquired from BR, TBL, and TBM with three linear arrays of 8 electrodes (5-mm inter-electrode intervals). Moreover, an isometric brace, used for limb fixation, was used for measuring the torque signal. Then, the signal was amplified using Force Amplifier MISO-II (LISiN, Politecnico di Torino, Italy) and then sampled at 2048 Hz.

Single differential (SD) and double differential (DD) signals were calculated along with Fiber direction. A non-causal digital low-pass filter (1 Hz, 4th-order Butterworth filter) derived the envelope of sEMG signals from rectified signals. The spatial mean of the recorded signals of each muscle was considered the global envelope for the muscle. The mean number of data samples for each experiment was 796. The data was provided in Figshare (https://figshare.com/s/6c772ef829faf53240c0).

The proposed fuzzy jump neural network model

In this section, the new fuzzy wavelet neural network model intended for system identification is introduced. The overall schematic diagram of the proposed FJWNN structure is shown in Fig 1. The structure is based on sub-jump wavelet neural network (sub-JWNN), fuzzy inference, and rule induction.

(A) The sub-JWNN structure and (B) the proposed FJWNN model structure.
Fig. 1. (A) The sub-JWNN structure and (B) the proposed FJWNN model structure.
In this figure, where u1, u2, , um are inputs, φ1, φ2, , φn are the selected dominant wavelet neurons, W1,W2, ,Wn+m are weights of the sub-JWNN output layer, JWNN1, JWNN2, , JWNNna are na sub-JWNN made from n dominant selected wavelets, η1,η2, ,ηna are outputs of the na sub-JWNN, v1,v2,, vna are na weights of fuzzy rules, μ-1,μ-2,…,μ-na are the membership function values of each rule in FJWNN modeling, and PH is the prediction horizon.

First, we focus on the sub-JWNN and its structure, as depicted in Fig 1(A). The sub-JWNNs, linear combinations of input regressors and wavelet neurons reside in the consequent parts of fuzzy rules in the proposed FJWNN model. The wavelet neurons include wavelets most effective on the output selected from a lattice of wavelets. The lattice of wavelets consists of different wavelets generated from a mother wavelet whose scale and shift parameters change at intervals [42]. In this study, the single-scale multi-dimensional Mexican hat wavelet is used as the mother wavelet:


where U and m are the input regressor vector and its dimension. The scaled and shifted variant of the Mexican hat wavelet was obtained using the following equation:

where a is the scale parameter, B is the shift parameter vector. In the simulations of this paper, the values of scale parameters of the lattice of wavelets ranged from −4 to 4.

The main effective wavelets have the greatest impact on system modeling and hence are chosen from the wavelet lattice while in line with input linear regressors. This selection is initially made through the orthogonal least squares (OLS) method multiple times. Each time, a different number of main effective wavelets are selected, and the wavelets are combined with linear input regressors to model the training data. After choosing the best number of wavelets that minimized the root mean square error (RMSE) of the validation data, the genetic algorithm (GA) is used to search for different main effective wavelets. The best number is selected by replacing the initially selected wavelets with various wavelets from the wavelet lattice based on linear regressors, checking the validation RMSE, and choosing the best ones. The initial wavelets in the GA are main effective wavelets selected by OLS in the initial step. In our experiments, other GA parameters are selected as (population size = 1000, Generation steps = 40 and tolerance = 1e-5).

Having chosen the most effective wavelets, the wavelets are classified with the same scale parameter. For example, if ith group has ni wavelets with the same scale parameter ai, the output of the sub-JWNN is calculated using the following equation:


where ai is the scale parameter, Bi = [b1i, b2i, , bmi] is the vector of the shift parameters of the ith dominant wavelet and U = [u1, u2, , um] is the input regressor vector. Hence, the sub-JWNN was made of a linear combination of input regressors with wavelets with the same scale parameter chosen from the selected dominant wavelets.

The proposed FJWNN model structure, which is based on the sub-JWNN, is depicted in Fig 1(B). The structure is composed of different layers which models the input-output relation. In the first layer, the inputs u1, u2, , um are entered into the fuzzification layer. The fuzzification step includes na fuzzy rules (Rl, l = 1, , na) to produce the final output model.


where each fuzzy rule corresponds to a single-scale parameter sub-JWNN, na is the number of rules (equal to the number of unique scale parameters of the selected dominant wavelets), the AND operator is the multiplication, and Ajl are Gaussian fuzzy membership functions calculated as follows:

where mulj and sulj are mean and standard deviation of Gaussian fuzzy membership functions. The lth sub-JWNN has ml inputs, (nl +m) nodes in the hidden layer, and one output (ηl).

In this study, to simplify the proposed FJWNN model and reduce the number of its parameters, fuzzy rule induction is applied through the imperialist competitive algorithm (ICA) [43]. Fuzzy rule induction consists of optimizing the structures of the antecedent part of fuzzy rules and allocating a weight for each rule to differentiate between fuzzy rules of different significance.

The antecedent part of a fuzzy rule includes an input membership function per input. Each input has na membership functions. In this study, the optimization of the structure of the antecedent part of fuzzy rules means firstly doubting the role of all inputs in the antecedent part of all fuzzy rules, and secondly choosing the optimal membership function among the na possible membership functions for an effective input in the antecedent part of each fuzzy rule. In the fuzzy rule induction procedure, for each fuzzy rule, there is an input vector and a corresponding antecedent vector, which specifies how the input vector participates in any of the rules. The antecedent vector members {cai} can be 0, 1, …, na. The zero value implies that the corresponding input does not play any role in that rule antecedent part, and nonzero numbers refer to the corresponding input membership function.

In (6), to determine the firing strength of the lth fuzzy rule, the geometric mean of the membership functions of input variables which contribute in each rule antecedent is calculated, instead of just multiplying the functions in common neuro-fuzzy models.


where dl=∑i=1mcil and cil(l=1,2,…,na,i=1,2,…,m) are antecedent assignments represented as 0 or 1. The antecedent assignments are calculated as follows:

For weight assignment, a continuous weight vi (i = 1, 2, …, na) ranging from 0 to 1 is allocated to each of na fuzzy rules. This weight sets the significance of the rule in the proposed FJWNN model. Then, fuzzy rules with weights of smaller than a threshold are eliminated from the FJWNN model. Next, the defuzzification step is implemented, and the final output is calculated using the following equation:


where PH is the prediction horizon and

The unknown parameters of the FJWNN model include the mean and standard deviation parameters of Gaussian fuzzy membership functions and the weights of fuzzy rules adjusted using ICA and the weights of sub-JWNNs learned by the LS method [Table 1]. It should be noted that the scale and shift parameters of the dominant wavelets extracted by the OLS and the GA methods are fixed by ICA in the training phase. The dominant wavelets are selected from the wavelet lattice in the initial steps by the OLS and GA methods.

Tab. 1. Different methods used to train the unknown parameters of the proposed FJWNN model.
Different methods used to train the unknown parameters of the proposed FJWNN model.

ICA is a computational method that is used to solve optimization problems. This method does not need the gradient of the cost function in its optimization process [43]. In our experiments, ICA parameters are selected as (Maximum Number of Iterations = 60, Population Size = 500, Number of Empires/Imperialists = 5, Selection Pressure = 1, Assimilation Coefficient = 1.5, Revolution Probability = 0.5, Revolution Rate = 0.3, Colonies Mean Cost Coefficient = 0.2).

The flowchart of the proposed FJWNN modeling steps is presented in Fig 2.

The proposed FJWNN modeling flowchart.
Fig. 2. The proposed FJWNN modeling flowchart.

Performance metrics

The performance of the FJWNN model was evaluated in terms of the goodness-of-fit. Since the results of the multiple examples are compared with those of previous studies based on different performance metrics, the following performance metrics are first introduced: RMSE, root relative square error (RRSE), relative error (rel ERR%) and variance accounted for (VAF%).

The RMSE between the predicted (y hat) and measured output (y) is calculated using the following equation:


In addition to RMSE, three fair performance indexes were used (RRSE, Rel ERR%, and VAF %). RRSE provides the RMSE of the predictions relative to the standard deviation of the measured output, and is obtained as follows:


The Rel ERR% provides the RMSE of the predictions relative to the mean square of the measured output, and is calculated using the following equation:


The VAF% provides the RMSE of the predictions relative to the mean square of the measured output, and is defined as follows:


Results and discussion

In this section, the effectiveness of the proposed FJWNN model is evaluated using simulated and experimental examples. In examples 1–3, for each data set, ten independent runs are performed, and the mean and standard deviations (std) values of the accuracy metrics are calculated for both training and testing data. Correspondingly because of using rule induction method, for the number of rules (NOR) and parameters (NOP), in each case mean and std values are reported.

In this research, initial parameters of fuzzy membership functions are chosen, so that mean initial values are chosen randomly, and standard deviation initial values are chosen as 0.2. Initial fuzzy rules weights are set 0.7 and all initial antecedent parameters are set 1.

For our evaluation, we used a PC with Intel(R) Core(TM) i7-4700MQ CPU @ (2.40 GHz) and 8 GB RAM. All the methods were realized by MATLAB 7.12.

Example 1—Function approximation

In this example, the proposed FJWNN model is evaluated on the piecewise single variable function formulated as (1). Moreover, the comparison between the present results and those obtained using the state-of-the-art models is provided in Table 2.

Tab. 2. The performance of different approximation methods in Example 1.
The performance of different approximation methods in Example 1.

According to the results, the proposed FJWNN model showed better performance compared to other models. Fig 3 shows how the cost function value reduced during epochs in the training procedure.

Cost function trend in ICA training procedure of FJWNN model for Example1 for ten independent runs.
Fig. 3. Cost function trend in ICA training procedure of FJWNN model for Example1 for ten independent runs.

The original piecewise single variable function and the FJWNN model output for the test data are depicted in Fig 4.

Comparison between the original piecewise single variable function and its FJWNN model output for the test data for ten independent runs.
Fig. 4. Comparison between the original piecewise single variable function and its FJWNN model output for the test data for ten independent runs.

Example 2—Dynamic nonlinear system identification

For dynamic nonlinear system identification, in the present study, five plants with different nonlinearity structure described in (2)(6) are considered. For example, 2–1, in Table 3, the RMSE value of its FJWNN modeling is presented along with the corresponding values reported for recent models with the excitation signal as that in the present study. As can be seen in this table, the RMSE value of the FJWNN model is lower than those of other models.

Tab. 3. Comparison between the simulation results of different methods, for Example 2–1.
Comparison between the simulation results of different methods, for Example 2–1.

Fig 5 shows the dynamic nonlinear system output and the output of the FJWNN model with three rules. Also, the cost function during the ICA training of the proposed FJWNN model is presented in Fig 6.

Comparison between the dynamic nonlinear system Example 2–1 and its FJWNN model estimation for the test data for ten independent runs.
Fig. 5. Comparison between the dynamic nonlinear system Example 2–1 and its FJWNN model estimation for the test data for ten independent runs.
The cost function during ICA training of the proposed FJWNN model for Example 2–1 for ten independent runs.
Fig. 6. The cost function during ICA training of the proposed FJWNN model for Example 2–1 for ten independent runs.

According to Table 3, the proposed FJWNN produced more acceptable results than other analyzed methods.

In the following, modeling of the other four nonlinear dynamics (examples 2–2 up to 2–5) are considered as system identification problem. In each case, in the training phase, a uniformly distributed random 1000 time steps signal over the interval [–1,1] is applied to the plant. While in [25], the training procedure continued for 50000 iterations during training phase. The comparison with other research works is summarized in Table 4.

Tab. 4. Comparison between the simulation results of different methods for Examples 2–2 up to 2–5.
Comparison between the simulation results of different methods for Examples 2–2 up to 2–5.

The dynamic nonlinear examples and their FJWNN predictions for the test data are illustrated in Figs 7 up to 10.

Comparison between the Example 2–2 and its FJWNN model estimation for test data for ten independent runs.
Fig. 7. Comparison between the Example 2–2 and its FJWNN model estimation for test data for ten independent runs.
Comparison between the Example 2–3 and its FJWNN model estimation for test data for ten independent runs.
Fig. 8. Comparison between the Example 2–3 and its FJWNN model estimation for test data for ten independent runs.
Comparison between the Example 2–4 and its FJWNN model estimation for test data for ten independent runs.
Fig. 9. Comparison between the Example 2–4 and its FJWNN model estimation for test data for ten independent runs.
Comparison between the Example 2–5 and its FJWNN model estimation for test data for ten independent runs.
Fig. 10. Comparison between the Example 2–5 and its FJWNN model estimation for test data for ten independent runs.

Simulation results for nonlinear dynamic identification systems described in (2)(6) show that the proposed FJWNN leads to acceptable accuracy based on RMSE and RRSE metrics.

Example 3—Predicting chaotic time series

This example dealt with the prediction of the chaotic Mackey-Glass time series to compare the prediction ability of the proposed FJWNN model with those of previous models in the presence of different SNR. The Mackey-Glass time series and its FJWNN prediction for the test data are illustrated in Fig 11, and the prediction error is shown in Fig 12.

Comparison between the Mackey-Glass time series and its FJWNN prediction for the test data for ten independent runs.
Fig. 11. Comparison between the Mackey-Glass time series and its FJWNN prediction for the test data for ten independent runs.
Mean of the test data prediction error of the proposed FJWNN model for Example 3 for ten independent runs.
Fig. 12. Mean of the test data prediction error of the proposed FJWNN model for Example 3 for ten independent runs.

The results of the comparison between the present model and some recent fuzzy neural and wavelet models are presented in Table 5. This table shows that the RMSE obtained by the FJWNN prediction is smaller than those obtained using the state-of-the-art models.

Tab. 5. Comparison of different methods results for Example 3.
Comparison of different methods results for Example 3.

Chaotic degree (τ) from Mackey–Glass series is used as an uncertainty source and variated to compare the performance of the proposed FJWNN model with state-of-the-art models. Corresponding results are reported in Table 6. It shows FJWNN architecture performs better than ANFIS and IT2FNN when uncertainty degree is increased.

Tab. 6. The noise-free Mackey–Glass chaotic time-series prediction (τ = 13, 30, 100).
The noise-free Mackey–Glass chaotic time-series prediction (τ = 13, 30, 100).

In the next step, the Mackey-Glass time series are corrupted with noise levels of 0dB, 10 dB, and 20 dB of SNR (signal-to-noise ratio) as a high source of uncertainty. Table 7 shows that when different levels of noise are added, FJWNN model performs better than ANFIS and IT2FNN. This is because FJWNN model handles noise better because of choosing dominant wavelets in the proposed structure optimization procedure.

Tab. 7. Corrupted by uniformly-distributed stationary additive noise Mackey–Glass chaotic time series prediction (τ = 17).
Corrupted by uniformly-distributed stationary additive noise Mackey–Glass chaotic time series prediction (τ = 17).

Furthermore, the proposed FJWNN consists of fuzzy rules with different weights as their importance optimized in the training procedure to make fuzzy rule induction plausible. During the training procedure, the fuzzy rules with weights of smaller than a given threshold (0.05) were eliminated from the FJWNN model by setting their weights to zero, as for one of FJWNN models shown in Table 8.

Tab. 8. Different weights of the FJWNN model for Example 3.
Different weights of the FJWNN model for Example 3.

Example 4—Real-world Box-Jenkins gas furnace system

In this real application, the proposed FJWNN model is evaluated on the gas furnace benchmark dataset. Moreover, the comparison between the present results and those obtained using the state-of-the-art models is provided in Table 9.

Tab. 9. Comparison of different methods results for Example 4.
Comparison of different methods results for Example 4.

The gas furnace output and its FJWNN prediction for the test data are illustrated in Fig 13. To compare these results with existing models that have been applied to the same process, the root mean square error (RMSE) is used. The comparison results in Table 9 indicate that the FJWNN model can achieve higher accuracies.

Comparison between the gas furnace output and its FJWNN prediction for the test data.
Fig. 13. Comparison between the gas furnace output and its FJWNN prediction for the test data.

Example 5—EMG signal modeling problem

In this example, the proposed FJWNN model was used to model EMG torques using experimental data. Fig 14. presents recorded and estimated torques and sEMG envelopes for the third participant. Considering this participant, an epoch of 47 sec was used for training, and the remaining time was used for testing the proposed FJWNN. As shown in Fig 14, the estimated torque signal followed the measured signal very well (Rel Err% = 7.13).

50% MVC elbow flexion-extension isometric ramps of the third participant.
Fig. 14. 50% MVC elbow flexion-extension isometric ramps of the third participant.
(A) recorded torque (solid blue) and estimated torque (dotted red) for each muscle (upper), and (B) the corresponding SD sEMG envelopes (bottom). The two sections of sEMG envelopes and recorded torque separated by thick black lines were used to train and test correspondingly.

Table 10 provides the results of Rel Err% on SD records for 5 participants during elbow flexion-extension isometric ramps at 30%, 50%, and 70% MVC in comparison with the results obtained by [36].

Tab. 10. The rel ERR% results of FJWNN model estimation on SD records obtained through elbow flexion-extension isometric ramps (test dataset) for five participants at 30%, 50%, and 70% maximum voluntary contractions (MVC).
The rel ERR% results of FJWNN model estimation on SD records obtained through elbow flexion-extension isometric ramps (test dataset) for five participants at 30%, 50%, and 70% maximum voluntary contractions (MVC).

Table 11, on the other hand, presents the comparison between the VAF% results of the present study (in terms of mean±std) and those obtained in [38, 52].

Tab. 11. The VAF% (mean±std) results of FJWNN modeling on SD records obtained through elbow flexion-extension isometric ramps (test dataset) for all participants at 30%, 50%, and 70% MVCs.
The VAF% (mean±std) results of FJWNN modeling on SD records obtained through elbow flexion-extension isometric ramps (test dataset) for all participants at 30%, 50%, and 70% MVCs.

Furthermore, the cross-checking results of the FJWNN model torque estimation on SD and DD records for elbow flexion-extension at 30% and 70% MVCs are presented in Table 12. The cross-checking test included training the proposed model by records for elbow flexion-extension at 50% MVC, and testing the model by the records for elbow flexion-extension at 30% and 70% MVCs.

Tab. 12. Cross-checking Rel Err% results of FJWNN model estimation in comparison with the results of [36] on SD and DD records obtained through elbow flexion-extension isometric ramps for five participants at 50% MVC (for model training) and 30% and 70% MVCs (for model testing).
Cross-checking Rel Err% results of FJWNN model estimation in comparison with the results of [<em class="ref">36</em>] on SD and DD records obtained through elbow flexion-extension isometric ramps for five participants at 50% MVC (for model training) and 30% and 70% MVCs (for model testing).

An overview of the number of cross-modeling parameters is presented in Table 13. For proposed FJWNN model, the number of fuzzy rules, number of antecedent parameters and total number of parameters with and without using fuzzy rule induction are compared. As shown, the effect of using fuzzy rule induction in reducing the number of model parameters is observed.

Tab. 13. The number of rules (NOR), the number of fuzzy rule antecedent parameters (NOA), and the number of parameters (NOP) for FJWNN modeling on SD records in the cross-checking.
The number of rules (NOR), the number of fuzzy rule antecedent parameters (NOA), and the number of parameters (NOP) for FJWNN modeling on SD records in the cross-checking.

Fuzzy rule induction is also used to organize the antecedent parts of the fuzzy rules, reducing the number of parameters. For example, in Table 13, less unknown parameters are to be determined for FJWNN model for the antecedent parts of fuzzy rules. In the same case, if the generation of the Gaussian membership functions for a TSK neuro-fuzzy model with four input, one output, and an equal number of fuzzy rules is intended, then the number of model parameters is expected to be more as [38]. So the number of parameters is significantly decreased for the present FJWNN model proposed for EMG cross-modeling.

The proposed fuzzy model resulted in %VAF (mean ± std) = 98.24 ± 0.71 for all trial signals. The best performance of the model in [38] yielded the %VAF (mean ± std) of 96.40 ± 3.38. Thus, the proposed FJWNN model improved torque modeling results. Overall, there is an improvement in the reconstructed torque performance criterion of the proposed FJWNN, compared to those of the models in [36, 38, 52].

One of the limitations of the proposed FJWNN method is its running time during the learning procedure. For example, the average running time of its Matlab implementation for the examples 1, 2–1, 2–2, 2–3, 2–4, 2–5, 3, 4, and experimental EMG signals were 27±10, 77±12, 3±4, 23±1, 23±10, 1±1, 44±14, 4±1, and 11±5 (in minutes). This implementation, in its current form, is not thus suitable for online applications. The Vectorization packages with C++ implementation could be used to reduce the running time, which is the focus of our future work. The running time of the trained system on the test set is, however, acceptable. For example, the running time of analyzing the test set for the first example was 0.02±0.23 (in sec).

Now, a comparison of the performance of the proposed FJWNN with those of the state-of-the-art models was presented, evidencing the superior performance of the proposed model. For nonlinear systems, using models with fewer parameters lead to missing essential relationships in the data, while using models with many parameters makes parameter estimation difficult. However, in most cases, many model terms are redundant, and only a few significant terms with a specific accuracy are necessary. In the present study, the OLS and GA methods were used to select dominant wavelets which described nonlinear behavior, delivering a restricted number of wavelets in the sub-JWNN model. Accordingly, the number of wavelet classes decreased, and the structure of the proposed FJWNN model became noticeably simpler.

Moreover, choosing dominant wavelets using the OLS and GA methods can reduce the initial values of the cost function (See Figs 3 and 6). In the presented approach, fuzzy rule induction removes ineffective parameters or rules to simplify the proposed FJWNN model. Based on the overall analysis using the simulation and real data, it can be concluded that the proposed FJWNN model has high precision and less complexity, compared to the state-of-the-art models.

According to the performance of the FJWNN model for the real data in examples 4 and 5, the proposed FJWNN model employs the learning ability of neural networks, time-frequency localization property of wavelets and approximate reasoning characteristics of fuzzy systems to present the effective technique to deal with uncertainty and disturbances in real data for complex hybrid nonlinear-linear problems.

Conclusions

In this paper, FJWNN combined with rule induction as a new wavelet-based identification model was proposed for the identification of real data dynamic nonlinear-linear systems. In the proposed approach, OLS and GA methods are respectively used to choose dominant wavelets along with the linear combination of input regressors. Fuzzy rules including wavelets with various scale parameters (different resolutions) can capture different behaviors (global or local) of the systems. Then, by applying fuzzy rule induction and assigning a weight to each fuzzy rule, which determines the importance of each fuzzy rule, insignificant rule is removed. Also fuzzy rule induction prunes unnecessary inputs from each of fuzzy rules. The obtained results of simulation and experimental examples demonstrate that the proposed model is quite useful in dynamic nonlinear system identification. Overall, the proposed FJWNN model can be considered a promising tool for EMG‑Torque modeling. Possible future work will be the utilization of the FJWNN model in on-line identification of dynamic nonlinear systems with real data applications.


Zdroje

1. Feng S, Chen CLP. Nonlinear system identification using a simplified Fuzzy Broad Learning System: Stability analysis and a comparative study. Neurocomputing. 2019;337:274–86. doi: 10.1016/j.neucom.2019.01.073

2. Labati RD, Genovese A, Muñoz E, Piuri V, Scotti F. Applications of computational intelligence in industrial and environmental scenarios. In: Sgurev V, Piuri V, Jotsov V, editors. Learning Systems: From Theory to Practice. 756: Springer; 2018. p. 29–46.

3. Jiang X, Adeli H. Dynamic Wavelet Neural Network for Nonlinear Identification of Highrise Buildings. Computer-Aided Civil and Infrastructure Engineering. 2005;20(5):316–30. doi: 10.1111/j.1467-8667.2005.00399.x

4. Abiyev RH, Kaynak O. Fuzzy Wavelet Neural Networks for Identification and Control of Dynamic Plants—A Novel Structure and a Comparative Study. IEEE Transactions on Industrial Electronics. 2008;55(8):3133–40. doi: 10.1109/TIE.2008.924018

5. Singh M, Srivastava S, Hanmandlu M, Gupta JRP. Type-2 fuzzy wavelet networks (T2FWN) for system identification using fuzzy differential and Lyapunov stability algorithm. Applied Soft Computing. 2009;9(3):977–89. doi: 10.1016/j.asoc.2008.03.017

6. Tzeng S-T. Design of fuzzy wavelet neural networks using the GA approach for function approximation and system identification. Fuzzy Sets and Systems. 2010;161(19):2585–96. doi: 10.1016/j.fss.2010.06.002

7. Bodyanskiy Y, Vynokurova O. Hybrid adaptive wavelet-neuro-fuzzy system for chaotic time series identification. Information Sciences. 2013;220:170–9. doi: 10.1016/j.ins.2012.07.044

8. Zekri M, Sadri S, Sheikholeslam F. Adaptive fuzzy wavelet network control design for nonlinear systems. Fuzzy Sets and Systems. 2008;159(20):2668–95. doi: 10.1016/j.fss.2008.02.008

9. Huang M, Zhang T, Ruan J, Chen X. A New Efficient Hybrid Intelligent Model for Biodegradation Process of DMP with Fuzzy Wavelet Neural Networks. Scientific Reports. 2017;7:41239. doi: 10.1038/srep41239 28120889

10. Ganjefar S, Tofighi M. Single-hidden-layer fuzzy recurrent wavelet neural network: Applications to function approximation and system identification. Information Sciences. 2015;294:269–85. doi: 10.1016/j.ins.2014.09.054

11. Cheng R, Bai Y. A novel approach to fuzzy wavelet neural network modeling and optimization. International Journal of Electrical Power & Energy Systems. 2015;64:671–8. doi: 10.1016/j.ijepes.2014.07.067

12. Ho DWC, Ping-Au Z, Jinhua X. Fuzzy wavelet networks for function learning. IEEE Transactions on Fuzzy Systems. 2001;9(1):200–11. doi: 10.1109/91.917126

13. Ong P, Zainuddin Z. Calibrating wavelet neural networks by distance orientation similarity fuzzy C-means for approximation problems. Applied Soft Computing. 2016;42:156–66. doi: 10.1016/j.asoc.2016.01.042

14. Huang M, Tian D, Liu H, Zhang C, Yi X, Cai J, et al. A Hybrid Fuzzy Wavelet Neural Network Model with Self-Adapted Fuzzy -Means Clustering and Genetic Algorithm for Water Quality Prediction in Rivers. Complexity. 2018;2018:11. doi: 10.1155/2018/8241342

15. Zarkogianni K, Mitsis K, Litsa E, Arredondo M-T, Ficο G, Fioravanti A, et al. Comparative assessment of glucose prediction models for patients with type 1 diabetes mellitus applying sensors for glucose and physical activity monitoring. Medical & Biological Engineering & Computing. 2015;53(12):1333–43. doi: 10.1007/s11517-015-1320-9 26049412

16. Kumar R, Srivastava S, Gupta JRP, Mohindru A. Self-recurrent wavelet neural network–based identification and adaptive predictive control of nonlinear dynamical systems. International Journal of Adaptive Control and Signal Processing. 2018;32(9):1326–58. doi: 10.1002/acs.2916

17. Huang W, Oh S, Pedrycz W. Fuzzy Wavelet Polynomial Neural Networks: Analysis and Design. IEEE Transactions on Fuzzy Systems. 2017;25(5):1329–41. doi: 10.1109/TFUZZ.2016.2612267

18. Huang M, Ma Y, Wan J, Chen X. A sensor-software based on a genetic algorithm-based neural fuzzy system for modeling and simulating a wastewater treatment process. Applied Soft Computing. 2015;27:1–10. doi: 10.1016/j.asoc.2014.10.034

19. Huang M, Han W, Wan J, Ma Y, Chen X. Multi-objective optimisation for design and operation of anaerobic digestion using GA-ANN and NSGA-II. Journal of Chemical Technology & Biotechnology. 2016;91(1):226–33. doi: 10.1002/jctb.4568

20. Abraham A, Grosan C, Ramos V. Swarm Intelligence in Data Mining (Studies in Computational Intelligence): Springer-Verlag; 2006.

21. Chia-Feng J, Chin-Teng L. A recurrent self-organizing neural fuzzy inference network. IEEE Transactions on Neural Networks. 1999;10(4):828–45. doi: 10.1109/72.774232 18252581

22. Chia-Feng J. A TSK-type recurrent fuzzy network for dynamic systems processing by neural network and genetic algorithms. IEEE Transactions on Fuzzy Systems. 2002;10(2):155–70. doi: 10.1109/91.995118

23. Elman JL. Finding structure in time. Cognitive Science. 1990;14(2):179–211. doi: 10.1016/0364-0213(90)90002-E

24. Madyastha RK, Aazhang B. An algorithm for training multilayer perceptrons for data classification and function interpolation. IEEE Transactions on Circuits and Systems I: Fundamental Theory and Applications. 1994;41(12):866–75. doi: 10.1109/81.340848

25. Narendra KS, Parthasarathy K. Identification and control of dynamical systems using neural networks. IEEE Transactions on Neural Networks. 1990;1(1):4–27. doi: 10.1109/72.80202 18282820

26. Patra JC, Kot AC. Nonlinear dynamic system identification using Chebyshev functional link artificial neural networks. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics). 2002;32(4):505–11. doi: 10.1109/TSMCB.2002.1018769 18238146

27. Zhao H, Zhang J. Nonlinear dynamic system identification using pipelined functional link artificial recurrent neural network. Neurocomputing. 2009;72(13):3046–54. doi: 10.1016/j.neucom.2009.04.001

28. Castro JR, Castillo O, Melin P, Rodríguez-Díaz A. A hybrid learning algorithm for a class of interval type-2 fuzzy neural networks. Information Sciences. 2009;179(13):2175–93. doi: 10.1016/j.ins.2008.10.016

29. Tamura H, Koichi T, Hisasi T, Vairappan C, Zheng T, editors. Recurrent type ANFIS using local search technique for time series prediction. APCCAS 2008–2008 IEEE Asia Pacific Conference on Circuits and Systems; 2008 30 Nov.-3 Dec. 2008.

30. Jang JR. ANFIS: adaptive-network-based fuzzy inference system. IEEE Transactions on Systems, Man, and Cybernetics. 1993;23(3):665–85. doi: 10.1109/21.256541

31. Oysal Y, Yilmaz S. An adaptive wavelet network for function learning. Neural Computing and Applications. 2010;19(3):383–92. doi: 10.1007/s00521-009-0297-4

32. Oysal Y, Yilmaz S. An adaptive fuzzy wavelet network with gradient learning for nonlinear function approximation. Journal of Intelligent Systems. 2014;23(2):201–12. doi: 10.1515/jisys-2013-0068

33. Gholipour A, Araabi BN, Lucas C. Predicting chaotic time series using neural and neuro-fuzzy models: a comparative study. Neural Processing Letters. 2006;24(3):217–39. doi: 10.1007/s11063-006-9021-x

34. Mackey M, Glass L. Oscillation and chaos in physiological control systems. Science. 1977;197(4300):287–9. doi: 10.1126/science.267326 267326

35. Box GEP, Jenkins G. Time Series Analysis, Forecasting and Control: Holden-Day, Inc.; 1990. 500 p.

36. Botter A, Marateb HR, Afsharipour B, Merletti R, editors. Solving EMG-force relationship using Particle Swarm Optimization. 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society; 2011 30 Aug.-3 Sept. 2011.

37. Afsharipour B. Estimation of load sharing among muscles acting on the same joint and Applications of surface electromyography: Politecnico di Torino; 2014.

38. Jafari Z, Edrisi M, Marateb HR. An Electromyographic-driven Musculoskeletal Torque Model using Neuro-Fuzzy System Identification: A Case Study. Journal of medical signals and sensors. 2014;4(4):237–46. 25426427.

39. Kuriki HU, De Azevedo FM, Takahashi LSO, Mello EM, de Faria Negrão Filho R, Alves N. The relationship between electromyography and muscle force. EMG Methods for evaluating muscle and nerve function: InTech; 2012.

40. Kutz M. Biomedical engineering and design handbook. 2nd ed. ed. New York: McGraw-Hill; 2009.

41. Afsharipour BJIPdT. Estimation of load sharing among muscles acting on the same joint and Applications of surface electromyography. 2014.

42. Sadri AR, Zekri M, Sadri S, Gheissari N, Mokhtari M, Kolahdouzan F. Segmentation of Dermoscopy Images Using Wavelet Networks. IEEE Transactions on Biomedical Engineering. 2013;60(4):1134–41. doi: 10.1109/TBME.2012.2227478 23193305

43. Atashpaz-Gargari E, Lucas C, editors. Imperialist competitive algorithm: An algorithm for optimization inspired by imperialistic competition. 2007 IEEE Congress on Evolutionary Computation; 2007 25–28 Sept. 2007.

44. Abiyev RH, Kaynak O, Kayacan E. A type-2 fuzzy wavelet neural network for system identification and control. Journal of the Franklin Institute. 2013;350(7):1658–85. doi: 10.1016/j.jfranklin.2013.04.020

45. Karatepe E, Alcı M. A new approach to fuzzy wavelet system modeling. International Journal of Approximate Reasoning. 2005;40(3):302–22. doi: 10.1016/j.ijar.2005.06.003

46. Srivastava S, Singh M, Hanmandlu M, Jha AN. New fuzzy wavelet neural networks for system identification and control %J Appl. Soft Comput. 2005;6(1):1–17. doi: 10.1016/j.asoc.2004.10.001

47. Davanipoor M, Zekri M, Sheikholeslam F. Fuzzy Wavelet Neural Network With an Accelerated Hybrid Learning Algorithm. IEEE Transactions on Fuzzy Systems. 2012;20(3):463–70. doi: 10.1109/TFUZZ.2011.2175932

48. Zhao H, Gao S, He Z, Zeng X, Jin W, Li T. Identification of Nonlinear Dynamic System Using a Novel Recurrent Wavelet Neural Network Based on the Pipelined Architecture. IEEE Transactions on Industrial Electronics. 2014;61(8):4171–82. doi: 10.1109/TIE.2013.2288196

49. Han H-G, Ge L-M, Qiao J-F. An adaptive second order fuzzy neural network for nonlinear system modeling. Neurocomputing. 2016;214:837–47. doi: 10.1016/j.neucom.2016.07.003

50. Li C, Zhou J, Fu B, Kou P, Xiao J. T–S Fuzzy Model Identification With a Gravitational Search-Based Hyperplane Clustering Algorithm. IEEE Transactions on Fuzzy Systems. 2012;20(2):305–17. doi: 10.1109/TFUZZ.2011.2173693

51. Zhang R, Tao J. A Nonlinear Fuzzy Neural Network Modeling Approach Using an Improved Genetic Algorithm. IEEE Transactions on Industrial Electronics. 2018;65(7):5882–92. doi: 10.1109/TIE.2017.2777415

52. Clancy EA, Liu L, Liu P, Moyer DVZ. Identification of Constant-Posture EMG–Torque Relationship About the Elbow Using Nonlinear Dynamic Models. IEEE Transactions on Biomedical Engineering. 2012;59(1):205–12. doi: 10.1109/TBME.2011.2170423 21968709


Článek vyšel v časopise

PLOS One


2019 Číslo 12
Nejčtenější tento týden
Nejčtenější v tomto čísle
Kurzy

Zvyšte si kvalifikaci online z pohodlí domova

Svět praktické medicíny 1/2024 (znalostní test z časopisu)
nový kurz

Koncepce osteologické péče pro gynekology a praktické lékaře
Autoři: MUDr. František Šenk

Sekvenční léčba schizofrenie
Autoři: MUDr. Jana Hořínková

Hypertenze a hypercholesterolémie – synergický efekt léčby
Autoři: prof. MUDr. Hana Rosolová, DrSc.

Význam metforminu pro „udržitelnou“ terapii diabetu
Autoři: prof. MUDr. Milan Kvapil, CSc., MBA

Všechny kurzy
Kurzy Podcasty Doporučená témata Časopisy
Přihlášení
Zapomenuté heslo

Zadejte e-mailovou adresu, se kterou jste vytvářel(a) účet, budou Vám na ni zaslány informace k nastavení nového hesla.

Přihlášení

Nemáte účet?  Registrujte se

#ADS_BOTTOM_SCRIPTS#