This section will present the results of our research, starting with the results of the statistical analysis of the dataset, followed by the results of the regression.
3.1. Dataset Analysis
The descriptive statistics shown in
Table 6 show great variance between the datasets for individual fluidic muscles. This indicates non-transferability between models. As expected, the dataset where all data are combined tends to be the average of other values in the dataset, indicating that such a dataset may have a chance of predicting the values from all datasets; in other words, it is not skewed too greatly to either of the sets. It should be noted that pressure, as the output, was not used in some of the presented statistical analyses due to it only having a few possible discrete values, instead of it being a continuous variable.
By observing the correlation table given in
Figure 2, we can see that there is not a high correlation between any of the targeted output values. Correlations are especially low for the pressure output, looking at all possible inputs that may be extracted from the available data. This is possibly due to the discrete nature of the pressure collected from the datasheet.
The dataset distributions are important to note. Even with the large amount of data, skewness of the data from the normal distribution could indicate possible issues with the resulting models, and, as such, would warrant a cross-validation scheme. This skewness, towards the start of the dataset, is shown for all datasets in
Figure 3, showing that most of the data are given for shorter changes in length, which is apparent in all of the datasets, including the combined dataset.
The same is apparent in the datasets for force, as given in
Figure 4. While the individual datasets here are less drastic in skewness, the combined dataset practically follows an exponential distribution, with a vast majority of data points being present in the lower end of the data range. Due to this, the use of cross-validation is key to guarantee that the entirety of the dataset is tested.
The histograms show that more data are present in the lower end of the data range. While this is less expressed for contraction, it is very apparent for the combined force dataset (
Figure 4). This may adversely affect the precision of the created models at the higher end of the data range due to the lower amount of data, and it is safe to assume that most errors will result from this part of data being predicted.
3.2. Regression Results
The complete regression results, for all methods, fluidic muscles, and targets are given in
Table 7. The results are given as a mean across the training folds for both
and
, as well as
—the standard error across folds in both cases. The rows are split by the target—force, pressure, or length (contraction)—and are then further split per each of the available models of fluidic muscles: 5-100N, 10-100N, 20-200N, and 40-400N. With the overview of the table, we can conclude that some of the models achieved satisfactory results. For an easier look into the data, the results are presented as figures for each of the fluidic muscles individually later in this manuscript. Overall, the results for pressure seem to be lower than either of the two other outputs, possibly due to the discrete nature of the data.
For the DMSP-5-100N fluidic muscle, we present its results in
Figure 5. In the case of contraction models, the best results were achieved by the SVR, at
. Still, this model exhibited a large
of 4.47%. Significantly smaller
was shown by the MLP, but with a much lower
—demonstrating the need for multiple metrics. In the case of pressure models, the MLP model achieved the highest
of 0.85, with the lowest
of 0.01. The SVR achieved a similar
, but a significantly higher
. Finally, the best results regarding force were obtained with the ENet and MLP models, achieving an
of 0.86, with the MLP achieving a significantly lower
—0.15% versus 1.16%.
In the case of 10-100N fluidic muscle, the pressure was regressed with
of 0.92 and
of 0.14% using the MLP. Contraction showed much poorer results, with
scores indicating that the models barely converged. Interestingly, the
errors were low, a phenomenon possibly caused by the models outputting a low-variance output (possibly constant), which was close to the average value of the output that had a comparatively small range. This again indicates the need for multiple metrics being applied, since none of these models can be used for the prediction of values. Force output is somewhat better, with
values being approximately 0.80, with the lowest
of 0.2 being achieved for the MLP. These results are presented in
Figure 6.
In the case of 20-200N fluidic muscle, the results are given in
Figure 7. The only model that achieved somewhat good results for contraction was using the SVR—with an
of 0.84 and a
of 0.80%. Good results ere achieved for pressure with the XGB, SVR, and MLP models, achieving an
of approximately 0.92, with the MLP having the lowest
of 0.14%. Same behavior was seen regarding force, with the MLP obtaining an
of 0.85 and a
of 0.17.
Results for the 40-400N muscle are shown in
Figure 8. In the case of contraction, the SVR showed the best results, with an
of 0.93 and a
slightly below 1%. Similar results for pressure were obtained as before, with the MLP achieving the best results with an
and a
of 0.1%. Force was best predicted with the SVR, but taking into account a high variance in this model’s results, the MLP model was overall a better solution, with an
of 0.84 and a
of 0.1%.
We finally arrived at the final results of the manuscript.
Figure 9 demonstrates the results achieved on the combined dataset, which was the main goal of this study—testing the possibility of creating a generalized model of fluidic muscles. In order, starting with the contraction percentage, as shown in
Figure 9a, the overall best results were achieved with the XGB algorithm, which achieved an
score of 0.9 and a
of 0.00%. XGB was also the best performing algorithm in both metrics, achieving a mean
of 0.92 and a
of 0.10% for pressure. For force, the best general prediction was achieved with the MLP, with an
of 0.91 and a
of 1.46%.
The improved results in the combined dataset were possibly caused by the introduction of new variables, which would have been constants in the individual datasets. As shown in the correlation table (
Figure 2, these values do have a certain correlation with outputs, indicating that they may be responsible for improved scores, especially with the models based on XGB.