Abstract
Keywords
Introduction
Case adaptation for parametric mechanical design
Parametric mechanical design1,2 is the key technology in realizing rapid development of mechanical product, but the massive parameters should be determined by designers in the process of complex machine design. Hence, case-based reasoning (CBR) methodology is proposed to figure out the solution parameter values of new mechanical product by referring the solution parameter values of existing cases in case base.3–6 So far, CBR has been employed successfully to design many mechanical products, such as micro-electromechanical device, 7 test turntable, 8 low power transformer, 9 extrusion die, 10 bearing device, 11 gear reducer 12 and welding fixture. 13 To obtain the feasible solution for new problem in CBR, finding a similar solution from existing case and further adapting this old solution are two main steps. Comparing with case retrieval, case adaptation is a more challenging issue, because in most situations, the old solution cannot be applied directly to the new problem.
Early CBR systems used in parametric mechanical design only focus on retrieving cases, and these systems commonly selected the solution of most similar case as the only candidate to be modified to satisfy the new problem.14–17 There are two main problems in classical case adaptation under 1-nearest neighbour (1-NN) principle: first, the adjustment of old solution values heavily depends on human subjective judgement, 18 and second, the solutions from other relative similar cases are ignored. Hence, the automatic case adaptation under K-nearest neighbour (KNN) principle, that is, using K (K > 1) similar cases to generate more accurate solution for new problem, is the greatest challenge for CBR researches.19,20
Early studies of CBR adaptation in the principle of KNN were based on the manual definition of adaptation rules, and engineers have to spend much effort on the acquisition of adaptation rules before adaptation process. 21 Thus, given a specific knowledge-light case adaptation method for CBR has significant practical implication. Statistical adaptation has been first proposed in the research of case adaptation since 1990s, such as the equal mean, 22 median, 23 weighted mean 24 and multivariate regression analysis (MRA). 25 The major shortcoming of statistical case adaptation is its relative low adaptation accuracy. Our previous studies26,27 have tried to extract various implicit knowledge hidden in similar case data to improve the performance of case adaptation based on weighted mean. Another way to overcome the shortcoming of statistical case adaptation is to perform case adaptation based on intelligent machine learning techniques, such as neural networks (NNs),28–30 decision tree 19 and hybrid method.31,32 These techniques have been introduced into CBR adaptation researches, which explore the utilization of inductive learning to acquire the differences between cases and their solutions, and apply the acquired knowledge to implement automatic case adaptation. 33 However, these enumerated intelligent case adaptation methods perform poorly when retrieved cases have a large number of attributes.
Support vector regression–based case adaptation
Recently, some scholars have paid attention to another intelligent algorithm, that is, support vector machine (SVM), because SVM shows its better performance in solving classification and regression problems. The formulation of SVM for solving regression problem is also called as support vector regression (SVR). Compared to other machine learning methods, SVR could generate a global optimal solution because of its optimum network structure and has better generalization performance for retrieved cases with many attributes.34,35 Hence, some studies have employed SVR to automatically perform CBR adaptation with various degrees of success.36–38
Despite the SVR has been employed preliminarily in CBR adaptation, the applications of SVR in CBR adaptation also expose a potential problem, namely, the standard formulation of SVR can only be used as a univariate modelling technique for CBR adaptation due to its inherent single-output structure.
39
Consequently, on one hand, SVR-based adaptation studies highlight the superiority of SVR for CBR adaptation. On the other hand, these studies have to construct different SVR-adaptation model for each solution feature separately, because of the inherent single-output structure of SVR, which would result in the ignorance of the possible interrelations among output solutions. If the parametric design case in case base has
In other research areas, to generalize the SVR from regression estimation and function approximation to multi-dimensional problems, Pérez-Cruz et al. 40 first proposed a multi-dimensional SVR (multiple-output support vector regression (MSVR)) that uses a cost function with a hyper-spherical intensive zone, capable of obtaining better predictions than using an SVR independently for each dimension. Subsequently, the new SVR algorithm has become a viable tool for solving the multiple-input, multiple-output regression problem, such as nonlinear channel estimation, 41 biophysical parameter evaluation, 42 converter gas tank level prediction, 43 dynamic load identification, 44 stock price index forecasting 45 and multiple-step-ahead time series prediction. 46 Although past studies have applied MSVR in various fields, as far as we know, there have been very few, if any, efforts to examine the feasibility of MSVR for CBR adaptation. So, we think it is worth to investigate the utilization of MSVR to conduct the adaptation engine, and apply MSVR-based adaptation method in actual mechanical design.
Motivation and originality of this research
As mentioned in section ‘Support vector regression–based case adaptation’, an alternative way to improve the performance of SVR-based adaptation is to utilize the MSVR. So, we have proposed an MSVR-based adaptation approach in our previous study.
47
We adopt the differential heuristic36–38,48 to build the training sample of MSVR. By using differential heuristic, the training sample consists of the differences between extracted case and retrieved case, rather than a single case in case base. As mentioned in section ‘Support vector regression–based case adaptation’, if the design case in case base has
Methodologies
Training sample generation
Before building the MSVR-SW-based CBR adaptation engine, a set of training sample needs to be produced beforehand. As mentioned in section ‘Support vector regression–based case adaptation’, if the design case in case base has
where
where
where
Similarity-related weight calculation
The key issue of similarity-based weight calculation is the similarity measurement (SM) of problem feature-values between extracted and retrieved cases. Referring to related SM studies,2,19,26,50 this article adopts multi-algorithm-oriented hybrid SM strategy to amplify the advantages of individual SM techniques and minimize their limitations. The hybrid SM strategy integrates four SM metrics, that is, Euclidian distance, Manhattan distance, Gaussian function and grey coefficient. The distance (Euclidian distance and Manhattan distance)-based SM metrics are in fact particular cases of the Minkowski measurement, expressed as follows
where
The third SM metric is SM based on Gaussian transformation function,
50
which transfers the differences between
where
The fourth SM method comes from the grey coefficient degree between
After the similarity between
where
MSVR-SW-based adaptation
Framework
The framework of multivariable CBR adaptation based on MSVR-SW is described in Figure 1. Sections ‘Training sample generation’ and ‘Similarity-related weight calculation’ have described the process of training sample and similarity-related weight generation. Training samples are used to conduct the MSVR model. Different from other MSVR researches, the proposed MSVR-SW introduces the

Framework of multivariable CBR adaptation based on MSVR-SW.
MSVR for case adaptation
Pérez-Cruz et al.
40
appointed out that ‘the key idea of MSVR is extending Vapnik
where
In equation (9) when
where
and
To optimize equation (10), an IRWLS procedure is constructed which linearly searched the next step solution along the descending direction based on the previous solution. According to the representer theorem,
51
the best solution to minimization of equation (10) can be expressed as
where
The proof of convergence of the above algorithm is given by Sánchez-Fernández et al.
41
Because
Example
Training sample and weight generation
Referring to our previous study,
37
this article also uses the power transformer design with
Adaptation engine construction
The parameter selection plays a crucial role on the adaptation engine construction. RBF
Meanwhile, Lee
56
suggested that an exponentially increasing sequence is a feasible method to obtain optimal parameters. Hence, the search spaces of

Parameter optimization results of MSVR-SW.
Comparison
Data description
We select 300 historical power transformer design cases from UCI Repository of ML Databases
57
as the raw data. The reasons for choosing such data are well-expressed parametrical case and moderate computational complexity. So, it is a representative data set for CBR adaptation researches.19,26,27,37,47 Furthermore, we would like to compare the performances between different SVRs with different number of inputs and outputs. Referring to Hu et al.,
27
we collect a total of 16 problem features and 10 solution features and build three datasets for this comparison according to different problem and solution features. Table 2 describes the selection of problem and solution features in three datasets. For dataset I, there are four problem features and four solution features. For dataset II, there are 10 problem features and 6 solution features, and all problem and solution features (sixteen problem features and ten solution features) are selected in dataset III. Besides, we intend to carry out adaptation processes using various
Three datasets with different problem and solution features. 27
To investigate the superiority of MSVR-SW in CBR adaptation, traditional adaptation methods are selected in this comparison. As mentioned in section ‘Case adaptation for parametric mechanical design’, the knowledge-light case adaptation falls into two categories: statistical adaptation and intelligent adaptation. One of the most used tools in intelligent adaptation is NNs. Therefore, two typical statistical methods, that is, MRA and median, and one NN, that is, back-propagation neural network (BPNN), are employed. In addition, classical SVRs, that is, standard SVR and MSVR, are also used as the comparative methods. So, there are a total of six examined methods, that is, MSVR-SW, MSVR, standard SVR, BPNN, MRA and median. Among them, MSVR-SW, MSVR-SW, MSVR and SVR are carried out using LIBSVM tool. 58 Meanwhile, BPNN and MRA are implemented using WEAK library 37 in this comparison.
Validation techniques
This comparison uses the hold-out methodology. In each dataset, two-thirds of the power transformer design cases are used as estimation samples, while reminders constitute the hold-out samples. Comparative adaptation methods were employed and trained on the estimation samples and produced the adaptations for the entire hold-out samples. The adaptation results were then compared to the hold-out sample set to evaluate the out-of-sample performance of each method. A five-fold cross-validation was used in the training phase to avoid over-fitting. To assess the adaptation abilities of the different methods, we compared the out-of-sample adaptations using two different approaches because it is generally impossible to specify an evaluation criterion that is universally acceptable. 45 First, we examined the adaptation accuracies of all comparative methods by calculating the mean absolute percentage error (MAPE), 46 as MAPE, rather than mean absolute error, can reflect the mean adaptation performance of all solution components. 37 The definition of MAPE is shown as follows
where

Process of comparative experiment.
Implementation of methodologies
Once the experimental dataset is set, the selected optimization method is employed for parameter space searching. Similar to MSVR-SW, the kernel functions of MSVR and SVR applied in this comparison experiment were RBF. Hence, the three hyper-parameters, namely,
where
Final parameters of SVR, MSVR and MSVR-SW in datasets I, II and III.
SVR: support vector regression; MSVR: multiple-output support vector regression; MSVR-SW: MSVR with similarity-related weight.
Results and discussion
Result
This section focuses on the out-of-sample adaptation abilities of MSVR-SW, MSVR, SVR, BPNN, MRA and median in terms of statistical accuracy and computational cost. Table 4 lists the mean MAPE values of examined methods on 50 times hold-out operations in three datasets to reflect the general adaptation performances, from which we can also abstract the change of accuracy of each method with increasing
MAPE values of six comparative methods with various
MAPE: mean absolute percentage error; MSVR-SW: MSVR with similarity-related weight; MSVR: multiple-output support vector regression; SVR: support vector regression; BPNN: back-propagation neural network; MRA: multivariate regression analysis.

Trends of adaptation accuracies of six comparative methods in (a) dataset I, (b) dataset II and (c) dataset III.
ANOVA test results for hold-out sample.
ANOVA: analysis of variance.
The mean difference among MSVR-SW, MSVR, SVR, BPNN, MRA and median is significant at the 0.05 level.
Comparison results with ranked comparative methods for hold-out sample.
SVR: support vector regression; MSVR: multiple-output support vector regression; MSVR-SW: MSVR with similarity-related weight; BPNN: back-propagation neural network; MRA: multivariate regression analysis.
The mean difference among MSVR-SW, MSVR, SVR, BPNN, MRA and median is significant at the 0.05 level.

Required time of six adaptation methods.
Comparison of adaptation accuracy
When the comparative results of adaptation accuracy achieved by six methods are presented in Table 4, we can deduce the following observations. Overall, by increasing the number of problem and solution features, the performances of CBR adaptation based on machine learning techniques (i.e. MSVR-SW, MSVR, SVR and BPNN) also increase, because of more abundant data to be trained. Meanwhile, the adaptation methods based on SVRs (MSVR-SW, MSVR and SVR) outperform classical adaptation methods based on BPNN and statistic methods in across the three datasets and six
Figure 4 concludes the changes of adaptation accuracies of adaptation methods with different
Comparison of adaptation difference
To compare the adaptation differences that exist among the six methods, we performed the ANOVA procedure in this experiment. All of the ANOVA results listed in Table 5 are significant at the 0.05 level, which means that there are significant differences among MSVR-SW, MSVR, SVR, BPNN, MRA and median. To further identify the significant difference between any two comparative methods, we also used Tukey’s HSD test to compare all pairwise adaptation differences at the 0.05 level. Table 6 summarizes the results of Tukey’s HSD test, from which we can find that when MSVR-SW with
Comparison of computation cost
It is important to note that the computational costs of six adaptation methods are different. From a practical viewpoint, the computational cost is an important and critical issue. Thus, the computational load of each method under
SVR and BPNN, which use problem and solution feature-values directly and output only one adapted solution values, are computationally much more expensive than multivariable CBR adaptation methods (MSVR-SW and MSVR), and the differences of computational costs are enlarging with increasing number of inputs and outputs. When comparing MSVR-SW with MSVR on computational cost index, the MSVR-SW is the winner, because MSVR-SW adopts differential strategy to generate the training samples from
Summation
According to the results of empirical comparisons, we can find that adaptation method using different number of inputs, outputs and retrieved cases could produce different adaptation results. In general, multivariable CBR adaptation with multiple output is more effective than univariate CBR adaptation with single output to handle the complicated case adaptation task from the views of the adaptation accuracy and computational cost. Furthermore, MSVR-SW improves the performance of MSVR-based adaptation by giving different weights to different training samples. Experiment results show that MSVR-SW is the best method among six comparative methods (MSVR-SW, MSVR, SVR, BPNN, MRA and median) across three datasets and six
Conclusion
In this article, because of its inherent single-output structure, standard SVR is incapable of handling the possible interrelations among solution outputs in CBR adaptation under
In future, we intend to investigate the parameter optimization of MSVR-SW for CBR adaptation. For example, in this comparison experiment, we just performed the coarse grid search on





