Support vector regression with genetic algorithms in forecasting tourism demand
Introduction
In the past few decades, the tourism industry has emerged as the fastest growing sector, and has spread widely around the world. Tourism expenditure has become an important source of economic activity, employment, tax revenue, income and foreign exchange. Therefore, every country needs to understand its international visitors and tourism receipts, to help formulate responsive policies on tourism quickly. Furthermore, the tourism industry is characterized by unstockpiled economics (such as empty hotel rooms and air flight seats), resulting in a requirement for accurate forecasts of tourism demand, in both the short and long term.
In tourism demand forecasting, the relevant researche models firstly emphasize econometrics (Hiemstra & Wong, 2002; Smeral, Witt, & Witt, 1992; Song & Witt, 2000). These investigations concluded that econometric models can help the policymakers formulate appropriate economic strategies in order to influence tourism demand and generate accurate demand forecasts. However, such methods are time-consuming, costly and the influent factors are difficult to research. Although time series methods cannot fully explain the relations between variables, they require less empirical evidence than econometrics and, hence, are appropriate for tourism research issues. For instance, Lim and McAleer (2002) employed Box–Jenkins’ autoregressive integrated moving average (ARIMA) model to forecast tourist arrivals to Australia from Hong Kong, Malaysia and Singapore. Goh and Law (2002) applied the time series SARIMA and MARIMA with interventions in forecasting tourism demand using ten arrival series for Hong Kong.
Recently, neural networks have been successfully employed for modelling time series. Unlike conventional statistical models, neural networks are data-driven, non-parametric, weak models which let “the data speak for themselves”. Law and Au (1999) first applied feed-forward neural networks to forecast Japanese demand for travel to Honk Kong. Law (2000) extended the applicability of neural networks in tourism demand forecasting by incorporating the back-propagation learning process into nonlinearly separable tourism demand data. Experimental results demonstrate that the neural networks model forecasts outperformed multiple regression, moving average and exponent smoothing. However, the neural networks suffer from several weaknesses, such as the need for a large number of controlling parameters, difficulty in obtaining a stable solution and the danger of over-fitting.
In 1995, Vapnik developed a neural network algorithm called support vector machine (SVM), which is a novel learning machine based on statistical learning theory, and which adheres to the principle of structural risk minimization seeking to minimize an upper bound of the generalization error, rather than minimize the training error (the principle followed by neural networks). This induction principle is based on the bounding of the generalization error by the sum of the training error and a confidence interval term depending on the Vapnik–Chervonenkis (VC) dimension. Based on this principle, SVM achieves an optimum network structure by striking a right balance between the empirical error and the VC-confidence interval. This balance eventually leads to better generalization performance than other neural network models (Tay & Cao, 2001). Additionally, the SVM training process is equivalent to solving linearly constrained quadratic programming problems, and the SVM-embedded solution meaning is unique, optimal and unlikely to generate local minima. Originally, SVM has been developed to solve pattern recognition problems. However, with the introduction of Vapnik's ε-insensitive loss function, SVM has been extended to solve nonlinear regression estimation problems, such as new techniques known as support vector regression (SVR), which have been shown to exhibit excellent performance (Vapnik, Golowich, & Smola, 1997). Recently, SVR has emerged as an alternative and powerful technique to solve the nonlinear regression problem. It has achieved great success in both academic and industrial platforms due to its many attractive features and promising generalization performance.
Despite its superior features, SVR is limited in academic research and industrial applications because the user must define various parameters (so-called hyper-parameters) appropriately. To construct the SVR model efficiently, SVR's parameters must be set carefully (Duan, Keerthi, & Poo, 2001; Keerthi, 2002; Lin, 2001). Inappropriate parameters in SVR lead to over-fitting or under-fitting (Lin, 2001). Different parameter settings can cause significant differences in performance. Therefore, selecting optimal hyper-parameter is an important step in SVR design. However, no general guidelines are available to select these parameters (Cristianini & Shawe-Taylor, 2000; Gunn, 1997; Schölkopf & Smola, 2002; Vapnik (1995), Vapnik (1998)). Therefore, in this study, we propose real-value genetic algorithms (RGA) to determine free parameters of SVR, known as GA-SVR, which optimizes all SVR's parameters simultaneously from the training data. Then, the tourism demand, as represented by the number of world-wide visitors to China, was predicted. The proposed approach was compared with the back-propagation neural networks (BPNN) and traditional time series models, such as ARIMA, so as to show that the SVR model is substantially featured with an excellent forecasting capacity.
This study includes eight sections. Section 2 introduces theories related to SVR. Section 3 summarizes the existing practical approaches to the choice of hyper-parameters. Section 4 elaborates on the GA-SVR model proposed in this study. Section 5 describes the data source and experimental settings. Section 6 analyzes the results of RGA and optimizes SVR's parameters, and also explains the determining parameters process of the BPNN and ARIMA models. Section 7 discusses and analyzes the experimental results. Section 8 concludes the study and suggests directions for future investigations.
Section snippets
Support vector regression
Recently, a regression version of SVM has emerged as an alternative and powerful technique to solve regression problems by introducting an alternative loss function. In the sequel, this version is referred to as SVR. Here, a brief description of SVR is given. Detailed descriptions of SVR can be found in Vapnik (1995), Vapnik (1998), Vapnik, Golowich, & Smola (1997), Schölkopf and Smola (2002) and Cristianini and Shawe-Taylor (2000).
The SVR formulation follows the principle of structural risk
Selection of SVR parameters
SVR has recently emerged as an alternative and highly effective means of solving the nonlinear regression problem. SVR has been quite successful in both academic and industrial platforms owing to its many attractive features and promising generalization performance. Some significant features of SVR are as follows: (i) it can model nonlinear relationships, (ii) the SVR training process is equivalent to solving linearly constrained quadratic programming problems, and the SVR embedded solution
GA-SVR optimization procedure
Generally, when selecting the parameters, most researchers still follow the trial-and-error procedure, first building a few SVR models based on different parameter sets, then testing them on a validation set to obtain optimal parameters. However, this procedure is time-consuming and requires some luck.
In contrast with optimizing SVR's parameters mentioned above, this study proposes a new method known as GA-SVR, which optimizes all SVR's parameters simultaneously. This model adopts RGA to seek
Model setting
For forecasting a univariate time series, the inputs of SVR are the past, lagged observations of the time series, and the outputs are the future values. Each set of input patterns is composed of any moving fixed-length window within the time series. Fig. 4 shows the basic architecture of SVR. The mapping function can be described as below:
In the above equation, yt is the observation of time t, and p is the dimension of the input vector (the number of input nodes) or the
Parameters determination of three models
In this study, determining the parameters in three models plays a significant role in obtaining good forecasting performance. Therefore, this section discusses the parameter determining process of three models.
Experimental results
After the RGA was applied to search for the optimal parameter sets, the SVR's forecasting models were built. The forecasting simulation was performed against the testing data. Table 5 and Fig. 6 compare the results obtained with the forecasting results from the ARIMA and BPNN models.
Table 5 shows that the forecasting performance from three various forecasting models are all appropriate. Generally, the results made by the GA-SVR models were superior to those from the other two models. MAPE and
Conclusions
This study applied SVR to the forecasting fields of tourism demand time series. To build stable and reliable forecasting models, the parameters of SVR must be specified carefully. Since the five-fold cross validation training error can be applied to the forecasting error estimation, the RGA, integrated using five-fold cross validation, was first applied to the training sets to obtain the optimal parameters. Thereafter, these optimal parameters were employed to build the actual GA-SVR
References (44)
- et al.
Modeling and forecasting tourism demand for arrivals with stochastic nonstationary seasonality and intervention
Tourism Management
(2002) Back-propagation learning in improving the accuracy of neural network-based tourism demand forecasting
Tourism Management
(2000)- et al.
A neural network model to forecast Japanese demand for travel to Hong Kong
Tourism Management
(1999) - et al.
Time series forecasts of international travel demand for Australia
Tourism Management
(2002) - et al.
Forecasting regional electric load based on recurrent support vector machines with genetic algorithms
Electric Power Systems Research
(2005) - et al.
Econometric forecasts: Tourism trends to 2000
Annals of Tourism Research
(1992) - et al.
Application of support vector machines in financial time series forecasting
Omega
(2001) - et al.
Evolutionary optimization: A tutorial
Trends in Analytical Chemistry
(1998) - et al.
Neural network forecasting of the British Pound/US Dollar exchange rate
Omega
(1998) - et al.
Time series analysis: Forecasting and control
(1976)
Learning from data: Concepts, theory, and methods
Feature selection by a genetic algorithm application to seed discrimination by artificial vision
Journal of Science: Food and Agriculture
An introduction to support vector machines and other kernel-based learning methods
Approximation by superpositions of a sigmoidal function
Mathematical Control Signal System
Simulated binary crossover for continuous search space
Complex Systems
A combined genetic adaptive search (geneAS) for engineering design
Computer Science and Informatics
Real-coded genetic algorithms with simulated binary crossover: Studies on multimodal and multiobjective problems
Complex Systems
Evaluation of simple performance measures for tuning SVM hyperparameters (Technical report)
Genetic algorithms in search, optimization and machine learning
Support vector machines for classification and regression
The elements of statistical learning. Data mining, inference and prediction
Cited by (309)
Forecasting daily tourism demand with multiple factors
2023, Annals of Tourism ResearchForecasting hourly attraction tourist volume with search engine and social media data for decision support
2023, Information Processing and ManagementIoT data analytics in dynamic environments: From an automated machine learning perspective
2022, Engineering Applications of Artificial IntelligenceArtificial intelligence in tourism and hospitality: Bibliometric analysis and research agenda
2022, International Journal of Hospitality ManagementHuman resource competency 4.0 and its impact on Bank Indonesia employees' readiness for transformational change
2022, Journal of Organizational Change ManagementUsing social media big data for tourist demand forecasting: A new machine learning analytical approach
2022, Journal of Digital Economy