Pm2 5 prediction using genetic algorthm based feature selection and encoder decoder model = dự đoán chỉ số pm2 5 sử dụng học sâu và thuật toán di truyền

60 15 0
Pm2 5 prediction using genetic algorthm based feature selection and encoder decoder model = dự đoán chỉ số pm2 5 sử dụng học sâu và thuật toán di truyền

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

Thông tin tài liệu

Acknowledgements First of all, I would like to deeply thank my family, especially my parents who have worked hard to raise me My parents have always been with me and created the best conditions for me to have all the necessities needed for my studies Parents are the spiritual fulcrum, helping me to have a springboard to overcome difficulties and challenges I would like to express my gratitude to my advisors, Dr Nguyen Phi Le for supporting my studies and research on this subject She is very kindhearted and supportive person, who has guided me from the first day I worked with her Moreover, I would like to thank Dr Nguyen Thanh Hung, who has spent his precious time supporting, giving me advice and along with Dr Nguyen Phi Le, giving me opportunities to work in many amazing projects My sincere thanks also go to all the people in the ICN laboratory of the BK.AI center I have a wonderful time working with talented and special peers I learned a lot from them and they always spread positive energy for me Finally, I would like to thank my friends who have always stood by me, shared joys and sorrows, and always supported and helped me all the time Abstract The concentration of fine particulate matter (PM2.5), which represents inhalable particles with diameters of 2.5 micrometers and smaller, is a vital air quality index Such particles can penetrate deep into the human lungs and severely affect human health This paper studies accurate PM2.5 prediction, which can potentially contribute to reducing or avoiding the negative consequences Our approach’s novelty is to utilize the genetic algorithm (GA) and an encoder-decoder (E-D) model for PM2.5 prediction The GA benefits feature selection and remove outliers to enhance the prediction accuracy The encoder-decoder model with long short-term memory (LSTM), which relaxes the restrictions between the input and output of the model, can be used to effectively predict the PM2.5 concentration We evaluate the proposed model on air quality datasets from Hanoi and Taiwan The evaluation results show that our model achieves excellent performance By merely using the E-D model, we can obtain more accurate (up to 53.7%) predictions than those of previous works Moreover, the GA in our model has the advantage of obtaining the optimal feature combination for predicting the PM2.5 concentration By combining the GA-based feature selection algorithm and the E-D model, our proposed approach further improves the accuracy by at least 13.7% Content Content i List of Figures iii List of Tables iv List of Equations v Acronyms vi Introduction 1.1 PM2.5 forecasting problem 1.2 Existing solutions and problems 1.3 Goals and approaches 1.4 Structure of thesis Chapter Related works Chapter Theoretical Background 2.1 Artificial Intelligence 2.2 Deep learning overview 2.3 Long short-term memory 2.4 Encoder-Decoder model 10 2.5 Feature engineering 10 2.5.1 The importance of features 11 2.5.2 Feature extraction 11 2.5.3 Feature selection 12 2.5.4 Feature construction 13 2.6 Genetic algorithm 14 2.6.1 Overview 14 2.6.2 Genetic operators 16 2.6.3 Initialization 17 2.6.4 Crossover 17 2.6.5 Mutation 17 2.6.6 Selection 17 2.7 Research methods 18 Chapter Proposed Forecasting Framework (OFFGED) 19 3.1 Overview 19 i 3.2 GA-based feature selection 20 3.3 Encoder-Decoder model-based prediction 21 3.4 New training strategy LTS2 23 Chapter Performance Evaluation .27 4.1 Dataset and evaluation settings 27 4.2 Impact of the GA’s number of generations 29 4.3 Comparing feature selection algorithms 30 4.4 Comparing prediction models 33 4.4.1 Comparing ED-LSTM, AE-BiLSTM, and AC-LSTM 34 4.4.2 Comparing ED-LSTM and ST-DNN 39 4.5 LTS2 evaluation 42 4.6 Discussion 44 4.6.1 Results of OFFGED 44 4.6.2 Results of LTS2 46 Conclusion 47 Published papers 48 References 49 ii List of Figures Figure An example of an artificial neural network Figure Structure of RNN Figure Structure of the LSTM unit Figure The basic structure of the encoder-decoder model 10 Figure An example of feature extraction 12 Figure An example of feature selection 13 Figure An example of feature construction 14 Figure The basic structure of Genetic Algorithm 16 Figure Overview of the proposed model 19 Figure 10 Encoding a feature combination (the white and gray cells represent the selected feature encoded by and 0, respectively) 20 Figure 11 Illustration of the GA’s crossover and mutation operations 21 Figure 12 Structure of the LSTM-based encoder-decoder model 23 Figure 13 Notation of the proposed GA-based training mechanism 24 Figure 14 Training strategy – 𝑓𝑓𝑖𝑖𝑥𝑥𝑥𝑥𝑥𝑥, 𝑠𝑠ℎ𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢 = (𝑡𝑡𝑡𝑡𝑡𝑡𝑡𝑡, 𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓) 24 Figure 15 Training strategy – 𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓, 𝑠𝑠ℎ𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢 = (𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓, 𝑓𝑓𝑓𝑓𝑙𝑙𝑙𝑙𝑙𝑙) 25 Figure 16 Training strategy – 𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓, 𝑠𝑠ℎ𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢 = (𝑡𝑡𝑡𝑡𝑡𝑡𝑡𝑡, 𝑡𝑡𝑡𝑡𝑡𝑡𝑡𝑡) 25 Figure 17 Training strategy – 𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓, 𝑠𝑠ℎ𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢 = (𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓, 𝑡𝑡𝑡𝑡𝑢𝑢𝑢𝑢) 26 Figure 18 Impact of the number of generations 30 Figure 19 Comparison of feature selection algorithms 31 Figure 20 Comparison of GA-based feature selection and using all the features for the Hanoi dataset 33 Figure 21 Comparison between models using Hanoi dataset with all features 37 Figure 22 Comparison between models using Hanoi dataset with feature selected by GA 38 Figure 23 MAE of the proposed model with different output lengths 39 Figure 24 Comparison between models using Taiwan dataset with features selected by [11] 42 iii List of Tables Table Details of missing data in the datasets 28 Table Hyperparameter settings 29 Table ED-LSTM, AE-BiLSTM, and AC-LSTM use all features (Hanoi dataset) 34 Table ED-LSTM, AE-BiLSTM and AC-LSTM use selected features by GA (Hanoi dataset) 35 Table Comparing the MAE of the proposed ED-LSTM model and the ST-DNN model (using the features proposed by [11]) 40 Table Hyperparameters of training strategy 42 Table Training strategy for different cases 43 Table Comparing proposed method combining new training strategy with related works 44 Table Correlation of features 45 iv List of Equations Equation RNN equation Equation LSTM equation Equation LSTM decoder equation of the first time step 22 Equation LSTM decoder equation 22 Equation Prediction result of one time step 23 v Acronyms Abbreviations and terms Meaning LSTM Long-Short Term Memory RNN Recurrent Neural Network MAE Mean Absolute Error RMSE Root Mean Squared Error 𝒍𝒍 Sequence Length 𝒉𝒉 Horizon vi Introduction 1.1 PM2.5 forecasting problem Industrialization and urbanization have brought considerable convenience to human lives However, they are generally associated with severe air pollution Accordingly, people have raised concerns about air quality, especially near living areas Particulate matter 2.5 (PM2.5) is one of the most important indexes to evaluate the severity of air quality, which is directly related to human health PM2.5 particles in the air can bypass the nose and throat and penetrate deep into the lungs, causing many diseases, such as cardiovascular disease and respiratory disease In [1], the authors reveal that long-term exposure to PM2.5 may lead to heart attack and stroke Therefore, accurate PM2.5 forecasting is crucial and may help governments and citizens find suitable solutions to control or prevent negative conditions 1.2 Existing solutions and problems PM2.5 forecasting is a time series prediction problem that is commonly solved using recurrent neural networks (RNNs), including LSTM [2] The LSTM-based model has advantages in air quality prediction [3] In [4], the authors also use LSTM but combine gas and PM2.5 concentrations to predict air quality in Taiwan The work in [5] exploits deep learning to build a hybrid neural network model that can forecast PM2.5 multiple steps ahead In [6], Yanlin et al present a hybrid model that integrates graph convolutional networks and LSTM to predict PM2.5 In [7], the authors utilize the knearest neighbor algorithm to mine spatial-temporal information The historical information of related locations is then used as the input of the LSTM, adaptive temporal extractor (ASE), and artificial neural network (ANN) models Several other deep learning models for predicting air quality can be found in [8] - [11] Despite considerable effort, air quality prediction models still suffer from two issues: restrictions of the input and output lengths and unoptimized feature selection The first issue indicates that the number of time steps in a model’s output cannot exceed that of the input; i.e., the model cannot predict the future with upcoming steps that exceed the input data’s length Therefore, it is essential to remove this limitation in PM2.5 prediction The second issue arises from the fact that air quality data include dozens of factors other than PM2.5, such as various concentrations, temperature, and humidity These factors may or may not be related to PM2.5 However, appropriate use of some of these factors may improve the prediction accuracy Meanwhile, misuse may not only degrade the accuracy but also add extra computational time Therefore, choosing the optimal feature combination is essential 1.3 Goals and approaches This paper aims to address the two issues described above As a solution, we propose a novel PM2.5 prediction model that combines a genetic algorithm (GA) and an encoder-decoder (E-D) model The GA is exploited to perform feature selection in a nearoptimal manner, thereby enriching the prediction model Additionally, we leverage the encoder-decoder model to build a PM2.5 prediction model with high accuracy As a result, the proposed model can efficiently handle different sizes (in terms of the number of time steps) of input and output To demonstrate the effectiveness of our proposed approach, we evaluate the GA-based feature selection on the Hanoi [12] and Taiwan datasets [11] The evaluations show that the GA-based feature selection outperforms other methods We then compare our model to the state-of-the-art method ST-DNN in [11] using the Taiwan dataset Compared to ST-DNN, our model improves the accuracy from 14.82% to 41.71% By combining the GA-based feature selection algorithm and the E-D model, our proposed approach further increases the accuracy by at least 3% 1.4 Structure of thesis The remainder of this paper is organized as follows We describe the motivations in Section II Section III presents our proposal The performance of evaluation is introduced in Section IV Section VI introduces related works Finally, Section VII concludes the paper 80 80 70 50 40 PM2.5 value 20 30 60 50 40 30 0 10 10 20 PM2.5 value Groundtruth AC−LSTM AE−BiLSTM ED−LSTM 60 70 Groundtruth AC−LSTM AE−BiLSTM ED−LSTM 200 600 1000 1400 1800 200 600 1400 1800 80 (d) Four time steps 80 (c) Three time steps 40 PM2.5 value 40 10 20 30 10 20 50 60 70 Groundtruth AC−LSTM AE−BiLSTM ED−LSTM 30 50 60 70 Groundtruth AC−LSTM AE−BiLSTM ED−LSTM PM2.5 value 1000 Time Time 200 600 1000 1400 Time (e) Five time steps 1800 200 600 1000 1400 1800 Time (f) Six time steps Figure 22 Comparison between models using Hanoi dataset with feature selected by GA 38 MAE 14 21 28 30 Number of the output’s time steps 31 Figure 23 MAE of the proposed model with different output lengths We visualize the prediction results on the Hanoi dataset with different time-step values ahead in Figure 21 and Figure 22 In all cases, our proposed ED-LSTM model outperforms the AE-BiLSTM model This is reflected by the fact that the ED-LSTM model’s prediction results are much closer to the ground truth than the AE-BiLSTM model In comparison between the ED-LSTM model and the AC-LSTM model, it can be seen that the ED-LSTM model captures more peaks than the AC-LSTM model By comparing Figure 21 and Figure 22, we can see that the prediction results with the selected features from GA are closer to the ground truth than the results when using all features This result again proves the data shown in Table and Table 4, in which the MAE value of GA is always better than that obtained when using all the attributes To further demonstrate the effectiveness of the proposed model, we input the features selected by our GA-based algorithm into the encoder-decoder model to predict PM2.5 one month ahead Because we have hourly data, we calculate the mean value of each day from 24-hour data We fix the input length to 70 and vary the length of the output from to 31 to see how the output length affects the model’s accuracy Figure 23 shows that the model produces low and stable MAEs as the number of time steps in the output varies Furthermore, the last three test cases in the chart achieve the lowest MAE 4.4.2 Comparing ED-LSTM and ST-DNN In this section, we compare the performance of our proposed model with the STDNN model Note that due to the impossibility of reproducing the ST-DNN results, we have copied the results from Fig 23 in the ST-DNN paper [11] to the third column The results are collected and shown in Table ST-DNN (A+L+C) indicates the best combination of adaptive artificial neural networks (A), long short-term memory (L), and 39 convolutional neural network (C) in ST-DNN Our model outperforms ST-DNN under all experimental settings The fourth column shows the performance gap between our proposed approach and ST-DNN, ranging from 14:82% to 41:89% Interestingly, the performance gap tends to increase as we increase the time steps of the output Table Comparing the MAE of the proposed ED-LSTM model and the ST-DNN model (using the features proposed by [11]) Case ED-LSTM (proposed method) ST-DNN (A+L+C) Improvement 𝑙𝑙 = 48, ℎ = 2.454 2.881 14.82% 𝑙𝑙 = 48, ℎ = 3.930 5.362 26.71% 𝑙𝑙 = 48, ℎ = 3.907 6.524 40.11% 𝑙𝑙 = 48, ℎ = 4.844 7.364 34.22% 𝑙𝑙 = 48, ℎ = 4.604 7.923 41.89% 𝑙𝑙 = 48, ℎ = 5.142 8.821 41.71% To illustrate the accuracy of our prediction model, we visualize the details of the prediction results in Table and Figure 24 The figure includes six subfigures, each of which presents the prediction over time with a different number of time steps ahead Moreover, we plot the ground truth in the subfigures for comparison When the number of the output timesteps is small (i.e., Figure 24(a) and (b)), the predicted data accurately match the ground truth data Even the peak points are successfully predicted In the other figures, as the number of time steps ahead increases, the prediction becomes slightly less accurate than that in Figure 24(a) In summary, ED-LSTM outperforms the existing models in terms of prediction accuracy The improvement comes from two reasons The first one is introducing the GAbased feature selection algorithm, which helps determine the optimal feature combination The second one is that the combination of encoder-decoder and LSTM network helps extract meaningful information from the input On the other hand, EDLSTM has a slightly higher computation time than the model with the lowest value However, ED-LSTM’s computation time is still sufficient for real-time prediction of PM2.5 40 90 90 50 PM2.5 value 70 Groundtruth ED−LSTM 10 30 50 10 30 PM2.5 value 70 Groundtruth ED−LSTM 200 600 1000 1400 1800 200 600 Time 1400 1800 (b) Two time steps 90 90 (a) One time step 30 50 70 Groundtruth ED−LSTM 10 10 30 50 PM2.5 value 70 Groundtruth ED−LSTM 0 PM2.5 value 1000 Time 200 600 1000 1400 Time (c) Three time steps 1800 200 600 1000 1400 1800 Time (d) Four time steps 41 90 70 Groundtruth ED−LSTM 0 10 10 30 50 PM2.5 value 50 30 PM2.5 value 70 90 Groundtruth ED−LSTM 200 600 1000 1400 1800 200 600 Time 1000 1400 1800 Time (e) Five time steps (f) Six time steps Figure 24 Comparison between models using Taiwan dataset with features selected by [11] 4.5 LTS2 evaluation In this part of the experiment, I compared new training strategy with traditional training Each training instance will be denoted as 𝑛𝑛%_𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓_𝑠𝑠ℎ𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢 100_any_any represents the initial training case, that is, remove all training data during the training process Therefore, the variable fixed and shuffling being true or false does not affect the result Table summarizes the hyperparameters used in this new training strategy Table Hyperparameters of training strategy Hyperparameters Values 𝑛𝑛% {10, 20, 30, 40, 50} 𝑠𝑠ℎ𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢𝑢 {true, false} 𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓𝑓 {true, false} The results in Table show that in any case, the new training strategy results in a significantly reduced training time compared to the original training strategy Although there is a decrease in accuracy, but not significantly when looking at the 10_true_false case, the MAE is only about 0.05 difference from the 3,592 of the 100_any_any case while the training time is reduced by almost 25 time 42 Table Training strategy for different cases Case (𝒏𝒏%_𝒇𝒇𝒇𝒇𝒇𝒇𝒇𝒇𝒇𝒇_𝒔𝒔𝒔𝒔𝒔𝒔𝒔𝒔𝒔𝒔𝒔𝒔𝒔𝒔𝒔𝒔𝒔𝒔) Training time MAE 100%_any_any 518400 3.592 10_false_false 23523 3.829 10_false_true 21450 3.693 10_true_false 23101 3.644 10_true_ true 22544 3.672 20_false_false 27746 4.061 20_false_true 27746 4.205 20_true_false 27140 3.881 20_true_ true 24860 4.278 30_false_false 30591 4.162 30_false_true 31292 5.640 30_true_false 35006 3.960 30_true_ true 33070 3.738 40_false_false 33070 3.805 40_false_true 32413 3.937 40_true_false 31756 3.752 40_true_ true 31370 3.675 50_false_false 34497 3.802 50_false_true 33528 3.737 50_true_false 52819 3.788 50_true_ true 51686 4.077 Table demonstrate that the features combinations generated by the new training method still ensures higher accuracy than related works when using the attribute set selected by GA when training the entire training data 43 Table Comparing proposed method combining new training strategy with related works h ED-LSTM (Proposed method) AEBiLSTM AC- ED-LSTM with new training LSTM strategy (LTS2) MAE MAE MAE MAE 3.592 8.242 3.919 3.693 4.603 8.567 4.762 4.488 5.173 8.830 5.531 5.237 5.491 9.083 6.168 5.891 6.135 9.323 6.748 5.946 6.684 9.634 7.710 6.678 Thus, with the proposed training tactic reduced the proposed method's training time by at least 10 and at most 25 times while keeping the MAE at least 10% lower than related works 4.6 Discussion 4.6.1 Results of OFFGED According to our GA-based feature selection methods, the feature combination that produces the best performance for predicting PM2.5 is {wind speed, temperature, radiation, PM10} Therefore, these features have the greatest effect on PM2.5 Indeed, we have measured the correlation between PM2.5 and all the features using three methods: Spearman’s correlation (SC), Pearson correlation (PC), and mutual information score (MIC) Table represents the absolute values of SC, PC, and MIC The wind speed, temperature, radiation, and PM10 show high correlations with PM2.5, which is reflected by the high absolute values of SC, PC, and MIC Our findings are consistent with the results of previous studies In [22], the authors showed that temperature is positively correlated with PM2.5, and a threshold decides the 44 correlation between wind speed and PM2.5 Specifically, when the wind speed is less than 𝑚𝑚/𝑠𝑠, it is negatively correlated with PM2.5, and beyond that point, it is positively correlated with PM2.5 The work in [23] also found that PM2.5 is affected the most by wind speed and temperature The relationship between radiation and temperature with PM2.5 is studied in [24] The authors in [24] showed that increasing temperature and decreasing radiation lead to increases in PM2.5 [25] – [27] have shown that PM10 is strongly related to PM2.5 Table Correlation of features SC PC MIC Time 0.345790 0.292338 0.992296 Month 0.184097 0.138431 0.401139 Day 0.003414 0.004860 0.511531 Year 0.295231 0.258422 0.130763 Hour 0.062853 0.072817 0.480540 Wind speed 0.263086 0.259349 0.989226 Wind direction 0.009217 0.040491 0.989369 Temperature 0.386038 0.391475 0.989201 Relative humidity 0.047072 0.084301 0.734780 Barometer 0.354467 0.366005 0.882035 Radiation 0.187909 0.214549 0.896700 Inner temperature 0.402831 0.361091 0.985135 PM10 0.924673 0.934090 0.994290 PM1 0.979046 0.985997 0.994193 45 In summary, our GA-based feature selection method has selected the optimal feature combination for predicting PM2.5 The optimal combination includes features that have a considerable effect on PM2.5 4.6.2 Results of LTS2 From the results from section 4.5, we can see that the new training strategy LTS2 offers nearly the same accuracy compared to training with the entire training data However, the training speed is significantly improved from 10 times up to 25 times Therefore, in real world, we can tradeoff between accuracy and training speed Besides, if we adjust the hyperparameters reasonably, the results from the new training strategy can still achieve better results when training with the whole training data 46 Conclusion In this paper, we have presented a novel prediction model for PM2.5 that combines PM2.5 with other air quality-related features Our model utilizes a GA-based feature selection algorithm and an ED-LSTM prediction model While the GA-based algorithm efficiently determines the near-optimal feature combination, the ED-LSTM model leverages LSTM units to loosen the restriction on the length of the input and output data The experimental results indicate that our ED-LSTM model can improve the MAE up to 53.7% compared to that of state-of-the-art PM2.5 prediction models Moreover, our proposed approach that includes the GA-based feature selection algorithm further improves the prediction accuracy by at least 13.7% In addition, I also propose a training mechanism that reduces the training time by at least 10 times compared to the traditional way while maintaining high accuracy Currently, the research has completed the goal of having a predictive model with high accuracy for the fine dust index, but there are still outstanding problems and can be further developed I have the following development directions: • Because the problem is to solve environmental problems in Vietnam, but only Hanoi's data is available, the geographical and spatial relationships have not been exploited For the model to be able to be applied, it is necessary to predict for many different areas, so I aim to find the data of other cities, then rely on geographical factors to increase the accuracy of the model up • Because the data is often missing because the sensors are not completely guaranteed to get the data, thus affecting the learning and prediction process I propose the development direction is to find a method to fill in the blank data so that when predicting real data, the accuracy is higher than some popular methods such as the averaging method 47 Published papers [1] Minh Hieu Nguyen, Phi Le Nguyen, Kien Nguyen, Van An Le, Thanh-Hung Nguyen, Yusheng Ji, “PM2.5 Prediction Using Genetic Algorithm-based Feature Selection and Encoder-Decoder Model,” IEEE Access, Vol 9, pp 57338 - 57350, 2021 48 References [1] C Arden Pope, III, Richard T Burnett, Michael J Thun, Eugenia E Calle, Daniel Krewski, Kazuhiko Ito, George D Thurston, “Lung Cancer, Cardiopulmonary Mortality, and Long-term Exposure to Fine Particulate Air Pollution”, JAMA, PMC Publisher, vol 287, 2002 [2] S Hochreiter and J J Schmidhuber, ‘‘Long short-term memory,’’ Neural Comput., vol 9, no 8, pp 1735–1780, 1997 [3] I Kok, M U Simsek, and S Ozdemir, ‘‘A deep learning model for air quality prediction in smart cities,’’ in Proc IEEE Int Conf Big Data (Big Data), Dec 2017, pp 1983–1990 [4] Y.-T Tsai, Y.-R Zeng, and Y.-S Chang, ‘‘Air pollution forecasting using RNN with LSTM,’’ in Proc IEEE 16th Int Conf Dependable, Autonomic Secure Comput., 16th Int Conf Pervasive Intell Comput., 4th Int Conf Big Data Intell Comput Cyber Sci Technol Congr (DASC/PiCom/DataCom/CyberSciTech), Aug 2018, pp 1074–1079 [5] Y Zhou, F.-J Chang, L.-C Chang, I.-F Kao, and Y.-S Wang, ‘‘Explore a deep learning multi-output neural network for regional multi-step-ahead air quality forecasts,’’ J Cleaner Prod., vol 209, pp 134–145, Feb 2019 [6] Y Qi, Q Li, H Karimian, and D Liu, ‘‘A hybrid model for spatiotemporal forecasting of PM2:5 based on graph convolutional neural network and long shortterm memory,’’ Sci Total Environ., vol 664, pp 1–10, May 2019 [7] C Wen, S Liu, X Yao, L Peng, X Li, Y Hu, and T Chi, ‘‘A novel spatiotemporal convolutional long short-term neural network for air pollution prediction,’’ Sci Total Environ., vol 654, pp 1091–1099, Mar 2019 [8] J Ma, Z Li, J C P Cheng, Y Ding, C Lin, and Z Xu, ‘‘Air quality prediction at new stations using spatially transferred bi-directional long short-term memory network,’’ Sci Total Environ., vol 705, Feb 2020, Art no 135771 [9] J Wang, P Du, Y Hao, X Ma, T Niu, and W Yang, ‘‘An innovative hybrid model based on outlier detection and correction algorithm and heuristic intelligent optimization algorithm for daily air quality index forecasting,’’ J Environ Manage., vol 255, Feb 2020, Art no 109855 49 [10] Z Qi, T Wang, G Song, W Hu, X Li, and Z Zhang, ‘‘Deep air learning: Interpolation, prediction, and feature analysis of fine-grained air quality,’’ IEEE Trans Knowl Data Eng., vol 30, no 12, pp 2285–2297, Dec 2018 [11] P.-W Soh, J.-W Chang, and J.-W Huang, ‘‘Adaptive deep learning-based air quality prediction model using the most relevant spatial-temporal relations,’’ IEEE Access, vol 6, pp 38186–38199, 2018 [12] Hanoi Dataset Accessed: Nov 2020 [Online] Available: https://bit.ly/hanoi-pm [13] M Z Joharestani, C Cao, X Ni, B Bashir, and S Talebiesfandarani, ‘‘PM2:5 prediction based on random forest, XGBoost, and deep learning using multisource remote sensing data,’’ Atmosphere, vol 10, no 7, p 373, Jul 2019 [14] Y Lin, N Mago, Y Gao, Y Li, Y.-Y Chiang, C Shahabi, and J L Ambite, ‘‘Exploiting spatiotemporal patterns for accurate air quality forecasting using deep learning,’’ in Proc 26th ACM SIGSPATIAL Int Conf Adv Geographic Inf Syst., Nov 2018, pp 359–368 [15] J Biesiada and W Duch, ‘‘Feature selection for high-dimensional data — A Pearson redundancy based filter,’’ in Computer Recognition Systems 2, vol 45 Berlin, Germany: Springer, 2007, pp 242–249 [16] D Panda, R Ray, A A Abdullah, and S R Dash, ‘‘Predictive systems: Role of feature selection in prediction of heart disease,’’ J Phys., Conf Ser., vol 1372, Nov 2019, Art no 012074 [17] R Genuer, J.-M Poggi, and C Tuleau-Malot, ‘‘Variable selection using random forests,’’ Pattern Recognit Lett., vol 31, no 14, pp 2225–2236, Oct 2010 [18] D P Kingma and J Ba, ‘‘Adam: A method for stochastic optimization,’’ 2014, arXiv:1412.6980 [Online] Available: https://arxiv.org/abs/1412.6980 [19] I Sutskever, O Vinyals, and V Q Le, ‘‘Sequence to sequence learning with neural networks,’’ 2014, arXiv:1409.3215 [Online] Available: https://arxiv.org/abs/1409.3215 [20] B Zhang, H Zhang, G Zhao, and J Lian, ‘‘Constructing a PM2:5 concentration prediction model by combining auto-encoder with biLSTM neural networks,’’ Environ Model Softw., vol 124, Feb 2020, Art no 104600 [21] S Li, G Xie, J Ren, L Guo, Y Yang, and X Xu, ‘‘Urban PM2:5 concentration prediction via attention-based CNN–LSTM,’’ Appl Sci., vol 10, no 6, p 1953, Mar 2020 50 [22] J Wang and S Ogawa, ‘‘Effects of meteorological conditions on PM2:5 concentrations in Nagasaki, Japan,’’ Int J Environ Res Public Health, vol 12, no 8, pp 9089–9101, 2015 [23] P D Hien, V T Bac, H C Tham, D D Nhan, and L D Vinh, ‘‘Influence of meteorological conditions on PM2:5 and PM2:5−10 concentrations during the monsoon season in Hanoi, Vietnam,’’ Atmos Environ., vol 36, no 21, pp 3473– 3484, Jul 2002 [24] M Préndez, M Egido, C TomỈs, J Seco, A Calvo, and H Romero, ‘‘Correlation between solar radiation and total syspended particulate matter in Santiago, Chile— Preliminary results,’’ Atmos Environ., vol 29, no 13, pp 1543–1551, Jul 1995 [25] X Zhou, Z Cao, Y Ma, L Wang, R Wu, and W Wang, ‘‘Concentrations, correlations and chemical species of PM2:5/PM10 based on published data in China: Potential implications for the revised particulate standard,’’ Chemosphere, vol 144, pp 518–526, Feb 2016 [26] E Maraziotis and N Marazioti, ‘‘Statistical analysis of inhalable (PM10) and fine particles (PM2:5) concentrations in urban region of Patras, Greece,’’ Global Nest J., vol 10, no 2, pp 123–131, 2008 [27] D Zhao, H Chen, E Yu, and T Luo, ‘‘PM2:5/PM10 ratios in eight economic regions and their relationship with meteorology in China,’’ Adv Meteorol., vol 2019, pp 1–15, Feb 2019 [28] B Liu, S Yan, J Li, G Qu, Y Li, J Lang, and R Gu, ‘‘A sequence-tosequence air quality predictor based on the n-step recurrent prediction,’’ IEEE Access, vol 7, pp 43331–43345, 2019 [29] L Yan, Y Wu, L Yan, and M Zhou, ‘‘Encoder-decoder model for forecast of PM2:5 concentration per hour,’’ in Proc 1st Int Cognit Cities Conf (IC3), Aug 2018, pp 45–50 [30] Y Zhang, Q Lv, D Gao, S Shen, R Dick, M Hannigan, and Q Liu, ‘‘Multi-group encoder-decoder networks to fuse heterogeneous data for next-day air quality prediction,’’ in Proc 28th Int Joint Conf Artif Intell., Aug 2019, pp 4341–4347 [31] Y Zhang, P Li, and X Wang, ‘‘Intrusion detection for IoT based on improved genetic algorithm and deep belief network,’’ IEEE Access, vol 7, pp 31711– 31722, 2019 [32] S Bouktif, A Fiaz, A Ouni, and M Serhani, ‘‘Optimal deep learning LSTM model for electric load forecasting using feature selection and genetic algorithm: 51 Comparison with machine learning approaches,’’ Energies, vol 11, no 7, p 1636, Jun 2018 [33] A Sagheer and M Kotb, ‘‘Time series forecasting of petroleum production using deep LSTM recurrent networks,’’ Neurocomputing, vol 323, pp 203–213, Jan 2019 [34] W Liu, G Guo, F Chen, and Y Chen, ‘‘Meteorological pattern analysis assisted daily PM2:5 grades prediction using SVM optimized by PSO algorithm,’’ Atmos Pollut Res., vol 10, no 5, pp 1482–1491, Sep 2019 [35] H Liu, Z Duan, and C Chen, ‘‘A hybrid multi-resolution multi-objective ensemble model and its application for forecasting of daily PM2:5 concentrations,’’ Inf Sci., vol 516, pp 266–292, Apr 2020 52 ... 10.0 452 16. 053 4 0. 353 559 1023.83 1.0026 24.0024 69 .59 2 48.9237 54 .049 0 .50 966 02/ 05/ 2021 342.787 16.0088 0. 354 644 1024. 15 1.0099 23.8 453 79.391 44. 151 5 54 .57 9 0.60866 29/04/2021 30/04/2021 Feature. .. 0 .56 33 01/ 05/ 2021 10.0 452 0. 353 559 1023.83 1.0026 0 .50 966 02/ 05/ 2021 342.787 0. 354 644 1024. 15 1.0099 0.60866 Figure An example of feature selection 2 .5. 4 Feature construction Building new features... Forecasting Framework using GA -based feature selection and EncoderDecoder model) , which includes GA -based feature selection and a prediction module that are detailed in Section 3.2 and Section 3.3,

Ngày đăng: 15/02/2022, 19:00

Mục lục

  • Content

  • Introduction

  • Chapter 1. Related works

  • Chapter 2. Theoretical Background

  • Chapter 3. Proposed Forecasting Framework (OFFGED)

  • Chapter 4. Performance Evaluation

  • Conclusion

  • Published papers

  • References

Tài liệu cùng người dùng

Tài liệu liên quan