Prof. Gualberto Asencio Cortés, Ph.D. is a Computer Science Engineer (University of Seville, 2008), Master in Software Engineering and Technology (University of Seville, 2010), Ph.D. (University of Pablo de Olavide, 2013) and he has an Executive Master in Innovation (EOI, Spain, 2016). He is Associate Professor of Computer Science (Profesor Titular de Universidad), in the area of Languages and Information Systems at the University of Pablo de Olavide. He is the author of more than 28 publications in impact journals according to JCR (20 of them between Q1 and Q2) and author of more than 30 articles in international and national conferences, most of them published in LNCS and LNBI. He has participated in three projects of the National Plan and three more of the Andalusian Research Plan. He is an editor of PLOS ONE (IF: 2.806, Q1), a regular reviewer of journals indexed in JCR (PLOS ONE, Bioinformatics, Neurocomputing, Computer and Geosciences, etc.) and member of the program committee in numerous international conferences. He has participated in more than 12 technology transfer contracts between the university and the company, including ISOTROL, Red Eléctrica Española and DETEA. He has 5 months of international research stays and 3 national months.
The research lines of Prof. Gualberto Asencio Cortés, Ph.D. are focused on data mining, machine learning, prediction of time series and bioinformatics, with different fields of application: prediction of natural series (seismic, air quality, meteorological, agronomic, …), prediction of electricity consumption and market prices, prediction of urban traffic, as well as bioinformatics in prediction of biological structures. He has also been data scientist and member of the steering committee responsible for artificial intelligence and data science technologies at the private company easytosee AgTech SL for more than 2 years (2015-2017).
Publications
2019 |
F. Martinez-Alvarez and A. Schmutz and G. Asencio-Cortes and J. Jacques A Novel Hybrid Algorithm to Forecast Functional Time Series Based on Pattern Sequence Similarity with Application to Electricity Demand Journal Article Energies, 12 (94), pp. 1-18, 2019, ISSN: 1996-1073. @article{en12010094b, title = {A Novel Hybrid Algorithm to Forecast Functional Time Series Based on Pattern Sequence Similarity with Application to Electricity Demand}, author = {F. Martinez-Alvarez and A. Schmutz and G. Asencio-Cortes and J. Jacques}, url = {http://www.mdpi.com/1996-1073/12/1/94}, doi = {10.3390/en12010094}, issn = {1996-1073}, year = {2019}, date = {2019-01-01}, journal = {Energies}, volume = {12}, number = {94}, pages = {1-18}, abstract = {The forecasting of future values is a very challenging task. In almost all scientific disciplines, the analysis of time series provides useful information and even economic benefits. In this context, this paper proposes a novel hybrid algorithm to forecast functional time series with arbitrary prediction horizons. It integrates a well-known clustering functional data algorithm into a forecasting strategy based on pattern sequence similarity, which was originally developed for discrete time series. The new approach assumes that some patterns are repeated over time, and it attempts to discover them and evaluate their immediate future. Hence, the algorithm first applies a clustering functional time series algorithm, i.e., it assigns labels to every data unit (it may represent either one hour, or one day, or any arbitrary length). As a result, the time series is transformed into a sequence of labels. Later, it retrieves the sequence of labels occurring just after the sample that we want to be forecasted. This sequence is searched for within the historical data, and every time it is found, the sample immediately after is stored. Once the searching process is terminated, the output is generated by weighting all stored data. The performance of the approach has been tested on real-world datasets related to electricity demand and compared to other existing methods, reporting very promising results. Finally, a statistical significance test has been carried out to confirm the suitability of the election of the compared methods. In conclusion, a novel algorithm to forecast functional time series is proposed with very satisfactory results when assessed in the context of electricity demand.}, keywords = {}, pubstate = {published}, tppubtype = {article} } The forecasting of future values is a very challenging task. In almost all scientific disciplines, the analysis of time series provides useful information and even economic benefits. In this context, this paper proposes a novel hybrid algorithm to forecast functional time series with arbitrary prediction horizons. It integrates a well-known clustering functional data algorithm into a forecasting strategy based on pattern sequence similarity, which was originally developed for discrete time series. The new approach assumes that some patterns are repeated over time, and it attempts to discover them and evaluate their immediate future. Hence, the algorithm first applies a clustering functional time series algorithm, i.e., it assigns labels to every data unit (it may represent either one hour, or one day, or any arbitrary length). As a result, the time series is transformed into a sequence of labels. Later, it retrieves the sequence of labels occurring just after the sample that we want to be forecasted. This sequence is searched for within the historical data, and every time it is found, the sample immediately after is stored. Once the searching process is terminated, the output is generated by weighting all stored data. The performance of the approach has been tested on real-world datasets related to electricity demand and compared to other existing methods, reporting very promising results. Finally, a statistical significance test has been carried out to confirm the suitability of the election of the compared methods. In conclusion, a novel algorithm to forecast functional time series is proposed with very satisfactory results when assessed in the context of electricity demand. |
2018 |
C. Rubio-Escudero and G. Asencio-Cortés and F. Martínez-Álvarez and A. Troncoso and J. C. Riquelme Impact of Auto-evaluation Tests as Part of the Continuous Evaluation in Programming Courses Conference ICEUTE 9th International Conference on European Transnational Education, 771 , Advances in Intelligent Systems and Computing 2018. @conference{RUBIO18, title = {Impact of Auto-evaluation Tests as Part of the Continuous Evaluation in Programming Courses}, author = {C. Rubio-Escudero and G. Asencio-Cortés and F. Martínez-Álvarez and A. Troncoso and J. C. Riquelme}, url = {https://link.springer.com/chapter/10.1007/978-3-319-94120-2_54}, doi = {https://doi.org/chapter/10.1007/978-3-319-94120-2_54}, year = {2018}, date = {2018-06-07}, booktitle = {ICEUTE 9th International Conference on European Transnational Education}, volume = {771}, pages = {553-561}, series = {Advances in Intelligent Systems and Computing}, abstract = {The continuous evaluation allows for the assessment of the progressive assimilation of concepts and the competences that must be achieved in a course. There are several ways to implement such continuous evaluation system. We propose auto-evaluation tests as a valuable tool for the student to judge his level of knowledge. Furthermore, these tests are also used as a small part of the continuous evaluation process, encouraging students to learn the concepts seen in the course, as they have the feeling that the time dedicated to this study will have an assured reward, binge able to answer correctly the questions in the continuous evaluation exams. New technologies are a great aid to improve the auto-evaluation experience both for the students and the teachers. In this research work we have compared the results obtained in courses where auto-evaluation tests were provided against courses where they were not provided, showing how the tests improve a set of quality metrics in the results of the course.}, keywords = {}, pubstate = {published}, tppubtype = {conference} } The continuous evaluation allows for the assessment of the progressive assimilation of concepts and the competences that must be achieved in a course. There are several ways to implement such continuous evaluation system. We propose auto-evaluation tests as a valuable tool for the student to judge his level of knowledge. Furthermore, these tests are also used as a small part of the continuous evaluation process, encouraging students to learn the concepts seen in the course, as they have the feeling that the time dedicated to this study will have an assured reward, binge able to answer correctly the questions in the continuous evaluation exams. New technologies are a great aid to improve the auto-evaluation experience both for the students and the teachers. In this research work we have compared the results obtained in courses where auto-evaluation tests were provided against courses where they were not provided, showing how the tests improve a set of quality metrics in the results of the course. |
A. Troncoso and P. Ribera and G. Asencio-Cortés and I. Vega and D. Gallego Imbalanced classification techniques for monsoon forecasting based on a new climatic time series Journal Article Environmental Modelling & Software, 106 (6), pp. 48-56, 2018. @article{ENV2018, title = {Imbalanced classification techniques for monsoon forecasting based on a new climatic time series}, author = {A. Troncoso and P. Ribera and G. Asencio-Cortés and I. Vega and D. Gallego}, url = {https://www.sciencedirect.com/science/article/pii/S1364815217301950}, doi = {10.1016/j.envsoft.2017.11.024}, year = {2018}, date = {2018-01-01}, journal = {Environmental Modelling & Software}, volume = {106}, number = {6}, pages = {48-56}, abstract = {Monsoons have been widely studied in the literature due to their climatic impact related to precipitation and temperature over different regions around the world. In this work, data mining techniques, namely imbalanced classification techniques, are proposed in order to check the capability of climate indices to capture and forecast the evolution of theWestern North Pacific Summer Monsoon. Thus, the main goal is to predict if the monsoon will be an extreme monsoon for a temporal horizon of a month. Firstly, a new monthly index of the monsoon related to its intensity has been generated. Later, the problem of forecasting has been transformed into a binary imbalanced classification problem and a set of representative techniques, such as models based on trees, models based on rules, black box models and ensemble techniques, are applied to obtain the forecasts. From the results obtained, it can be concluded that the methodology proposed here reports promising results according to the quality measures evaluated and predicts extreme monsoons for a temporal horizon of a month with a high accuracy.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Monsoons have been widely studied in the literature due to their climatic impact related to precipitation and temperature over different regions around the world. In this work, data mining techniques, namely imbalanced classification techniques, are proposed in order to check the capability of climate indices to capture and forecast the evolution of theWestern North Pacific Summer Monsoon. Thus, the main goal is to predict if the monsoon will be an extreme monsoon for a temporal horizon of a month. Firstly, a new monthly index of the monsoon related to its intensity has been generated. Later, the problem of forecasting has been transformed into a binary imbalanced classification problem and a set of representative techniques, such as models based on trees, models based on rules, black box models and ensemble techniques, are applied to obtain the forecasts. From the results obtained, it can be concluded that the methodology proposed here reports promising results according to the quality measures evaluated and predicts extreme monsoons for a temporal horizon of a month with a high accuracy. |
A. Gomez-Losada and G. Asencio-Cortes and F. Martinez-Alvarez and J. C. Riquelme A novel approach to forecast urban surface-level ozone considering heterogeneous locations and limited information Journal Article Environmental Modelling and Software, (110), pp. 52-61, 2018, ISSN: 1364-8152. @article{Gomez-Losada2018b, title = {A novel approach to forecast urban surface-level ozone considering heterogeneous locations and limited information}, author = {A. Gomez-Losada and G. Asencio-Cortes and F. Martinez-Alvarez and J. C. Riquelme}, doi = {10.1016/j.envsoft.2018.08.013}, issn = {1364-8152}, year = {2018}, date = {2018-01-01}, journal = {Environmental Modelling and Software}, number = {110}, pages = {52-61}, keywords = {}, pubstate = {published}, tppubtype = {article} } |
E. Florido and G. Asencio-Cortes and J. L. Aznarte and C. Rubio-Escudero and F. Martinez-Alvarez A novel tree-based algorithm to discover seismic patterns in earthquake catalogs Journal Article Computers and Geosciences, (115), pp. 96-104, 2018, ISSN: 0098-3004. @article{Florido2018, title = {A novel tree-based algorithm to discover seismic patterns in earthquake catalogs}, author = {E. Florido and G. Asencio-Cortes and J. L. Aznarte and C. Rubio-Escudero and F. Martinez-Alvarez}, doi = {10.1016/j.cageo.2018.03.005}, issn = {0098-3004}, year = {2018}, date = {2018-01-01}, journal = {Computers and Geosciences}, number = {115}, pages = {96-104}, abstract = {A novel methodology is introduced in this research study to detect seismic precursors. Based on an existing approach, the new methodology searches for patterns in the historical data. Such patterns may contain statistical or soil dynamics information. It improves the original version in several aspects. First, new seismicity indicators have been used to characterize earthquakes. Second, a machine learning clustering algorithm has been applied in a very flexible way, thus allowing the discovery of new data groupings. Third, a novel search strategy is proposed in order to obtain non-overlapped patterns. And, fourth, arbitrary lengths of patterns are searched for, thus discovering long and short-term behaviors that may influence in the occurrence of medium-large earthquakes. The methodology has been applied to seven different datasets, from three different regions, namely the Iberian Peninsula, Chile and Japan. Reported results show a remarkable improvement with respect to the former version, in terms of all evaluated quality measures. In particular, the number of false positives has decreased and the positive predictive values increased, both of them in a very remarkable manner.}, keywords = {}, pubstate = {published}, tppubtype = {article} } A novel methodology is introduced in this research study to detect seismic precursors. Based on an existing approach, the new methodology searches for patterns in the historical data. Such patterns may contain statistical or soil dynamics information. It improves the original version in several aspects. First, new seismicity indicators have been used to characterize earthquakes. Second, a machine learning clustering algorithm has been applied in a very flexible way, thus allowing the discovery of new data groupings. Third, a novel search strategy is proposed in order to obtain non-overlapped patterns. And, fourth, arbitrary lengths of patterns are searched for, thus discovering long and short-term behaviors that may influence in the occurrence of medium-large earthquakes. The methodology has been applied to seven different datasets, from three different regions, namely the Iberian Peninsula, Chile and Japan. Reported results show a remarkable improvement with respect to the former version, in terms of all evaluated quality measures. In particular, the number of false positives has decreased and the positive predictive values increased, both of them in a very remarkable manner. |
G. Asencio-Cortes and A. Morales-Esteban and X. Shang and F. Martinez-Alvarez Earthquake prediction in California using regression algorithms and cloud-based big data infrastructure Journal Article Computers and Geosciences, (115), pp. 198-210, 2018, ISSN: 0098-3004. @article{Asencio-Cortes2018b, title = {Earthquake prediction in California using regression algorithms and cloud-based big data infrastructure}, author = {G. Asencio-Cortes and A. Morales-Esteban and X. Shang and F. Martinez-Alvarez}, doi = {10.1016/j.cageo.2017.10.011}, issn = {0098-3004}, year = {2018}, date = {2018-01-01}, journal = {Computers and Geosciences}, number = {115}, pages = {198-210}, abstract = {Earthquake magnitude prediction is a challenging problem that has been widely studied during the last decades. Statistical, geophysical and machine learning approaches can be found in literature, with no particularly satisfactory results. In recent years, powerful computational techniques to analyze big data have emerged, making possible the analysis of massive datasets. These new methods make use of physical resources like cloud based architectures. California is known for being one of the regions with highest seismic activity in the world and many data are available. In this work, the use of several regression algorithms combined with ensemble learning is explored in the context of big data (1 GB catalog is used), in order to predict earthquakes magnitude within the next seven days. Apache Spark framework, H2O library in R language and Amazon cloud infrastructure were been used, reporting very promising results.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Earthquake magnitude prediction is a challenging problem that has been widely studied during the last decades. Statistical, geophysical and machine learning approaches can be found in literature, with no particularly satisfactory results. In recent years, powerful computational techniques to analyze big data have emerged, making possible the analysis of massive datasets. These new methods make use of physical resources like cloud based architectures. California is known for being one of the regions with highest seismic activity in the world and many data are available. In this work, the use of several regression algorithms combined with ensemble learning is explored in the context of big data (1 GB catalog is used), in order to predict earthquakes magnitude within the next seven days. Apache Spark framework, H2O library in R language and Amazon cloud infrastructure were been used, reporting very promising results. |
X. Shang and X. Li and A. Morales-Esteban and G. Asencio-Cortes and Z. Wang Data field-based K-means clustering for spatio-temporal seismicity analysis and hazard assessment Journal Article Remote Sensing, 10 (461), pp. 1-22, 2018, ISSN: 2072-4292. @article{Shang2018b, title = {Data field-based K-means clustering for spatio-temporal seismicity analysis and hazard assessment}, author = {X. Shang and X. Li and A. Morales-Esteban and G. Asencio-Cortes and Z. Wang}, doi = {10.3390/rs10030461}, issn = {2072-4292}, year = {2018}, date = {2018-01-01}, journal = {Remote Sensing}, volume = {10}, number = {461}, pages = {1-22}, abstract = {Microseismic sensing taking advantage of sensors can remotely monitor seismic activities and evaluate seismic hazard. Compared with experts' seismic event clusters, clustering algorithms are more objective, and they can handle many seismic events. Many methods have been proposed for seismic event clustering and the K-means clustering technique has become the most famous one. However, K-means can be affected by noise events (large location error events) and initial cluster centers. In this paper, a data field-based K-means clustering methodology is proposed for seismicity analysis. The application of synthetic data and real seismic data have shown its effectiveness in removing noise events as well as finding good initial cluster centers. Furthermore, we introduced the time parameter into the K-means clustering process and applied it to seismic events obtained from the Chinese Yongshaba mine. The results show that the time-event location distance and data field-based K-means clustering can divide seismic events by both space and time, which provides a new insight for seismicity analysis compared with event location distance and data field-based K-means clustering. The Krzanowski-Lai (KL) index obtains a maximum value when the number of clusters is five: the energy index (EI) shows that clusters C1, C3 and C5 have very critical periods. In conclusion, the time-event location distance, and the data field-based K-means clustering can provide an effective methodology for seismicity analysis and hazard assessment. In addition, further study can be done by considering time-event location-magnitude distances.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Microseismic sensing taking advantage of sensors can remotely monitor seismic activities and evaluate seismic hazard. Compared with experts' seismic event clusters, clustering algorithms are more objective, and they can handle many seismic events. Many methods have been proposed for seismic event clustering and the K-means clustering technique has become the most famous one. However, K-means can be affected by noise events (large location error events) and initial cluster centers. In this paper, a data field-based K-means clustering methodology is proposed for seismicity analysis. The application of synthetic data and real seismic data have shown its effectiveness in removing noise events as well as finding good initial cluster centers. Furthermore, we introduced the time parameter into the K-means clustering process and applied it to seismic events obtained from the Chinese Yongshaba mine. The results show that the time-event location distance and data field-based K-means clustering can divide seismic events by both space and time, which provides a new insight for seismicity analysis compared with event location distance and data field-based K-means clustering. The Krzanowski-Lai (KL) index obtains a maximum value when the number of clusters is five: the energy index (EI) shows that clusters C1, C3 and C5 have very critical periods. In conclusion, the time-event location distance, and the data field-based K-means clustering can provide an effective methodology for seismicity analysis and hazard assessment. In addition, further study can be done by considering time-event location-magnitude distances. |
2017 |
N. Bokde and A. Troncoso and G. Asencio-Cortés and K. Kulat and F. Martínez-Álvarez Pattern sequence similarity based techniques for wind speed forecasting Conference ITISE International Work-Conference on Time Series Analysis, 2017. @conference{ITISE2017, title = {Pattern sequence similarity based techniques for wind speed forecasting}, author = {N. Bokde and A. Troncoso and G. Asencio-Cortés and K. Kulat and F. Martínez-Álvarez}, year = {2017}, date = {2017-01-01}, booktitle = {ITISE International Work-Conference on Time Series Analysis}, keywords = {}, pubstate = {published}, tppubtype = {conference} } |
N. Bokde and G. Asencio-Cortes and F. Martinez-Alvarez and K. Kulat PSF: Introduction to R Package for Pattern Sequence Based Forecasting Algorithm Journal Article R Journal, 1 (9), pp. 324-333, 2017, ISSN: 2073-4859. @article{Bokde2016a, title = {PSF: Introduction to R Package for Pattern Sequence Based Forecasting Algorithm}, author = {N. Bokde and G. Asencio-Cortes and F. Martinez-Alvarez and K. Kulat}, issn = {2073-4859}, year = {2017}, date = {2017-01-01}, journal = {R Journal}, volume = {1}, number = {9}, pages = {324-333}, abstract = {This paper discusses about an R package that implements the Pattern Sequence based Forecasting (PSF) algorithm, which was developed for univariate time series forecasting. This algorithm has been successfully applied to many different fields. The PSF algorithm consists of two major parts: clustering and prediction. The clustering part includes selection of the optimum number of clusters. It labels time series data with reference to such clusters. The prediction part includes functions like optimum window size selection for specific patterns and prediction of future values with reference to past pattern sequences. The PSF package consists of various functions to implement the PSF algorithm. It also contains a function which automates all other functions to obtain optimized prediction results. The aim of this package is to promote the PSF algorithm and to ease its implementation with minimum efforts. This paper describes all the functions in the PSF package with their syntax. It also provides a simple example of usage. Finally, the usefulness of this package is discussed by comparing it to auto.arima and ets, well-known time series forecasting functions available on CRAN repository.}, keywords = {}, pubstate = {published}, tppubtype = {article} } This paper discusses about an R package that implements the Pattern Sequence based Forecasting (PSF) algorithm, which was developed for univariate time series forecasting. This algorithm has been successfully applied to many different fields. The PSF algorithm consists of two major parts: clustering and prediction. The clustering part includes selection of the optimum number of clusters. It labels time series data with reference to such clusters. The prediction part includes functions like optimum window size selection for specific patterns and prediction of future values with reference to past pattern sequences. The PSF package consists of various functions to implement the PSF algorithm. It also contains a function which automates all other functions to obtain optimized prediction results. The aim of this package is to promote the PSF algorithm and to ease its implementation with minimum efforts. This paper describes all the functions in the PSF package with their syntax. It also provides a simple example of usage. Finally, the usefulness of this package is discussed by comparing it to auto.arima and ets, well-known time series forecasting functions available on CRAN repository. |
J. L. Amaro-Mellado and A. Morales-Esteban and G. Asencio-Cortes and F. Martinez-Alvarez Comparing seismic parameters for different source zone models in the Iberian Peninsula Journal Article Tectonophysics, (717), pp. 449-472, 2017, ISSN: 0040-1951. @article{Amaro-Mellado2017, title = {Comparing seismic parameters for different source zone models in the Iberian Peninsula}, author = {J. L. Amaro-Mellado and A. Morales-Esteban and G. Asencio-Cortes and F. Martinez-Alvarez}, doi = {10.1016/j.tecto.2017.08.032}, issn = {0040-1951}, year = {2017}, date = {2017-01-01}, journal = {Tectonophysics}, number = {717}, pages = {449-472}, abstract = {Seismical parameters of five seismogenic zonings for the Iberian Peninsula have been determined in this work. For that purpose, this research has two key goals. The first is to generate a seismic catalog. The second to calculate the seismical parameters of all the zones of the seismogenic zonings selected. The first key goal has been the creation of a catalog of earthquakes for the Iberian Peninsula and adjacent areas. First, the National Geographic Institute of Spain's catalog has been completed and reviewed with the information from other catalog reviews and specific studies. Second, all magnitude calculations have been homogenized. Third, all dependent data have been eliminated through declustering. Finally, the year of completeness for each magnitude has been considered. The Quaternary active faults database of Iberia has also been used as input data. All of this information has been integrated into a geographic information system. The second key aim is the calculation of the seismical parameters. The first parameter obtained has been the b-value. A method which considers different years of completeness in accordance with the magnitude has been used. Also, the annual rate of earthquakes per square kilometer has been calculated. Moreover, the maximum magnitude known that Quaternary active faults might generate and maximum magnitude recorded in the catalog have been determined. Finally, based solely on the statistical parameters obtained, a critical discussion of the seismogenic zonings of the Iberian Peninsula has been conducted. The results show that some zonings possess insufficient data for a proper calculation of the seismic parameters, from a statistical point of view.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Seismical parameters of five seismogenic zonings for the Iberian Peninsula have been determined in this work. For that purpose, this research has two key goals. The first is to generate a seismic catalog. The second to calculate the seismical parameters of all the zones of the seismogenic zonings selected. The first key goal has been the creation of a catalog of earthquakes for the Iberian Peninsula and adjacent areas. First, the National Geographic Institute of Spain's catalog has been completed and reviewed with the information from other catalog reviews and specific studies. Second, all magnitude calculations have been homogenized. Third, all dependent data have been eliminated through declustering. Finally, the year of completeness for each magnitude has been considered. The Quaternary active faults database of Iberia has also been used as input data. All of this information has been integrated into a geographic information system. The second key aim is the calculation of the seismical parameters. The first parameter obtained has been the b-value. A method which considers different years of completeness in accordance with the magnitude has been used. Also, the annual rate of earthquakes per square kilometer has been calculated. Moreover, the maximum magnitude known that Quaternary active faults might generate and maximum magnitude recorded in the catalog have been determined. Finally, based solely on the statistical parameters obtained, a critical discussion of the seismogenic zonings of the Iberian Peninsula has been conducted. The results show that some zonings possess insufficient data for a proper calculation of the seismic parameters, from a statistical point of view. |
M. J. Fernández-Gómez and G. Asencio-Cortés and A. Troncoso and F. Martínez-Álvarez Large earthquake magnitude prediction in Chile with imbalanced classifiers and ensemble learning Journal Article Applied Sciences, 7 (6), pp. 625, 2017. @article{APSCI2017, title = {Large earthquake magnitude prediction in Chile with imbalanced classifiers and ensemble learning}, author = {M. J. Fernández-Gómez and G. Asencio-Cortés and A. Troncoso and F. Martínez-Álvarez}, url = {https://www.mdpi.com/2076-3417/7/6/625}, doi = {10.3390/app7060625}, year = {2017}, date = {2017-01-01}, journal = {Applied Sciences}, volume = {7}, number = {6}, pages = {625}, abstract = {This work presents a novel methodology to predict large magnitude earthquakes with horizon of prediction of five days. For the first time, imbalanced classification techniques are applied in this field by attempting to deal with the infrequent occurrence of such events. So far, classical classifiers were not able to properly mine these kind of datasets and, for this reason, most of the methods reported in the literature were only focused on moderate magnitude prediction. As an additional step, outputs from different algorithms are combined by applying ensemble learning. Since false positives are quite undesirable in this field, due to the social impact that they might cause, ensembles have been designed in order to reduce these situations. The methodology has been tested on different cities of Chile, showing very promising results in terms of accuracy.}, keywords = {}, pubstate = {published}, tppubtype = {article} } This work presents a novel methodology to predict large magnitude earthquakes with horizon of prediction of five days. For the first time, imbalanced classification techniques are applied in this field by attempting to deal with the infrequent occurrence of such events. So far, classical classifiers were not able to properly mine these kind of datasets and, for this reason, most of the methods reported in the literature were only focused on moderate magnitude prediction. As an additional step, outputs from different algorithms are combined by applying ensemble learning. Since false positives are quite undesirable in this field, due to the social impact that they might cause, ensembles have been designed in order to reduce these situations. The methodology has been tested on different cities of Chile, showing very promising results in terms of accuracy. |
G. Asencio-Cortés and F. Martínez-Álvarez and A. Troncoso and A. Morales-Esteban Medium-Large earthquake magnitude prediction in Tokyo with artificial neural networks Journal Article Neural Computing and Applications, 28 (5), pp. 1043-1055, 2017. @article{NCA2017, title = {Medium-Large earthquake magnitude prediction in Tokyo with artificial neural networks}, author = {G. Asencio-Cortés and F. Martínez-Álvarez and A. Troncoso and A. Morales-Esteban}, url = {http://link.springer.com/article/10.1007/s00521-015-2121-7}, doi = {10.1007/s00521-015-2121-7}, year = {2017}, date = {2017-01-01}, journal = {Neural Computing and Applications}, volume = {28}, number = {5}, pages = {1043-1055}, abstract = {This work evaluates artificial neural networks’ accuracy when used to predict earthquakes magnitude in Tokyo. Several seismicity indicators have been retrieved from the literature and used as input for the networks. Some of them have been improved and parameterized in order to extract more valuable knowledge from datasets. The experimental set-up includes predictions for five consecutive datasets referring to year 2015, earthquakes with magnitude larger than 5.0 and for a temporal horizon of seven days. Results have been compared to four well-known machine learning algorithms, reporting very promising results in terms of all quality parameters evaluated. The statistical tests applied conclude that differences between the proposed artificial neural network and the other methods are significant.}, keywords = {}, pubstate = {published}, tppubtype = {article} } This work evaluates artificial neural networks’ accuracy when used to predict earthquakes magnitude in Tokyo. Several seismicity indicators have been retrieved from the literature and used as input for the networks. Some of them have been improved and parameterized in order to extract more valuable knowledge from datasets. The experimental set-up includes predictions for five consecutive datasets referring to year 2015, earthquakes with magnitude larger than 5.0 and for a temporal horizon of seven days. Results have been compared to four well-known machine learning algorithms, reporting very promising results in terms of all quality parameters evaluated. The statistical tests applied conclude that differences between the proposed artificial neural network and the other methods are significant. |
G. Asencio-Cortes and S. Scitovski and R. Scitovski and F. Martinez-Alvarez Temporal analysis of croatian seismogenic zones to improve earthquake magnitude prediction Journal Article Earth Science Informatics, 3 (10), pp. 303-320, 2017, ISSN: 1865-0481. @article{AsencioCortes2017, title = {Temporal analysis of croatian seismogenic zones to improve earthquake magnitude prediction}, author = {G. Asencio-Cortes and S. Scitovski and R. Scitovski and F. Martinez-Alvarez}, doi = {10.1007/s12145-017-0295-5}, issn = {1865-0481}, year = {2017}, date = {2017-01-01}, journal = {Earth Science Informatics}, volume = {3}, number = {10}, pages = {303-320}, keywords = {}, pubstate = {published}, tppubtype = {article} } |
G. Asencio-Cortés and F. Martínez-Álvarez and A. Morales-Esteban and J. Reyes and A. Troncoso Using principal component analysis to improve earthquake magnitude prediction in Japan Journal Article Logical Journal of the IGPL, 25 (6), pp. 949-966, 2017. @article{IGPL2017, title = {Using principal component analysis to improve earthquake magnitude prediction in Japan}, author = {G. Asencio-Cortés and F. Martínez-Álvarez and A. Morales-Esteban and J. Reyes and A. Troncoso}, url = {https://academic.oup.com/jigpal/article/25/6/949/4565822}, doi = {https://doi.org/10.1093/jigpal/jzx049}, year = {2017}, date = {2017-01-01}, journal = {Logical Journal of the IGPL}, volume = {25}, number = {6}, pages = {949-966}, abstract = {Increasing attention has been paid to the prediction of earthquakes with data mining techniques during the last decade. Several works have already proposed the use of certain features serving as inputs for supervised classifiers. However, they have been successfully used without any further transformation so far. In this work, the use of principal component analysis (PCA) to reduce data dimensionality and generate new datasets is proposed. In particular, this step is inserted in a successfully already used methodology to predict earthquakes. Tokyo, one of the cities mostly threatened by large earthquakes occurrence in Japan, is studied. Several well-known classifiers combined with PCA have been used. Noticeable improvement in the results is reported.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Increasing attention has been paid to the prediction of earthquakes with data mining techniques during the last decade. Several works have already proposed the use of certain features serving as inputs for supervised classifiers. However, they have been successfully used without any further transformation so far. In this work, the use of principal component analysis (PCA) to reduce data dimensionality and generate new datasets is proposed. In particular, this step is inserted in a successfully already used methodology to predict earthquakes. Tokyo, one of the cities mostly threatened by large earthquakes occurrence in Japan, is studied. Several well-known classifiers combined with PCA have been used. Noticeable improvement in the results is reported. |
2016 |
G. Asencio-Cortés and E. Florido and A. Troncoso and F. Martínez-Álvarez A novel methodology to predict urban traffic congestion with ensemble learning Journal Article Knowledge and Information Systems, 20 , pp. 4205–4216, 2016. @article{ASENCIO16, title = {A novel methodology to predict urban traffic congestion with ensemble learning}, author = {G. Asencio-Cortés and E. Florido and A. Troncoso and F. Martínez-Álvarez}, url = {https://link.springer.com/article/10.1007/s00500-016-2288-6}, doi = {https://doi.org/10.1007/s00500-016-2288-6}, year = {2016}, date = {2016-11-01}, journal = {Knowledge and Information Systems}, volume = {20}, pages = {4205–4216}, keywords = {}, pubstate = {published}, tppubtype = {article} } |