Machine learning models based on thermal data predict solar radiation
A research team at the University of Córdoba has developed and evaluated models for the prediction of solar radiation in nine locations in southern Spain and North Carolina (USA).
Measuring solar radiation is costly, as are all the tasks related to the maintenance and calibration of the most commonly used sensors: pyranometers and radiometers. The result is a paucity of reliable data. Hence, a research group from the University of Córdoba has developed and evaluated several Machine Learning models to predict solar radiation in nine locations (southern Spain and North Carolina, USA) spanning a range of different geo-climatic conditions (aridity, distance to the sea, and elevation). The work has been featured in the journal Applied Energy.
One of the main breakthroughs of the models they created is that they only need thermal data to estimate daily solar radiation. “Measuring and having air temperature data today, thanks to low-cost sensorization and IoT (Internet of Things) technologies, is quite affordable,” added Javier Estévez Gualda, a researcher and professor at the UCO. Most of the existing meteorological stations around the world have thermal and rainfall sensors, but very few measure solar radiation.
Juan Antonio Bellido, the main author of the work, highlights that one of the problems that currently exist in models based on Artificial Intelligence is the configuration of internal parameters, termed hyperparameters. “We might compare these hyperparameters to the controls on a sound technician’s mixing board; their adjustment is essential, as the modification of potentiometers must be constantly prevented from causing sound problems,” he explains. To solve this problem, they used an automatic algorithm called Bayesian Optimization (based on Bayes’ theory), which is responsible for efficiently and quickly searching for suitable parameters so that the models obtain efficient and accurate results.
All the models used can be classified as supervised models; that is, they require data training. However, they can be classified into several groups: models based on neural networks, whose functioning is similar to that of neurons (Multilayer Perceptron—MLP, Extreme Learning Machine—ELM—and Generalized Regression Neural Networks—GRNN); models based on a tree typology (Random Forest—RF—and Extreme Gradient Boost—XGBoost), and others, such as Support Vector Machine (SVM).
These have been tested in different arid conditions, such as southern Spain, and in North Carolina (USA); that is, in arid, semi-arid, sub-humid, dry and humid areas. 31% of the earth’s surface is arid/semi-arid/dry sub-humid, while 67% is humid. In this way, they are applicable to the entire surface of the Earth. When evaluating the efficiency of the models trained in new locations whose data has not been used during the same, significant improvements are yielded in all the locations. For example, those trained with time series from Cabra (Córdoba) and then applied in El Campillo (Huelva) show a considerable improvement.
Professor and researcher Amanda P. García-Marín, who is also part of the team, stated that the results improve on current models, obtaining great precision in estimating daily solar radiation values. “This is crucial in locations with no available or missing/low-quality data sets, and can be used to optimize the determination of potential locations for the construction of solar power plants,” concluded researcher Javier Estévez Gualda.
Another of the work’s strengths is that the models are available to any researcher,via the free-access GitHub repository, so their use can be extended to any area of the planet, depending on its aridity. In addition, from the agronomic point of view, the precise estimation of solar radiation is vital, since it is a key variable in the development of crops.
This study is part of the Smarity project (AGL2017-87658-R), funded by the Ministry of Science’s Social Challenges National R&D Plan, which seeks solutions to climatic aridity in southern Spain and to intelligently predict the spatio-temporal variability of aridity and its effects on agriculture and the environment.
This article originally appeared on techxplore.com, to read the full article, click here.
Nastel Technologies helps companies achieve flawless delivery of digital services powered by middleware. Nastel delivers Middleware Management, Monitoring, Tracking, and Analytics to detect anomalies, accelerate decisions, and enable customers to constantly innovate. To answer business-centric questions and provide actionable guidance for decision-makers, Nastel’s Navigator X fuses:
Nastel Technologies is the global leader in Integration Infrastructure Management (i2M). It helps companies achieve flawless delivery of digital services powered by integration infrastructure by delivering Middleware Management, Monitoring, Tracking, and Analytics to detect anomalies, accelerate decisions, and enable customers to constantly innovate, to answer business-centric questions, and provide actionable guidance for decision-makers. It is particularly focused on IBM MQ, Apache Kafka, Solace, TIBCO EMS, ACE/IIB and also supports RabbitMQ, ActiveMQ, Blockchain, IOT, DataPower, MFT and many more.
The Nastel i2M Platform provides:
- Secure self-service configuration management with auditing for governance & compliance
- Message management for Application Development, Test, & Support
- Real-time performance monitoring, alerting, and remediation
- Business transaction tracking and IT message tracing
- AIOps and APM
- Automation for CI/CD DevOps
- Analytics for root cause analysis & Management Information (MI)
- Integration with ITSM/SIEM solutions including ServiceNow, Splunk, & AppDynamics