Products
menu item
Prospect
Quick estimate of PV site's solar potential
menu item
Evaluate
Time Series & TMY data for energy modelling
menu item
Monitor
Real-time PV output assessment
menu item
Forecast
Solar power output forecast for up to 14 days
menu item
Analyst
Simplified & unified solar data management
menu item
Integrations
Automate delivery of Solargis data
Use cases
menu item
Site selection
Find the right solar project location
menu item
Energy yield simulation
Analyze potential gains and risks
menu item
Optimizing power plant design
Find optimum power plant design
menu item
Real power plant performance
Discover the true output
menu item
Power output forecast
Predict solar project energy output
menu item
Ground data verification
Verify quality of solar & meteo measurements
Solar Resource & Meteo Assessment
Detailed solar resource validation and assessment
Site Adaptation of Solargis Models
Combining satellite data with on-site measurements
Quality Control of Solar & Meteo Measurements
Correction of errors in ground-measured data
Customized GIS Data
Customized Solargis GIS data for your applications
PV Energy Yield Assessment
Estimated energy uncertainties and related data inputs
PV Performance Assessment
Energy estimate for refinancing or asset acquisition
PV Variability & Storage Optimization Study
Understand output variability across wide geo regions
Regional Solar Energy Potential Study
Identification of locations for solar power plants
Our expertise
How our technology works
Methodology
How we transform science into technology
API & integration
How to integrate Solargis data via API
Product guides & documentation
Release notes
Success stories
Blog
Ebooks & Whitepapers
Webinars
Collaterals
Publications
Events
Free Maps & GIS Data
Solar performance maps
About Solargis
Partners
ISO Certification
Careers

Solargis’ Technical Director, Tomas Cebecauer, and Managing Director, Marcel Suri speak about the core ingredients for a reliable database, Solargis’ ongoing efforts to enhance and validate its data services, and how users of Solargis’ data can most effectively undertake their own validation using on-site measurements.

marcel tomas interview

Why is a reliable solar database important for the continued global growth of the solar industry?

Marcel: As solar energy continues to deliver ever lower energy costs than traditional forms of power generation, so it has expanded around the world. We’ve seen increasing uptake of solar in countries where it had not previously been competitive, or where infrastructure is less developed.

It is vital that solar data keeps pace with the rapid expansion of the industry, supporting asset owners and operators as they optimise ever more complex projects and ensure correct valuations.

What are the core ingredients of a reliable database?

Tomas: A good model must be able to handle outputs from a range of different sources, and ensure that any conflict is handled in a stable manner to ensure data is representative - for example in overlapping satellite zones or when modern satellites replace old ones.

A special challenge was aligning the modern satellite data with the older archives via in-house developed geometric and radiometric pre-processing tools. These helped us to include valuable long-term history, while assuring its high positional and temporal stability and overall reliability. Today Solargis archives are complete and date back to years 1994, 1999 and 2006, respectively, depending on the satellite region.

With the expansion of solar energy globally, it is essential to accurately represent all climate zones and geographical peculiarities. The data should work just as well in tropical rainforests or high latitudes as in the extreme deserts of Atacama.

Good models should deliver reliable outputs for all primary parameters: Global Horizontal Irradiance (GHI), Direct Normal Irradiance (DNI) and Diffuse Horizontal Irradiance (DIF) – and these must be in equilibrium with each other.

Typically, ensuring the accuracy of DNI data is more demanding. However, without good DNI, calculation of Global Tilted Irradiance (GTI) – which is vital for trackers – can be subject to large errors. A database that boasts highly accurate GHI, but falls short on DNI, may not be fit for purpose in many contexts.

What we observe with some other data products is ‘overfitting’, where the model is forced to match on-site measurements perfectly, while disregarding imperfections and the resulting uncertainty in those on-site measurements. There must be low bias, but crucially, also low Root Mean Square Error (RMSE) at all points. Critical - for a good model - is to maintain also a harmonised representation at the level of sub-hourly data values in all energy ranges. Again, balance is important – no single measurement, parameter or metric should dominate. This is indeed tricky, as we are working with imperfect reference data - the ground measurements.

The model should also be able to differentiate different types of aerosols and clouds properly, including high and rapidly changing atmospheric turbidity, occurrence of snow, ice, fog and thin clouds. Without this, it is hard to attain reliable data for mountainous, densely populated, industrial and coastal areas.

Finally, data must be available at high spatial resolution to accurately reflect unique site conditions. A digital terrain model, such as the one we have harmonised from multiple sources and implemented within the Solargis Prospect app, makes it possible to calculate global data layers at a grid resolution of 90 metres.

What has Solargis done to ensure maximum reliability from day one?

Marcel: At Solargis, we have not only put years of R&D effort into ensuring our models are robust and accurate, we have also ensured that our solar datasets are the most extensively validated.

We decided to develop our own computational and delivery systems even before we delivered our first commercial data products. At the beginning Dr. Richard Perez (a leading academic and professor at the State University of New York) provided support with definition of the models, and later we developed our own solar-computing methodology, also published in several academic papers.

We started in Europe and extended the coverage globally in steps. It took us several years to acquire the necessary historical satellite images. There were sticking points, such as availability of data on Aerosol Optical Depth, but now there are several options.

Tomas: We knew that it was important to ensure fast delivery to make sure asset owners had the most up to date intelligence, down to only a few minutes globally.

For the optimisation of a PV design, the ability to deliver data in 10 and 15-minute resolution is needed. Hourly data is not sufficient, anymore. It became also important to ensure reliable distribution of sub-hourly values, especially high and low energy values for the correct evaluation of combined performance of PV modules and inverters. We are increasingly receiving requests for 1-minute data to enable the simulation of hybrid PV- battery solutions.

We use data from meteorological models and global data processing projects to further increase the accuracy of our energy simulation models and PV performance analysis. This data is well-validated and in case of parameters such as air temperature, and atmospheric pressure, also recomputed at higher resolution and better accuracy. We have also developed a new high-resolution global albedo database and calculation scheme based on a complex methodology combining the best features of several public databases.

Marcel: Extensive validation is a critical step in demonstrating the accuracy of satellite-based solar data used to make development, financing and operational decisions. To prove reliability requires rigorous validation of all solar components. We have validated the model at over 1000 sites across the world, far in excess of other players in the market.

Ultimately, this combination of technology and experience enables us to save our customers money. Every 1% reduced uncertainty increases return on equity by approximately 5% for the investor.

What measurement data should you provide to support validation?

Tomas: Our customers can validate our data themselves, via one-to-one comparison to their on-site meteorological measurements or to PV production data. This further solidifies trust in our processes, once people see for themselves how accurate our data and energy calculations are.

Unfortunately, solar measurements are prone to many errors and issues, therefore prior to any data comparison, it’s important to control the quality of the measured data – otherwise the comparison suffers from errors. Optimally the evaluation of comparison should take into account the uncertainty of ground measurements.

Typical errors we see in solar measurements are due to polluted sensors, shading, misaligned instruments and outdated calibration. In addition, there can be missing records due to issues with data loggers, power supply or incorrect data manipulation and postprocessing.

Even solar data coming from high-grade scientific measuring networks suffer from various errors and need to be rigorously quality-screened before further use.

Marcel: The good news for the industry is that there are already professional commercial measurement providers who offer reliable services, helping to ensure smooth continuity between historical data, real-time updates and forecasting.

We believe strongly in transparency – everybody can see key data features and long-term values of Solargis data parameters globally, within the free version of our Prospect app. Reliable data come with a guarantee of validation: Solargis time series are available for any location and can be validated if good quality local measurements exist.

How can better data deliver better results for solar asset owners?

Marcel:  Our data can be used at all stages of the project lifecycle – from prefeasibility assessments, throughout project development, engineering and financial calculations. After the power plant enters operation, our models continue supplying data for regular performance evaluation and solar power forecasting.

Through our consultancy services, we also help large-scale asset owners to re-evaluate initial assessments on the basis of the improved accuracy and reduced uncertainty via site-adaptation of the models with the use of high quality local measurements.

Tomas: Solargis data and services are constantly undergoing further refinement and long-term development – we are systematically working on making the calculation schemes more accurate, while continuing to develop new models which make better use of the wealth of data from more detailed modern observation systems.

This allows us to plan for delivery of new products and services such as the new tools for manipulation, visualization, quality control and analysis of large-volume time series data. We are also focusing on developing tools for more efficient combined use of measurements and modelled data for reduced uncertainty. Ultimately, our photovoltaic simulation software will benefit from the improved data sets and support tools.

Keep reading

Why to use satellite-based solar resource data in PV performance assessment
Best practices

Why to use satellite-based solar resource data in PV performance assessment

It is widely accepted that high-standard pyranometers operated under rigorously controlled conditions are to be used for bankable performance assessment of photovoltaic (PV) power systems.

How Solargis is improving accuracy of solar power forecasts
Best practices

How Solargis is improving accuracy of solar power forecasts

Just as there are horses for courses, different forecasting techniques are more suitable depending on the intended forecast lead time.

How to calculate P90 (or other Pxx) PV energy yield estimates
Best practices

How to calculate P90 (or other Pxx) PV energy yield estimates

To assess the solar resource or energy yield potential of a site, we model the solar resource/energy yield using best available information and methods. The resulting estimate is the P50 estimate, or in other words, the “best estimate”. P50 is essentially a statistical level of confidence suggesting that we expect to exceed the predicted solar resource/energy yield 50% of the time. However, ...