The University of Southampton
University of Southampton Institutional Repository

Vision transformers for estimating irradiance using data scarce sky images

Vision transformers for estimating irradiance using data scarce sky images
Vision transformers for estimating irradiance using data scarce sky images
Accurate estimation of diffuse horizontal irradiance (DHI) is critical for optimising photovoltaic system performance and energy forecasting yet remains challenging in regions lacking comprehensive ground-based instrumentation. Recent advancements using Vision Transformers (ViTs) trained on extensive sky image datasets have shown promise in replacing costly irradiance measurement equipment, but the scarcity of long-term, high-quality sky imagery significantly restricts practical implementation. Addressing this critical gap, this study proposes a novel dual-framework approach designed for data-scarce scenarios. First, calculated atmospheric parameters, including extraterrestrial irradiance and cyclic time encodings, are integrated to represent sky conditions without utilising any instrumentation. Next, a sequential pipeline initially predicts synthetic global horizontal irradiance (GHI) and uses it as a feature, to refine DHI estimation. Finally, a dual-parallel architecture simultaneously processes raw and overlay-enhanced fisheye sky images. Overlays are generated through unsupervised, physics-informed cloud segmentation to highlight dynamic sky features. Empirical validation is performed using data from the Chilbolton Observatory, chosen for its temperate climate and frequent cloud variability. To simulate data-scarce conditions, models are trained on a single month (e.g., January) and evaluated across a temporally disjoint, full-year test set. Under this setup, the sequential and dual-parallel frameworks achieve RMSE values within 2–3 W/m² and 1–6 W/m², respectively, of a state-of-the-art ViT trained on the complete dataset. By combining physics-informed modelling with unsupervised segmentation, the proposed method provides a scalable and cost-effective solution for DHI estimation, advancing solar resource assessment in data-constrained environments.
Computer Vision, Machine learning, Sky imaging, Solar irradiance, Vision transformer (ViT), Computer vision
2666-5468
Hamlyn, David
cf128cee-171b-45d1-9963-1dead6092167
Chaudhary, Sunny
25f0d213-03ef-4909-8cfc-29a8498aa28f
Rahman, Tasmiat
e7432efa-2683-484d-9ec6-2f9c568d30cd
Hamlyn, David
cf128cee-171b-45d1-9963-1dead6092167
Chaudhary, Sunny
25f0d213-03ef-4909-8cfc-29a8498aa28f
Rahman, Tasmiat
e7432efa-2683-484d-9ec6-2f9c568d30cd

Hamlyn, David, Chaudhary, Sunny and Rahman, Tasmiat (2025) Vision transformers for estimating irradiance using data scarce sky images. Energy and AI, 21, [100560]. (doi:10.1016/j.egyai.2025.100560).

Record type: Article

Abstract

Accurate estimation of diffuse horizontal irradiance (DHI) is critical for optimising photovoltaic system performance and energy forecasting yet remains challenging in regions lacking comprehensive ground-based instrumentation. Recent advancements using Vision Transformers (ViTs) trained on extensive sky image datasets have shown promise in replacing costly irradiance measurement equipment, but the scarcity of long-term, high-quality sky imagery significantly restricts practical implementation. Addressing this critical gap, this study proposes a novel dual-framework approach designed for data-scarce scenarios. First, calculated atmospheric parameters, including extraterrestrial irradiance and cyclic time encodings, are integrated to represent sky conditions without utilising any instrumentation. Next, a sequential pipeline initially predicts synthetic global horizontal irradiance (GHI) and uses it as a feature, to refine DHI estimation. Finally, a dual-parallel architecture simultaneously processes raw and overlay-enhanced fisheye sky images. Overlays are generated through unsupervised, physics-informed cloud segmentation to highlight dynamic sky features. Empirical validation is performed using data from the Chilbolton Observatory, chosen for its temperate climate and frequent cloud variability. To simulate data-scarce conditions, models are trained on a single month (e.g., January) and evaluated across a temporally disjoint, full-year test set. Under this setup, the sequential and dual-parallel frameworks achieve RMSE values within 2–3 W/m² and 1–6 W/m², respectively, of a state-of-the-art ViT trained on the complete dataset. By combining physics-informed modelling with unsupervised segmentation, the proposed method provides a scalable and cost-effective solution for DHI estimation, advancing solar resource assessment in data-constrained environments.

Text
Vision transformers for estimating irradiance using data scarce sky images - Version of Record
Available under License Creative Commons Attribution.
Download (5MB)

More information

Submitted date: 8 May 2025
e-pub ahead of print date: 14 July 2025
Published date: 29 July 2025
Keywords: Computer Vision, Machine learning, Sky imaging, Solar irradiance, Vision transformer (ViT), Computer vision

Identifiers

Local EPrints ID: 504862
URI: http://eprints.soton.ac.uk/id/eprint/504862
ISSN: 2666-5468
PURE UUID: a93bf9a7-cb05-413e-9e54-f37cd7682093
ORCID for David Hamlyn: ORCID iD orcid.org/0009-0005-9801-595X
ORCID for Sunny Chaudhary: ORCID iD orcid.org/0000-0003-2664-7083
ORCID for Tasmiat Rahman: ORCID iD orcid.org/0000-0002-6485-2128

Catalogue record

Date deposited: 19 Sep 2025 16:47
Last modified: 20 Sep 2025 02:28

Export record

Altmetrics

Contributors

Author: David Hamlyn ORCID iD
Author: Sunny Chaudhary ORCID iD
Author: Tasmiat Rahman ORCID iD

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×