Oct 27 – 31, 2025
Europe/Stockholm timezone

Session

APL2 - Bridging the Gap: Reproducibility, Deployment, Operation, Updates, and Monitoring of Machine Learning-Based Solar Flare Forecast Models

APL2
Oct 31, 2025, 8:30 AM
Miklagård

Miklagård

Conveners

APL2 - Bridging the Gap: Reproducibility, Deployment, Operation, Updates, and Monitoring of Machine Learning-Based Solar Flare Forecast Models: Orals

  • Sophie Murray (Dublin Institute for Advanced Studies)
  • Anna Maria Massone (Dipartimento di Matematica, Università degli Studi di Genova)
  • Paul Wright (University of Exeter)
  • Shane Maloney (Dublin Institute for Advanced Studies)

Description

Machine learning (ML) has shown promise in solar flare forecasting, yet major challenges remain in moving beyond single studies often published without the associated code and datasets. Are forecast models truly reproducible, deployable, operable, updatable, and monitorable? It has been nearly a decade since the influential work of Bobra & Couvidat (2015), where Support Vector Machine (SVM) was used to predict solar flares using features derived from solar magnetic field observations, what progress has been made? While the field has made strides in improving flare forecasting, little to no progress has been made in ensuring the reproducibility, deployment, and sustained operation of ML models. As models become increasingly complex, how can we guarantee that their results are reproducible and transparent? How do we ensure that models remain accurate as new data becomes available and do not degrade over time? The traditional approach of publishing research papers—often without the accompanying code, data, or reproducibility frameworks—is no longer sufficient. To truly advance the field, we must move beyond current academic practices and adopt best practices from ML and meteorology, which have well-established methodologies for real-time prediction systems. This includes:

Reproducibility: Establishing standardized benchmarks, dataset and model versioning, and open-source implementations.
Deployment: Models must be easily deployed from zero to a running deployment with as little human intervention as possible.
Operation: Ensuring once deployed models can easily be run in research or operational environments and they are robust to missing data, latency, and changing solar cycle conditions.
Updates: Implementing retraining strategies to include new data and prevent model degradation over time.
Monitoring: Developing frameworks for continuous evaluation, explainability, and reliability of forecasts.
We must leverage platforms like Hugging Face, Kaggle, Comet, Neptune, WandB and open source solutions like MLFlow to facilitate transparent and collaborative development. The space weather community must also engage MLops, automated retraining pipelines, and robust monitoring tools to transition ML-based forecasting from one off publications to an operational reality.

This session invites submission focusing on the use or implementation of any of the above aspects and interdisciplinary approaches to move ML-based space weather forecasting from promise to practice.

Presentation materials

There are no materials yet.
Building timetable...