You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Sep 7, 2023. It is now read-only.
Combing multiple ML models often wins Kaggle competitions.
The ensemble could include:
Model which takes single step of satellite imagery from optical flow and predicts PV yield.
Multiple Optical flow predictions using flow computed at t-1, t-2, etc.
Model which predicts PV yield from ML predictions of future imagery (from SatFlow).
NWP to PV.
Similar cloud sequences from the past.
Context
Whilst ensemble models work well in many instances, perhaps our Perceiver IO will already perform as well as an ensemble model?
Possible Implementation
Maybe our "joint model" (based on Perceiver IO) would be the "meta model" which combines the ensemble? Maybe the joint model directly just gets historical data (satellite, historical pv) and the predictions from component models as inputs.
From a software engineering perspective, it might be nice to have several discrete models.
The text was updated successfully, but these errors were encountered:
Detailed Description
Combing multiple ML models often wins Kaggle competitions.
The ensemble could include:
Context
Whilst ensemble models work well in many instances, perhaps our Perceiver IO will already perform as well as an ensemble model?
Possible Implementation
Maybe our "joint model" (based on Perceiver IO) would be the "meta model" which combines the ensemble? Maybe the joint model directly just gets historical data (satellite, historical pv) and the predictions from component models as inputs.
From a software engineering perspective, it might be nice to have several discrete models.
The text was updated successfully, but these errors were encountered: