Code Monkey home page Code Monkey logo

crossvivit's People

Contributors

danassou avatar gitbooo avatar jaggbow avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar

crossvivit's Issues

Fluctuating and Non-decreasing Loss during CrossViVit Training on my own dataset

I really appreciate your outstanding work! The work you have done on the fusion of image and time-series modalities is exactly the direction I have been researching recently. Therefore, I tested your model on my own dataset and compared it with some models I have used before, but I found that CrossViVit did not perform as well as expected. I would appreciate some advice on how to improve the forecasting performance of the model.

Similar to the dataset used in your work, my dataset consists of 15-minute satellite images with only one channel (image size = 96*96) and corresponding photovoltaic power. However, the difference is that the time span of the data I used is only 2.8 years, and I tested it on the last 5 months. The model takes in the past 4 hours of data (step=16) and predicts the future 4 hours of photovoltaic power (step=16).

I trained the model using the default parameters of CrossViVit as in your experiments, with the following differences:

  1. Optical flow was not used.
  2. Loss criterion = nn.MSE (the paper used L1 loss?).
  3. AdamW optimizor with a learning rate of 0.001.
  4. The batch_size was 16.

During the experiments, the train loss and valid loss kept fluctuating and did not decrease. I am not sure what the issue might be. Below are the train loss and valid loss from the wandb logs. I compared them with Perceiver-RNN, which is a model used in the OCF project: https://github.com/openclimatefix/predict_pv_yield (experiment/003*.py)

train_loss: https://api.wandb.ai/links/740402059/uji2orxi

train_step_loss: https://api.wandb.ai/links/740402059/lwz71qqh

valid_loss: https://api.wandb.ai/links/740402059/k7a2abvx

I believe that the cross-attention used in CrossViVit would perform better compared to the concatenation used in Perceiver-RNN, but I am not sure what might be causing the loss to not decrease.

Share parameters

Hi guys,
Thank you for sharing your amazing work!
For training I am using the dataset posted here https://app.activeloop.ai/crossvivit/SunLake, which can be accessed with deeplake.
Furthermore, I am using some of the parameters which you have shared in your paper, however, I got some errors that mat1 could not be multiplied with mat2, because of different matrix size.
In order to solve this little problem, I have changed ctx_channels from 8 to 23, ts_channels from 16 to 8 in cross_vivit_bis.yaml.
However, I am not feeling that it is the correct solution to the problem.
Could you please share all your parameters from cross_vivit_bis.yaml, txcontext_datamodule.yaml and cross_vivit.yaml?

the data in PCCI_20082022_IZA seems to be broken

Thanks for great work!
I got an IndexError after I moved PCCI_20082022_IZA to the training split to be same as the paper.
I was wondering the data in PCCI_20082022_IZA whether it is broken. Is that correct or I need to redownload the section of data in PCCI_20082022_IZA ?
(I'm asure this problem is caused by PCCI_20082022_IZA, since I removed it from training split and everything goes fine)

Here is the detail of my configuration in forecast_datamodule.yaml
stations: train: ["PCCI_20082022_IZA", "PCCI_20082022_CNR", "PCCI_20082022_PAL"] val: ["PCCI_20082022_PAY"] test: ["PCCI_20082022_CAB", "PCCI_20082022_TAM"]

Spatio-temporal prediction

Awesome work and it is really interesting that the model seems to learn physics ? Or do you think the good results are because of using Locations in the Center of the satellite Images ?

Nevertheless, i asked myself whether it is possible to generate predictions for locations without observed GHI?

E.g. providing the latest satellite images and all available observations and Receive a prediction for and other Location in the area ?

Thanks and best regards!

Which is a better approach: forecast satellite image pixels first or direct forecasting (in this paper)?

I have a question about using information from satellite cloud images. Previous studies mostly involved establishing spatiotemporal forecast models to forecast satellite images first, and then using the forecasted results of corresponding pixel points in satellite cloud images as part of the features, along with time series data, for another round of forecast. This staged approach is easy to comprehend and yields decent results in my tests.

In this work, you directly built an end-to-end model for the fusion of the two types of data. I believe this approach could better utilize information from other pixel points in the satellite images . However, I understand that training the model is more challenging compared to the previous method.

Therefore, I would like to inquire if you have compared the method of forecasting satellite cloud images first and whether direct forecasting is superior to staged forecasting. Additionally, how do you perceive these two methods?

Can you please share the weight of the model and a testing script?

Unfortunately, to train the model, it is required expensive hardware RTX8000 (6000$ + per one), but you have trained your model on 2. Can you please share the weights which you have trained?

Also, could you please share a testing script, where the parameters are, for example, 1-2, or 7 days historical time series, 1-2 or 7 days satellite images and the model?

Dataset in the paper

Hi, I enjoyed the paper and wanted to check the dataset that you presented in the paper. Do you think it would be possible to access it ?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.