-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Discussion of methods alternative to CNNs (e.g. RNNs or transformers) #5
Comments
So we would be able to incorporate time-invariant features (such as population size, etc) to the RNN, which is great news! Roughly, it would update the hidden state as: It is simple to implement it:
@gvegayon: What would be the time-invariant variables we'd like to include? |
Also, after reading about options and thinking about our problem. I think that LSTM (which is a fancy RNN) might be an overkill, given that our time series isn't that long and it is univariate. In fact, we might even have problems with overfitting the data. My suggestion:
Happy to talk about it more next time we meet. |
Thank you so much, Bernardo! |
That's a great question, @modenesi. For the moment, the only one I can think of is population size. We could add other things, but generally that would make the model less usable. For instance, we could add Rt and generation interval estimates, but that information is not always available. |
From chatGPT, I asked for a specific architecture for a simple RNN in R, with 1 time invariant and 1 time variant variable:
|
Hi all,
I'm not used to using issues on github, but I love the idea and I want to be using it more. Feel free to give me feedback on it anytime.
I just want to start an issue to discuss whether or not we want to implement models alternative to CNNs in order to make our ML model able to take inputs of any size. It might be that the current joint CNN solution that you both have is enough to tackle the issue. I'll try to learn a bit more about it.
ALTERNATIVE SOLUTIONs:
It might also be interesting to try alternative models, especially if the cost of running them is low? I wonder if we can handle our current code to chatGPT and ask it to write a 2nd model directly suited for time series. It might even capture temporal aspects better than CNNs. Some options:
I need to check, but I think we want to train these models w/ very long vectors, but with padding and masking. We can talk more about it during our meeting today.
I read that RNNs or transformers are great to model temporal dependencies, while CNNs with global pooling is more efficient to capture broad patters in the time sequence.
The text was updated successfully, but these errors were encountered: