Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[RL agent] study the impact of model scaling and model architecture in the overall performances #25

Open
3rdCore opened this issue Feb 15, 2022 · 0 comments
Labels
help wanted Extra attention is needed

Comments

@3rdCore
Copy link
Collaborator

3rdCore commented Feb 15, 2022

The current RL agent uses a very simple GNN +MLP architecture. From the supremacy of ResNET to the advent of autoregressive Transformers, the latest papers in either natural language processing or image processing have shown the benefits of using extremely complex architectures sometimes made of more than a billion parameters.

To what extent these conclusions can be applied to SeaPearl ? Why do we need very little CPU computing power to achieve good results ? I would be very interesting to study given a specific CP problem, the performance of different deep-NN architectures and to more widely study the impact of different scaling parameters regarding the input data size, the model size, the computing power available.

@3rdCore 3rdCore added the help wanted Extra attention is needed label Feb 15, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
help wanted Extra attention is needed
Projects
None yet
Development

No branches or pull requests

1 participant