Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Genericness of model loading #44

Open
annaproxy opened this issue Jul 18, 2021 · 0 comments
Open

Genericness of model loading #44

annaproxy opened this issue Jul 18, 2021 · 0 comments
Assignees
Labels
enhancement New feature or request

Comments

@annaproxy
Copy link
Collaborator

See #38 (comment)

Right now, model loader reconstructs model according to state_dict by using size of lstm.hh, which is not awful but also not too beautiful if the architecture changes to e.g. two layers, or if the model parameter names slightly change (e.g. lstm to rnn).

Could definitely be improved even though it is not an issue for models right now.

(The most generic solution would be to save the model directly (as a "binary") and not a state_dict, which makes the model instantly torch.load-able, but does introduce dependencies on the current exact RNN implementation/location of the definition, torch version, etc, which is why it is not the recommended way to save.)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant