Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Distilbert 모델 추가합니다. #42

Merged
merged 10 commits into from
Dec 21, 2021
Merged

Distilbert 모델 추가합니다. #42

merged 10 commits into from
Dec 21, 2021

Conversation

abbymark
Copy link
Contributor

모델 동작까지 확인하였습니다.
다만 성능이 예상보다 많이 떨어집니다.

기존 monologg님의 distilkobert를 encoder와 decoder로 사용하였습니다.
encoder layer 3개, decoder layer 3개 로써 기존 distilbert 모델 그대로 사용하였습니다.

5에폭에 rougelsum은 12가 약간 넘는 수치였습니다.
대신 eval time이 기존 bart모델의 encoder layer 3, decoder layer 3 보다는 빨랐습니다.
https://wandb.ai/final_project/optimization/runs/z88c6i2r?workspace=user-abbymark

@abbymark abbymark requested a review from gistarrr December 18, 2021 12:02
@changyong93 changyong93 merged commit 295fc22 into dev Dec 21, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants