-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add test for deepseek_vl #1136
base: main
Are you sure you want to change the base?
Add test for deepseek_vl #1136
Conversation
d680aab
to
07bb2f6
Compare
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Don't we have HuggingFace versions for which we don't need to copy model implementations?
|
07bb2f6
to
b774e54
Compare
|
1 similar comment
|
b774e54
to
42bf67b
Compare
|
3 similar comments
|
|
|
eccc9ef
to
9baabd6
Compare
|
1 similar comment
|
a9a7143
to
563b03e
Compare
|
2 similar comments
|
|
|
adfe853
to
273ee1b
Compare
return framework_model, vl_gpt, tokenizer, inputs_embeds | ||
|
||
|
||
def generation(max_new_tokens, model, inputs_embeds, tokenizer, vl_gpt): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Great work @meenakshiramanathan1 . Can we reuse this script for other generation models like t5, llama, and more? BTW, we can expect a conflict related to vl_gpt
used here.
cc: @nvukobratTT May I know your thoughts in this?
273ee1b
to
d75cca4
Compare
d75cca4
to
8373b96
Compare
Model doesn't have huggingface direct version, hence model implementation is from corresponding github repository. |
b3431b2
to
20c7194
Compare
|
3 similar comments
|
|
|
…ps config (#1231) The [generate model ops test pipeline](https://github.com/tenstorrent/tt-forge-fe/actions/runs/13328380954/job/37226649520) is currently freezing during the unique ops configuration extraction phase Error: `Failed on "DecomposeEinsum" TVM callback` This error is encountered in the test case: `forge/test/models/pytorch/vision/detr/test_detr.py::test_detr_segmentation[facebook/detr-resnet-50-panoptic]` test cases. To prevent the extraction process from hanging indefinitely, a timeout of 1200 seconds (20 minutes) has been added. This ensures that if the unique ops configuration extraction takes too long, the test will be terminated.
20c7194
to
76b1781
Compare
|
|
1 similar comment
|
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
As this is a lot of new code that is mostly copied from GitHub, let me see can we get new repository for this purpose, and then link it to ours.
E.g. have a repo that will contain all third-party GitHub repositories as submodules, from which we can load this and similar models to run compile against them. That way, in FFE we'll only store model tests + wrappers if some modification is needed.
What do you think @meenakshiramanathan1 ?
Sure @nvukobratTT , I will keep this in draft for now. |
That sound good @nvukobratTT . We can do that. As we are going to focus on P1 models and models supported in tt-torch which is not yet supported in tt-forge-fe. What will be the priority of moving all the GitHub copied code to a new repo? Because there are some other models also fall in this case as you know. Could you please clarify this? |
Synced offline |
Model doesn't have huggingface direct version, hence model implementation is from corresponding github repository.
Model requires timm version>=0.9.16 which was conflicting with segmentation_models_pytorch package so both versions have been upgraded to resolve conflicts.
The custom generate function generates text by iteratively predicting tokens while maintaining a fixed sequence length using padding.
The model input shape is kept static by preallocating padded_inputs_embeds, which is updated each step
with new token embeddings corresponding to next_token_id until the sequence is complete.
Model is currently failing in verify with pcc drop of 0.96.