You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi,
I am working to prune BART model for seq2seq purpose. Currently, I have replaced this code with BART based functionalities. After executing I am getting drop in number of parameters for both attention and FFN but dimension reduction happens only for FFN which results in slowness. My questions are following:
Hi, I am working to prune BART model for seq2seq purpose. Currently, I have replaced this code with BART based functionalities. After executing I am getting drop in number of parameters for both attention and FFN but dimension reduction happens only for FFN which results in slowness. My questions are following:
Hi,
I am working to prune BART model for seq2seq purpose. Currently, I have replaced this code with BART based functionalities. After executing I am getting drop in number of parameters for both attention and FFN but dimension reduction happens only for FFN which results in slowness. My questions are following:
The text was updated successfully, but these errors were encountered: