You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi Wafaa!
Currently, Stopes is focused only on translation models, and ALTI+ was implemented only for seq2seq transformers, such as NLLB. We are not currently planning to adapt ALTI+ to other architectures.
If I learn that our other colleagues who are working with LLMs implement ALTI+ or similar attribution methods for their models, I will update this thread accordingly.
Otherwise, I suggest that maybe you make such a contribution. LLaMA is a decoder-only transformer, so the part of ALTI+ code responsible for the decoder can in principle be adapted to it. But of course, this adaptation will depend on the chosen framework and other exact implementation details of the LLM code.
Can you offer support for the ALTI attribution method for LLMs such as LLAMA?
The text was updated successfully, but these errors were encountered: