You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
With transformers=3.31.0 and accelerate=0.22.0.dev0 (latest main) On 2 x A6000 48G, fine-tuning LLaMA 7B , the warning is:
FutureWarning: using `--fsdp_transformer_layer_cls_to_wrap` is deprecated. Use fsdp_config instead
FSDP Warning: When using FSDP, it is efficient and recommended to call prepare for the model before creating the optimizer.
FSDP Warning: When using FSDP, several parameter groups will be conflated into a single one due to nested module wrapping and parameter flattening.
I use this hacking method to revise the code in trainer.py ; the warning is disappeared but still hit OOM;
Still warning:
FutureWarning: using `--fsdp_transformer_layer_cls_to_wrap` is deprecated. Use fsdp_config instead
the following warning disappear:
FSDP Warning: When using FSDP, it is efficient and recommended to call prepare for the model before creating the optimizer.
FSDP Warning: When using FSDP, several parameter groups will be conflated into a single one due to nested module wrapping and parameter flattening.
With transformers=3.31.0 and accelerate=0.22.0.dev0 (latest main) On 2 x A6000 48G, fine-tuning LLaMA 7B , the warning is:
And i use the default command:
I install fastchat from source.
And i also try your solution, it is also not worked.
I found you have the problem, please help me to fix this !
The text was updated successfully, but these errors were encountered: