You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Your question
How to save a sharded checkpoint when using MoE parallel folding -- specifically, when ETP != TP? save_checkpoint only supports PP, TP, and EP (as in expert_model_parallel not expert_tensor_parallel)?