davanstrien HF Staff commited on
Commit
61f5a7b
·
1 Parent(s): b1ee428

Fix device_map for 120B model - use tp_plan instead

Browse files
Files changed (1) hide show
  1. gpt_oss_transformers.py +1 -1
gpt_oss_transformers.py CHANGED
@@ -242,7 +242,7 @@ def main(
242
  # For 120B model, use tensor parallel planning
243
  if "120b" in model_id:
244
  model_kwargs = {
245
- "device_map": {"tp_plan": "auto"},
246
  "enable_expert_parallel": True,
247
  }
248
  else:
 
242
  # For 120B model, use tensor parallel planning
243
  if "120b" in model_id:
244
  model_kwargs = {
245
+ "tp_plan": "auto",
246
  "enable_expert_parallel": True,
247
  }
248
  else: