kshitijthakkar commited on
Commit
f93db33
·
1 Parent(s): 337ea13

fix: Use L40S (48GB) for 8-9B models on Modal

Browse files

L40S provides 48GB VRAM which is sufficient for 8-9B models with agentic workloads

Files changed (1) hide show
  1. utils/modal_job_submission.py +2 -2
utils/modal_job_submission.py CHANGED
@@ -391,8 +391,8 @@ def _auto_select_modal_hardware(provider: str, model: str) -> Optional[str]:
391
  # 13-15B models: ~52-75GB needed -> A100 40GB or A100 80GB
392
  return "A100-80GB"
393
  elif "8b" in model_lower or "9b" in model_lower:
394
- # 8-9B models: ~32-45GB needed -> A10G 24GB may OOM, use A100
395
- return "A100-40GB"
396
  elif "7b" in model_lower:
397
  # 7B models: ~28-35GB needed -> A10G can work with quantization
398
  return "A10G"
 
391
  # 13-15B models: ~52-75GB needed -> A100 40GB or A100 80GB
392
  return "A100-80GB"
393
  elif "8b" in model_lower or "9b" in model_lower:
394
+ # 8-9B models: ~32-45GB needed -> L40S (48GB VRAM)
395
+ return "L40S"
396
  elif "7b" in model_lower:
397
  # 7B models: ~28-35GB needed -> A10G can work with quantization
398
  return "A10G"