Spaces:
Running
Running
Commit
·
f93db33
1
Parent(s):
337ea13
fix: Use L40S (48GB) for 8-9B models on Modal
Browse filesL40S provides 48GB VRAM which is sufficient for 8-9B models with agentic workloads
utils/modal_job_submission.py
CHANGED
|
@@ -391,8 +391,8 @@ def _auto_select_modal_hardware(provider: str, model: str) -> Optional[str]:
|
|
| 391 |
# 13-15B models: ~52-75GB needed -> A100 40GB or A100 80GB
|
| 392 |
return "A100-80GB"
|
| 393 |
elif "8b" in model_lower or "9b" in model_lower:
|
| 394 |
-
# 8-9B models: ~32-45GB needed ->
|
| 395 |
-
return "
|
| 396 |
elif "7b" in model_lower:
|
| 397 |
# 7B models: ~28-35GB needed -> A10G can work with quantization
|
| 398 |
return "A10G"
|
|
|
|
| 391 |
# 13-15B models: ~52-75GB needed -> A100 40GB or A100 80GB
|
| 392 |
return "A100-80GB"
|
| 393 |
elif "8b" in model_lower or "9b" in model_lower:
|
| 394 |
+
# 8-9B models: ~32-45GB needed -> L40S (48GB VRAM)
|
| 395 |
+
return "L40S"
|
| 396 |
elif "7b" in model_lower:
|
| 397 |
# 7B models: ~28-35GB needed -> A10G can work with quantization
|
| 398 |
return "A10G"
|