“Modal lets us deploy new ML models in hours rather than weeks. We use it across spam detection, recommendations, audio transcription, and video pipelines, and it’s helped us move faster with far less complexity.”
Define your training function with Modal’s SDK. Easily keep ML dependencies and GPU requirements in sync with application code.
image = (
modal.Image.from_registry(
f"nvidia/cuda:{tag}"
)
.uv_pip_install(
"accelerate",
"torch",
)
)
@app.function(gpu="B200:8", image=image)
@modal.clustered(size=4, rdma=True)
def train_multi_node():
...
Ingest training data from anywhere: Modal’s distributed Volumes, cloud buckets, or your local filesystem.
volume = modal.Volume.from_name(
"training_data_vol"
)
@app.function(
volumes={
"/my-s3-mount": modal.CloudBucketMount(
"training_data_s3",
secret=secret,
),
"/my-volume": volume,
}
)
def train():
...
Modal’s container stack launches GPUs for your function in < 1s. Fan out experiments to accelerate your research.
Scale from 1 GPU to 64 with just one line of code
Spin up a cluster in a second with no minimum commitments
B200, H200, and H100 clusters equipped with Infiniband and private networking