AI infrastructure that
developers love
Designed to help AI teams
deploy faster.
Programmable infra
Define everything in code, no YAML or config files. Keep environment and hardware requirements in sync.
Built for performance
Launch and scale containers in seconds to keep feedback loops tight and latency low.
Elastic GPU scaling
Elastic GPU capacity and access to thousands of GPUs across clouds. No quotas or reservations. Scale back to zero when not in use.
Unified observability
Integrated logging and full visibility into every function, container, and workload.
Powering any ML workload
Build on a powerful foundation
From filesystem to runtime, every layer of Modal’s platform is engineered to give you the tools to build robust, scalable data applications.
AI-native runtime
Built-in storage layer
First-party integrations
Multi-cloud capacity pool
Security and governance
“We use Modal to run edge inference with <10ms overhead and batch jobs at large scale. Our team loves the platform for the power and flexibility it gives us.”
Brian Ichter, Co-founder
“Modal makes it easy to write code that runs on 100s of GPUs in parallel, transcribing podcasts in a fraction of the time.”
Mike Cohen, Head of Data
“Everyone here loves Modal because it helps us move so much faster. We rely on it to handle massive spikes in volume for evals, RL environments, and MCP servers. Whenever a team asks about compute, we tell them to use Modal.”
Aakash Sabharwal, VP of Engineering
“We've previously managed to break services like GitHub because of our load, so Modal handling our massive scale so smoothly means a lot. We trust Modal to keep up with our growth, and we're excited to build together in the long term.”
Anton Osika, CEO & Founder
If you building AI stuff with Python and haven't tried @modal you are missing out big time
Bullish on @modal - Great Docs + Examples - Healthy Free Plan (30$ free compute / month) - Never have to worry about infra / just Python
@modal continues to be magical... 10 minutes of effort and the `joblib`-based parallelism I use to test on my local machine can trivially scale out on the cloud. Makes life so easy!
@modal has got a bunch of stuff just worked out this should be how you deploy python apps. wow
This tool is awesome. So empowering to have your infra needs met with just a couple decorators. Good people, too!
If you are still using AWS Lambda instead of @modal you're not moving fast enough
Recently built an app on Lambda and just started to use @modal, the difference is insane! Modal is amazing, virtually no cold start time, onboarding experience is great 🚀
special shout out to @modal and @_hex_tech for providing the crucial infrastructure to run this! Modal is the coolest tool I’ve tried in a really long time— cannnot say enough good things.
Probably one of the best piece of software I'm using this year: modal.com
I use @modal because it brings me joy. There isn't much more to it.
feels weird at this point to use anything else than @modal for this — absolutely the GOAT of dynamic sandboxes
Used @modal for the first time today - immediate "oh, this is how backends should work" moment, similar to using Vercel for the first time for frontend deployments.
Nothing beats @modal when it comes to deploying a quick POC
I've realized @modal is actually a great fit for ML training pipelines. If you're running model-based evals, why not just call a serverless Modal function and have it evaluate your model on a separate worker GPU? This makes evaluation during training really easy.
Late to the party, but finally playing with @modal to run some backend jobs. DX is sooo nice (compared to Docker, Cloud Run, Lambda, etc). Just decorate a Python function and deploy. And it's fast! Love it.
If you building AI stuff with Python and haven't tried @modal you are missing out big time
Bullish on @modal - Great Docs + Examples - Healthy Free Plan (30$ free compute / month) - Never have to worry about infra / just Python
@modal continues to be magical... 10 minutes of effort and the `joblib`-based parallelism I use to test on my local machine can trivially scale out on the cloud. Makes life so easy!
@modal has got a bunch of stuff just worked out this should be how you deploy python apps. wow
This tool is awesome. So empowering to have your infra needs met with just a couple decorators. Good people, too!
If you are still using AWS Lambda instead of @modal you're not moving fast enough
Recently built an app on Lambda and just started to use @modal, the difference is insane! Modal is amazing, virtually no cold start time, onboarding experience is great 🚀
special shout out to @modal and @_hex_tech for providing the crucial infrastructure to run this! Modal is the coolest tool I’ve tried in a really long time— cannnot say enough good things.
Probably one of the best piece of software I'm using this year: modal.com
I use @modal because it brings me joy. There isn't much more to it.
feels weird at this point to use anything else than @modal for this — absolutely the GOAT of dynamic sandboxes
Used @modal for the first time today - immediate "oh, this is how backends should work" moment, similar to using Vercel for the first time for frontend deployments.
Nothing beats @modal when it comes to deploying a quick POC
I've realized @modal is actually a great fit for ML training pipelines. If you're running model-based evals, why not just call a serverless Modal function and have it evaluate your model on a separate worker GPU? This makes evaluation during training really easy.
Late to the party, but finally playing with @modal to run some backend jobs. DX is sooo nice (compared to Docker, Cloud Run, Lambda, etc). Just decorate a Python function and deploy. And it's fast! Love it.