modal.com Open in urlscan Pro
34.230.160.51  Public Scan

Submitted URL: https://modal.support/
Effective URL: https://modal.com/
Submission: On October 25 via api from US — Scanned from US

Form analysis 1 forms found in the DOM

<form class="flex w-full bg-light-green/60 rounded-full h-10 backdrop-blur-sm"><input type="text" placeholder="Type an image prompt (e.g. blue tulips)"
    class="flex-grow w-full min-w-0 pl-6 pr-2 h-full text-sm text-black !bg-transparent border-none focus:ring-0 placeholder:text-black/70">
  <div class="p-0.5"><button type="submit" class="px-6 h-full rounded-full text-sm text-light-green bg-black hover:bg-gray-800 focus:outline-none transition-colors">Generate</button></div>
</form>

Text Content

Use Cases
Pricing Customers Blog Docs Company
Log In Sign Up




RUN ANYTHING
IN THE CLOUD

Serverless cloud infrastructure
for AI, ML, and data applications
Get Started Book a Demo




CLOUD DEVELOPMENT MADE FRICTIONLESS



Run generative AI models, large-scale batch jobs, job queues, and much more.
Bring your own code — we run the infrastructure.

View Docs



ITERATE AT THE SPEED OF THOUGHT



Make code changes and watch your app rebuild instantly. Never write a single
line of YAML again.

View Docs


BUILT FOR LARGE-SCALE WORKLOADS



Engineered in Rust, our custom container stack allows you to scale to hundreds
of GPUs and then back down to zero in seconds. Pay only while it's running.

View Docs


USE CASES


GENERATIVE AI INFERENCE THAT SCALES WITH YOU

View Examples

--------------------------------------------------------------------------------

Fast cold boots

Load gigabytes of weights in seconds with our optimized container file system.

--------------------------------------------------------------------------------

Bring your own code

Deploy anything from custom models to popular frameworks.

--------------------------------------------------------------------------------

Seamless autoscaling

Handle bursty and unpredictable load by scaling to thousands of GPUs and back
down to zero.


Generate

--------------------------------------------------------------------------------

Fast cold boots

--------------------------------------------------------------------------------

Bring your own code

--------------------------------------------------------------------------------

Seamless autoscaling

View Examples


FINE-TUNING AND TRAINING WITHOUT MANAGING INFRASTRUCTURE

View Examples

--------------------------------------------------------------------------------

Start training immediately

Serverless, but for high-performance compute. Run things on massive amounts of
CPU and memory.

--------------------------------------------------------------------------------

Never wait in line

Run as many experiments as you need to, in parallel. Stop paying for idle GPUs
when you’re done.

--------------------------------------------------------------------------------

Cloud storage

Mount weights and data in distributed volumes, then access them wherever they’re
needed.




--------------------------------------------------------------------------------

Start training immediately

--------------------------------------------------------------------------------

Never wait in line

--------------------------------------------------------------------------------

Cloud storage

View Examples


BATCH PROCESSING OPTIMIZED FOR HIGH-VOLUME WORKLOADS

View Examples

--------------------------------------------------------------------------------

Supercomputing scale

Serverless, but for high-performance compute. Run things on massive amounts of
CPU and memory.

--------------------------------------------------------------------------------

Serverless pricing

Pay only for resources consumed, by the second, as you spin up containers.

--------------------------------------------------------------------------------

Powerful compute primitives

Simple fan-out parallelism that scales to thousands of containers, with a single
line of Python.




--------------------------------------------------------------------------------

Supercomputing scale

--------------------------------------------------------------------------------

Serverless pricing

--------------------------------------------------------------------------------

Powerful compute primitives

View Examples
Build anything with Modal

Language Models

Image, Video, 3D

Audio Processing

Fine-Tuning

Batch Processing

Sandboxed Code

Language Models

Image, Video, 3D

Audio Processing

Fine-Tuning

Batch Processing

Sandboxed Code





FEATURES

Flexible Environments

Bring your own image or build one in Python, scale resources as needed, and
leverage state-of-the-art GPUs like H100s & A100s for high-performance
computing.

Seamless Integrations

Export function logs to Datadog or any OpenTelemetry-compatible provider, and
easily mount cloud storage from major providers (S3, R2 etc.).

Data Storage

Manage data effortlessly with storage solutions (network volumes, key-value
stores and queues). Provision storage types and interact with them using
familiar Python syntax.

Job Scheduling

Take control of your workloads with powerful scheduling. Set up cron jobs,
retries, and timeouts, or use batching to optimize resource usage.

Web Endpoints

Deploy and manage web services with ease. Create custom domains, set up
streaming and websockets, and serve functions as secure HTTPS endpoints.

Built-In Debugging

Troubleshoot efficiently with built-in debugging tools. Use the modal shell for
interactive debugging and set breakpoints to pinpoint issues quickly.

--------------------------------------------------------------------------------

Flexible Environments

--------------------------------------------------------------------------------

Seamless Integrations

--------------------------------------------------------------------------------

Data Storage

--------------------------------------------------------------------------------

Job Scheduling

--------------------------------------------------------------------------------

Web Endpoints

--------------------------------------------------------------------------------

Built-In Debugging

Only pay when your
code is running
Scale up to hundreds of nodes and down to zero within seconds. Pay for actual
compute, by the CPU cycle. With $30 of compute on us, every month.


COMPUTE COSTS

Per hour
Per second

--------------------------------------------------------------------------------

GPU Tasks

Nvidia H100

$0.001267 / sec

Nvidia A100, 80 GB

$0.000944 / sec

Nvidia A100, 40 GB

$0.000772 / sec

Nvidia A10G

$0.000306 / sec

Nvidia L4

$0.000222 / sec

Nvidia T4

$0.000164 / sec

--------------------------------------------------------------------------------

CPU

Physical core
(2 vCPU equivalent)

$0.000038 / core / sec

*minimum of 0.125 cores per container

--------------------------------------------------------------------------------

Memory

$0.00000667 / GiB / sec

For teams
of all scales
Starter
For small teams and independent developers looking to level up.
Team
For startups and larger organizations looking to scale quickly.
Enterprise
For organizations prioritizing security, support, and reliability.
View Pricing



SECURITY AND GOVERNANCE

--------------------------------------------------------------------------------

Built on top of gVisor

--------------------------------------------------------------------------------

SOC 2 and HIPAA

--------------------------------------------------------------------------------

Region support

--------------------------------------------------------------------------------

SSO sign in for enterprise


Learn More



--------------------------------------------------------------------------------

Built on top of gVisor

--------------------------------------------------------------------------------

SOC 2 and HIPAA

--------------------------------------------------------------------------------

Region support

--------------------------------------------------------------------------------

SSO sign in for enterprise


Learn More


BUILT WITH MODAL

View all



DEPLOY AN OPENAI-COMPATIBLE LLM SERVICE

Run large language models with a drop-in replacement for the OpenAI API.


CUSTOM PET ART FROM FLUX WITH HUGGING FACE AND GRADIO

Fine-tune an image generation model on pictures of your pet.


VOICE CHAT WITH LLMS

Build an interactive voice chat app.


SERVE DIFFUSION MODELS

Serve Flux on Modal with a number of optimizations for blazingly fast inference.


FAST PODCAST TRANSCRIPTIONS

Build an end-to-end podcast transcription app that leverages dozens of
containers for super-fast processing.


DEPLOY A HACKER NEWS SLACKBOT

Periodically post new Hacker News posts to Slack.


RETRIEVAL-AUGMENTED GENERATION (RAG) FOR Q&A

Build a question-answering web endpoint that can cite its sources.


DOCUMENT OCR JOB QUEUE

Use Modal as an infinitely scalable job queue that can service async tasks from
a web app.


PARALLEL PROCESSING OF PARQUET FILES ON S3

Analyze data from the Taxi and Limousine Commission of NYC in parallel.



> “Modal makes it easy to write code that runs on 100s of GPUs in parallel,
> transcribing podcasts in a fraction of the time.”

Mike Cohen, Head of Data

> “Tasks that would have taken days to complete take minutes instead. We’ve
> saved thousands of dollars deploying LLMs on Modal.”

Rahul Sengottuvelu, Head of Applied AI

> “The beauty of Modal is that all you need to know is that you can scale your
> function calls in the cloud with a few lines of Python.”

Georg Kucsko, Co-founder and CTO


Case Study
Join Modal's developer
community
Modal Community Slack
Igor Kotua
Engineer, The Linux Foundation

If you building AI stuff with Python and haven't tried @modal_labs you are
missing out big time

Daniel Rothenberg
Co-founder, Brightband

@modal_labs continues to be magical... 10 minutes of effort and the
`joblib`-based parallelism I use to test on my local machine can trivially scale
out on the cloud. Makes life so easy!

Erin Boyle
ML Engineer, Tesla

This tool is awesome. So empowering to have your infra needs met with just a
couple decorators. Good people, too!

Lindsay Pettingill
Data @ Replit

Modal has the most magical onboarding I've ever seen and it's not even close.
And Erik's walk through of how they approached it is a Masterclass.

Izzy Miller
DevRel, Hex

special shout out to @modal_labs and @_hex_tech for providing the crucial
infrastructure to run this! Modal is the coolest tool I’ve tried in a really
long time— cannnot say enough good things.

Mark Tenenholtz
Head of AI, PredeloHQ

I use @modal_labs because it brings me joy. There isn't much more to it.

Nick Schrock
Founder, Dagster Labs

I have tried @modal_labs and am now officially Modal-pilled. Great work
@bernhardsson and team. Every hyperscalar should be trying this out and
immediately pivoting their compute teams' roadmaps to match this DX.

Moin Nadeem
Co-founder, Phonic

I've realized @modal_labs is actually a great fit for ML training pipelines. If
you're running model-based evals, why not just call a serverless Modal function
and have it evaluate your model on a separate worker GPU? This makes evaluation
during training really easy.

Igor Kotua
Engineer, The Linux Foundation

If you building AI stuff with Python and haven't tried @modal_labs you are
missing out big time

Daniel Rothenberg
Co-founder, Brightband

@modal_labs continues to be magical... 10 minutes of effort and the
`joblib`-based parallelism I use to test on my local machine can trivially scale
out on the cloud. Makes life so easy!

Erin Boyle
ML Engineer, Tesla

This tool is awesome. So empowering to have your infra needs met with just a
couple decorators. Good people, too!

Lindsay Pettingill
Data @ Replit

Modal has the most magical onboarding I've ever seen and it's not even close.
And Erik's walk through of how they approached it is a Masterclass.

Izzy Miller
DevRel, Hex

special shout out to @modal_labs and @_hex_tech for providing the crucial
infrastructure to run this! Modal is the coolest tool I’ve tried in a really
long time— cannnot say enough good things.

Mark Tenenholtz
Head of AI, PredeloHQ

I use @modal_labs because it brings me joy. There isn't much more to it.

Nick Schrock
Founder, Dagster Labs

I have tried @modal_labs and am now officially Modal-pilled. Great work
@bernhardsson and team. Every hyperscalar should be trying this out and
immediately pivoting their compute teams' roadmaps to match this DX.

Moin Nadeem
Co-founder, Phonic

I've realized @modal_labs is actually a great fit for ML training pipelines. If
you're running model-based evals, why not just call a serverless Modal function
and have it evaluate your model on a separate worker GPU? This makes evaluation
during training really easy.

Caleb
Product ML Engineer, Hugging Face

Bullish on @modal_labs - Great Docs + Examples - Healthy Free Plan (30$ free
compute / month) - Never have to worry about infra / just Python

Matt
AI Engineer, StackOne

@modal_labs has got a bunch of stuff just worked out this should be how you
deploy python apps. wow

Aman Kishore
Research Engineer, Harvey

If you are still using AWS Lambda instead of @modal_labs you're not moving fast
enough

Jai Chopra
Product, LanceDB

Recently built an app on Lambda and just started to use @modal_labs, the
difference is insane! Modal is amazing, virtually no cold start time, onboarding
experience is great 🚀

Diego Fernandes
Co-founder & CTO, RocketSeat

Probably one of the best piece of software I'm using this year: modal.com

Adam Azzam
Product, Prefect

feels weird at this point to use anything else than @modal_labs for this —
absolutely the GOAT of dynamic sandboxes

Rémi 📎
Co-founder & CEO, .txt

Nothing beats @modal_labs when it comes to deploying a quick POC

Matt Holden
Founder

Late to the party, but finally playing with @modal_labs to run some backend
jobs. DX is sooo nice (compared to Docker, Cloud Run, Lambda, etc). Just
decorate a Python function and deploy. And it's fast! Love it.

Caleb
Product ML Engineer, Hugging Face

Bullish on @modal_labs - Great Docs + Examples - Healthy Free Plan (30$ free
compute / month) - Never have to worry about infra / just Python

Matt
AI Engineer, StackOne

@modal_labs has got a bunch of stuff just worked out this should be how you
deploy python apps. wow

Aman Kishore
Research Engineer, Harvey

If you are still using AWS Lambda instead of @modal_labs you're not moving fast
enough

Jai Chopra
Product, LanceDB

Recently built an app on Lambda and just started to use @modal_labs, the
difference is insane! Modal is amazing, virtually no cold start time, onboarding
experience is great 🚀

Diego Fernandes
Co-founder & CTO, RocketSeat

Probably one of the best piece of software I'm using this year: modal.com

Adam Azzam
Product, Prefect

feels weird at this point to use anything else than @modal_labs for this —
absolutely the GOAT of dynamic sandboxes

Rémi 📎
Co-founder & CEO, .txt

Nothing beats @modal_labs when it comes to deploying a quick POC

Matt Holden
Founder

Late to the party, but finally playing with @modal_labs to run some backend
jobs. DX is sooo nice (compared to Docker, Cloud Run, Lambda, etc). Just
decorate a Python function and deploy. And it's fast! Love it.





SHIP YOUR FIRST APP IN MINUTES.

Get Started

$30 / month free compute

© Modal 2024

Use Cases

Language Model Inference

Image, Video & 3D

Audio Processing

Fine-Tuning

Job Queues & Batch Processing

Sandboxing Code

Resources

Documentation

Pricing

Slack Community

Security & Privacy

Articles

Popular Examples

Serve LLM APIs with vLLM

Create Custom Art of Your Pet

Analyze Parquet files from S3 with DuckDB

Run hundreds of LoRAs from one app

Replace your CEO with an LLM

Company

About

Blog

Careers

Privacy Policy

Terms

© Modal 2024