modal.com
Open in
urlscan Pro
34.230.160.51
Public Scan
Submitted URL: https://modal.support/
Effective URL: https://modal.com/
Submission: On October 25 via api from US — Scanned from US
Effective URL: https://modal.com/
Submission: On October 25 via api from US — Scanned from US
Form analysis
1 forms found in the DOM<form class="flex w-full bg-light-green/60 rounded-full h-10 backdrop-blur-sm"><input type="text" placeholder="Type an image prompt (e.g. blue tulips)"
class="flex-grow w-full min-w-0 pl-6 pr-2 h-full text-sm text-black !bg-transparent border-none focus:ring-0 placeholder:text-black/70">
<div class="p-0.5"><button type="submit" class="px-6 h-full rounded-full text-sm text-light-green bg-black hover:bg-gray-800 focus:outline-none transition-colors">Generate</button></div>
</form>
Text Content
Use Cases Pricing Customers Blog Docs Company Log In Sign Up RUN ANYTHING IN THE CLOUD Serverless cloud infrastructure for AI, ML, and data applications Get Started Book a Demo CLOUD DEVELOPMENT MADE FRICTIONLESS Run generative AI models, large-scale batch jobs, job queues, and much more. Bring your own code — we run the infrastructure. View Docs ITERATE AT THE SPEED OF THOUGHT Make code changes and watch your app rebuild instantly. Never write a single line of YAML again. View Docs BUILT FOR LARGE-SCALE WORKLOADS Engineered in Rust, our custom container stack allows you to scale to hundreds of GPUs and then back down to zero in seconds. Pay only while it's running. View Docs USE CASES GENERATIVE AI INFERENCE THAT SCALES WITH YOU View Examples -------------------------------------------------------------------------------- Fast cold boots Load gigabytes of weights in seconds with our optimized container file system. -------------------------------------------------------------------------------- Bring your own code Deploy anything from custom models to popular frameworks. -------------------------------------------------------------------------------- Seamless autoscaling Handle bursty and unpredictable load by scaling to thousands of GPUs and back down to zero. Generate -------------------------------------------------------------------------------- Fast cold boots -------------------------------------------------------------------------------- Bring your own code -------------------------------------------------------------------------------- Seamless autoscaling View Examples FINE-TUNING AND TRAINING WITHOUT MANAGING INFRASTRUCTURE View Examples -------------------------------------------------------------------------------- Start training immediately Serverless, but for high-performance compute. Run things on massive amounts of CPU and memory. -------------------------------------------------------------------------------- Never wait in line Run as many experiments as you need to, in parallel. Stop paying for idle GPUs when you’re done. -------------------------------------------------------------------------------- Cloud storage Mount weights and data in distributed volumes, then access them wherever they’re needed. -------------------------------------------------------------------------------- Start training immediately -------------------------------------------------------------------------------- Never wait in line -------------------------------------------------------------------------------- Cloud storage View Examples BATCH PROCESSING OPTIMIZED FOR HIGH-VOLUME WORKLOADS View Examples -------------------------------------------------------------------------------- Supercomputing scale Serverless, but for high-performance compute. Run things on massive amounts of CPU and memory. -------------------------------------------------------------------------------- Serverless pricing Pay only for resources consumed, by the second, as you spin up containers. -------------------------------------------------------------------------------- Powerful compute primitives Simple fan-out parallelism that scales to thousands of containers, with a single line of Python. -------------------------------------------------------------------------------- Supercomputing scale -------------------------------------------------------------------------------- Serverless pricing -------------------------------------------------------------------------------- Powerful compute primitives View Examples Build anything with Modal Language Models Image, Video, 3D Audio Processing Fine-Tuning Batch Processing Sandboxed Code Language Models Image, Video, 3D Audio Processing Fine-Tuning Batch Processing Sandboxed Code FEATURES Flexible Environments Bring your own image or build one in Python, scale resources as needed, and leverage state-of-the-art GPUs like H100s & A100s for high-performance computing. Seamless Integrations Export function logs to Datadog or any OpenTelemetry-compatible provider, and easily mount cloud storage from major providers (S3, R2 etc.). Data Storage Manage data effortlessly with storage solutions (network volumes, key-value stores and queues). Provision storage types and interact with them using familiar Python syntax. Job Scheduling Take control of your workloads with powerful scheduling. Set up cron jobs, retries, and timeouts, or use batching to optimize resource usage. Web Endpoints Deploy and manage web services with ease. Create custom domains, set up streaming and websockets, and serve functions as secure HTTPS endpoints. Built-In Debugging Troubleshoot efficiently with built-in debugging tools. Use the modal shell for interactive debugging and set breakpoints to pinpoint issues quickly. -------------------------------------------------------------------------------- Flexible Environments -------------------------------------------------------------------------------- Seamless Integrations -------------------------------------------------------------------------------- Data Storage -------------------------------------------------------------------------------- Job Scheduling -------------------------------------------------------------------------------- Web Endpoints -------------------------------------------------------------------------------- Built-In Debugging Only pay when your code is running Scale up to hundreds of nodes and down to zero within seconds. Pay for actual compute, by the CPU cycle. With $30 of compute on us, every month. COMPUTE COSTS Per hour Per second -------------------------------------------------------------------------------- GPU Tasks Nvidia H100 $0.001267 / sec Nvidia A100, 80 GB $0.000944 / sec Nvidia A100, 40 GB $0.000772 / sec Nvidia A10G $0.000306 / sec Nvidia L4 $0.000222 / sec Nvidia T4 $0.000164 / sec -------------------------------------------------------------------------------- CPU Physical core (2 vCPU equivalent) $0.000038 / core / sec *minimum of 0.125 cores per container -------------------------------------------------------------------------------- Memory $0.00000667 / GiB / sec For teams of all scales Starter For small teams and independent developers looking to level up. Team For startups and larger organizations looking to scale quickly. Enterprise For organizations prioritizing security, support, and reliability. View Pricing SECURITY AND GOVERNANCE -------------------------------------------------------------------------------- Built on top of gVisor -------------------------------------------------------------------------------- SOC 2 and HIPAA -------------------------------------------------------------------------------- Region support -------------------------------------------------------------------------------- SSO sign in for enterprise Learn More -------------------------------------------------------------------------------- Built on top of gVisor -------------------------------------------------------------------------------- SOC 2 and HIPAA -------------------------------------------------------------------------------- Region support -------------------------------------------------------------------------------- SSO sign in for enterprise Learn More BUILT WITH MODAL View all DEPLOY AN OPENAI-COMPATIBLE LLM SERVICE Run large language models with a drop-in replacement for the OpenAI API. CUSTOM PET ART FROM FLUX WITH HUGGING FACE AND GRADIO Fine-tune an image generation model on pictures of your pet. VOICE CHAT WITH LLMS Build an interactive voice chat app. SERVE DIFFUSION MODELS Serve Flux on Modal with a number of optimizations for blazingly fast inference. FAST PODCAST TRANSCRIPTIONS Build an end-to-end podcast transcription app that leverages dozens of containers for super-fast processing. DEPLOY A HACKER NEWS SLACKBOT Periodically post new Hacker News posts to Slack. RETRIEVAL-AUGMENTED GENERATION (RAG) FOR Q&A Build a question-answering web endpoint that can cite its sources. DOCUMENT OCR JOB QUEUE Use Modal as an infinitely scalable job queue that can service async tasks from a web app. PARALLEL PROCESSING OF PARQUET FILES ON S3 Analyze data from the Taxi and Limousine Commission of NYC in parallel. > “Modal makes it easy to write code that runs on 100s of GPUs in parallel, > transcribing podcasts in a fraction of the time.” Mike Cohen, Head of Data > “Tasks that would have taken days to complete take minutes instead. We’ve > saved thousands of dollars deploying LLMs on Modal.” Rahul Sengottuvelu, Head of Applied AI > “The beauty of Modal is that all you need to know is that you can scale your > function calls in the cloud with a few lines of Python.” Georg Kucsko, Co-founder and CTO Case Study Join Modal's developer community Modal Community Slack Igor Kotua Engineer, The Linux Foundation If you building AI stuff with Python and haven't tried @modal_labs you are missing out big time Daniel Rothenberg Co-founder, Brightband @modal_labs continues to be magical... 10 minutes of effort and the `joblib`-based parallelism I use to test on my local machine can trivially scale out on the cloud. Makes life so easy! Erin Boyle ML Engineer, Tesla This tool is awesome. So empowering to have your infra needs met with just a couple decorators. Good people, too! Lindsay Pettingill Data @ Replit Modal has the most magical onboarding I've ever seen and it's not even close. And Erik's walk through of how they approached it is a Masterclass. Izzy Miller DevRel, Hex special shout out to @modal_labs and @_hex_tech for providing the crucial infrastructure to run this! Modal is the coolest tool I’ve tried in a really long time— cannnot say enough good things. Mark Tenenholtz Head of AI, PredeloHQ I use @modal_labs because it brings me joy. There isn't much more to it. Nick Schrock Founder, Dagster Labs I have tried @modal_labs and am now officially Modal-pilled. Great work @bernhardsson and team. Every hyperscalar should be trying this out and immediately pivoting their compute teams' roadmaps to match this DX. Moin Nadeem Co-founder, Phonic I've realized @modal_labs is actually a great fit for ML training pipelines. If you're running model-based evals, why not just call a serverless Modal function and have it evaluate your model on a separate worker GPU? This makes evaluation during training really easy. Igor Kotua Engineer, The Linux Foundation If you building AI stuff with Python and haven't tried @modal_labs you are missing out big time Daniel Rothenberg Co-founder, Brightband @modal_labs continues to be magical... 10 minutes of effort and the `joblib`-based parallelism I use to test on my local machine can trivially scale out on the cloud. Makes life so easy! Erin Boyle ML Engineer, Tesla This tool is awesome. So empowering to have your infra needs met with just a couple decorators. Good people, too! Lindsay Pettingill Data @ Replit Modal has the most magical onboarding I've ever seen and it's not even close. And Erik's walk through of how they approached it is a Masterclass. Izzy Miller DevRel, Hex special shout out to @modal_labs and @_hex_tech for providing the crucial infrastructure to run this! Modal is the coolest tool I’ve tried in a really long time— cannnot say enough good things. Mark Tenenholtz Head of AI, PredeloHQ I use @modal_labs because it brings me joy. There isn't much more to it. Nick Schrock Founder, Dagster Labs I have tried @modal_labs and am now officially Modal-pilled. Great work @bernhardsson and team. Every hyperscalar should be trying this out and immediately pivoting their compute teams' roadmaps to match this DX. Moin Nadeem Co-founder, Phonic I've realized @modal_labs is actually a great fit for ML training pipelines. If you're running model-based evals, why not just call a serverless Modal function and have it evaluate your model on a separate worker GPU? This makes evaluation during training really easy. Caleb Product ML Engineer, Hugging Face Bullish on @modal_labs - Great Docs + Examples - Healthy Free Plan (30$ free compute / month) - Never have to worry about infra / just Python Matt AI Engineer, StackOne @modal_labs has got a bunch of stuff just worked out this should be how you deploy python apps. wow Aman Kishore Research Engineer, Harvey If you are still using AWS Lambda instead of @modal_labs you're not moving fast enough Jai Chopra Product, LanceDB Recently built an app on Lambda and just started to use @modal_labs, the difference is insane! Modal is amazing, virtually no cold start time, onboarding experience is great 🚀 Diego Fernandes Co-founder & CTO, RocketSeat Probably one of the best piece of software I'm using this year: modal.com Adam Azzam Product, Prefect feels weird at this point to use anything else than @modal_labs for this — absolutely the GOAT of dynamic sandboxes Rémi 📎 Co-founder & CEO, .txt Nothing beats @modal_labs when it comes to deploying a quick POC Matt Holden Founder Late to the party, but finally playing with @modal_labs to run some backend jobs. DX is sooo nice (compared to Docker, Cloud Run, Lambda, etc). Just decorate a Python function and deploy. And it's fast! Love it. Caleb Product ML Engineer, Hugging Face Bullish on @modal_labs - Great Docs + Examples - Healthy Free Plan (30$ free compute / month) - Never have to worry about infra / just Python Matt AI Engineer, StackOne @modal_labs has got a bunch of stuff just worked out this should be how you deploy python apps. wow Aman Kishore Research Engineer, Harvey If you are still using AWS Lambda instead of @modal_labs you're not moving fast enough Jai Chopra Product, LanceDB Recently built an app on Lambda and just started to use @modal_labs, the difference is insane! Modal is amazing, virtually no cold start time, onboarding experience is great 🚀 Diego Fernandes Co-founder & CTO, RocketSeat Probably one of the best piece of software I'm using this year: modal.com Adam Azzam Product, Prefect feels weird at this point to use anything else than @modal_labs for this — absolutely the GOAT of dynamic sandboxes Rémi 📎 Co-founder & CEO, .txt Nothing beats @modal_labs when it comes to deploying a quick POC Matt Holden Founder Late to the party, but finally playing with @modal_labs to run some backend jobs. DX is sooo nice (compared to Docker, Cloud Run, Lambda, etc). Just decorate a Python function and deploy. And it's fast! Love it. SHIP YOUR FIRST APP IN MINUTES. Get Started $30 / month free compute © Modal 2024 Use Cases Language Model Inference Image, Video & 3D Audio Processing Fine-Tuning Job Queues & Batch Processing Sandboxing Code Resources Documentation Pricing Slack Community Security & Privacy Articles Popular Examples Serve LLM APIs with vLLM Create Custom Art of Your Pet Analyze Parquet files from S3 with DuckDB Run hundreds of LoRAs from one app Replace your CEO with an LLM Company About Blog Careers Privacy Policy Terms © Modal 2024