fireworks.ai
Open in
urlscan Pro
76.76.21.21
Public Scan
URL:
https://fireworks.ai/
Submission: On March 28 via api from US — Scanned from DE
Submission: On March 28 via api from US — Scanned from DE
Form analysis
0 forms found in the DOMText Content
ModelsDocsPricing Company Login / Sign up ModelsDocsPricingBlogCareersContactLogin / Sign up THE PRODUCTION AI PLATFORM BUILT FOR DEVELOPERS Fireworks partners with the world's leading generative AI researchers to serve the best models, at the fastest speeds. Get Started for FreeContact Sales Companies of all sizes trust Fireworks to power their production AI use-cases MODELS CURATED AND OPTIMIZED BY FIREWORKS Chat LLM Mixtral MoE 8x7B Instruct Mistral MoE 8x7B Instruct v0.1 model with Sparse Mixture of Experts. Fine tuned for instruction following Try now Chat LLM FireFunction V1 Fireworks' open-source function calling model. Try now Chat LLM Llama 2 70B Chat A fine-tuned version of Llama 2 70B, optimized for dialogue applications using Reinforcement Learning from Human Feedback (RLHF), and perform comparably to ChatGPT according to human evaluations. Try now Chat LLM Mistral 7B Instruct The Mistral-7B-Instruct-v0.1 Large Language Model (LLM) is a instruct fine-tuned version of the Mistral-7B-v0.1 generative text model using a variety of publicly available conversation datasets. Try now See all 89 models THE FASTEST AND MOST UNCOMPROMISING AI PLATFORM! 0 Fireworks AI tokens / second 0 Next provider tokens / second 0 average provider tokens / second Industry Leading Performance Independently benchmarked to have the top speed of all inference providers 1 Sec Cutoff2 Sec Cutoff3 Sec Cutoff0 req/sec20 req/sec40 req/sec60 req/sec80 req/sec vLLM Quant vLLM FP16 Fireworks FP16 Fireworks FP8 Enterprise Scale Throughput Our proprietary stack blows open source options out of the water (see blog) FireLLaVA: the first commercially permissive OSS LLaVA model State-of-the-art Models Use powerful models curated by Fireworks or our in-house trained multi-modal and function-calling models 0 Billion+ tokens served in a day Battle Tested for Reliability Fireworks is the 2nd most used open-source model provider and also generates over 1M images/day TSPython fetch("https://api.fireworks.ai/inference/v1/chat/completions", { method: "POST", headers: { "Content-Type": "application "Authorization: "Bearer <API KEY>", }, body: JSON.stringify({ model: "accounts/fireworks/mixtral-8x7b", prompt: "Say this is a test", max_tokens: 700, }), }) Built for Developers Our OpenAI-compatible API makes it easy to start building with Fireworks! > "We've had a really great experience working with Fireworks to host open > source models, including SDXL, Llama, and Mistral. After migrating one of our > models, we noticed a 3x speedup in response time, which made our app feel much > more responsive and boosted our engagement metrics." Spencer Chan Product Lead, Poe by Quora > "Fireworks is a go-to platform for deploying OSS models. Not only do they have > superb tech - fast inference and low costs - but their team also offers > incredible support on all things related to inference, finetuning, and OSS > LLMs in general. LangChain users find great value for the ease of setting it > up, the speed at which they add new models, and the ability to easily use the > latest and greatest OSS models without breaking the bank" Harrison Chase CEO & Co-Founder, LangChain > "Fireworks has been a fantastic partner as we build AI dev tools at > Sourcegraph. Model inference runs super fast and reliably, which allows us to > focus on core differentiators like finetuning, AI-powered code search, and > deep code context that have made Cody the best AI coding assistant for > completions and codebase-aware chat. They are also super responsive with > feedback and ship at an amazing pace." Beyang Liu CTO, SourceGraph > "Fireworks enabled us to deliver industry-leading speed for both LLM and image > generation inference. It makes a huge difference for the customer experience > when the AI responses are fast and interactive." Ves Stoyanov Head of AI, Tome > "Fireworks has been marvelous, both in terms of features and support. > Fireworks allows Artera's AI engineers to operate easily running both > interactive and large scale batch jobs with a few clicks." Nathan Silberman Head of AI, Artera > "We've had a really great experience working with Fireworks to host open > source models, including SDXL, Llama, and Mistral. After migrating one of our > models, we noticed a 3x speedup in response time, which made our app feel much > more responsive and boosted our engagement metrics." Spencer Chan Product Lead, Poe by Quora > "Fireworks is a go-to platform for deploying OSS models. Not only do they have > superb tech - fast inference and low costs - but their team also offers > incredible support on all things related to inference, finetuning, and OSS > LLMs in general. LangChain users find great value for the ease of setting it > up, the speed at which they add new models, and the ability to easily use the > latest and greatest OSS models without breaking the bank" Harrison Chase CEO & Co-Founder, LangChain > "Fireworks has been a fantastic partner as we build AI dev tools at > Sourcegraph. Model inference runs super fast and reliably, which allows us to > focus on core differentiators like finetuning, AI-powered code search, and > deep code context that have made Cody the best AI coding assistant for > completions and codebase-aware chat. They are also super responsive with > feedback and ship at an amazing pace." Beyang Liu CTO, SourceGraph > "Fireworks enabled us to deliver industry-leading speed for both LLM and image > generation inference. It makes a huge difference for the customer experience > when the AI responses are fast and interactive." Ves Stoyanov Head of AI, Tome > "Fireworks has been marvelous, both in terms of features and support. > Fireworks allows Artera's AI engineers to operate easily running both > interactive and large scale batch jobs with a few clicks." Nathan Silberman Head of AI, Artera LEVEL UP WITH FIREWORKS AI ENTERPRISE Get dedicated deployments for your models to ensure uptime and speed Fireworks is proudly compliant with HIPAA and SOC2 and offers secure VPC and VPN connectivity Meet your needs with data privacy - own your data and your models © 2024 Fireworks AI All rights reserved. Toggle theme Dark Modem Pages HomePricingModelsDocs Company BlogCareers Legal Terms of ServicePrivacy Policy 0 req/sec