petals.dev
Open in
urlscan Pro
2606:50c0:8000::153
Public Scan
URL:
https://petals.dev/
Submission: On September 19 via api from CA — Scanned from CA
Submission: On September 19 via api from CA — Scanned from CA
Form analysis
1 forms found in the DOM<form class="d-flex col-xl-8 mx-auto">
<input class="form-control me-2" type="email" name="email" id="email" placeholder="Leave your email" style="font-size: 11pt;">
<button class="btn btn-outline-dark my-0" type="submit">Subscribe</button>
</form>
Text Content
PETALS Run large language models at home, BitTorrent‑style * Generate text with Llama 3.1 (up to 405B), Mixtral (8x22B), Falcon (40B+) or BLOOM (176B) and fine‑tune them for your tasks — using a consumer-grade GPU or Google Colab. * You load a part of the model, then join a network of people serving its other parts. Single‑batch inference runs at up to 6 tokens/sec for Llama 2 (70B) and up to 4 tokens/sec for Falcon (180B) — enough for chatbots and interactive apps. * Beyond classic LLM APIs — you can employ any fine-tuning and sampling methods, execute custom paths through the model, or see its hidden states. You get the comforts of an API with the flexibility of PyTorch and 🤗 Transformers. THANKS FOR SUBSCRIBING! We will email you only if we have really exciting updates. Try now in Colab Docs on GitHub Top contributors right now: Loading... Oops, can't load the network status... • and more Contribute my GPU Follow development in Discord or via email: Subscribe We send updates once a few months. No spam. Submitting... We sent you an email to confirm your address. Click it and you're in! Featured on: This project is a part of the BigScience research workshop.