petals.dev Open in urlscan Pro
2606:50c0:8000::153  Public Scan

URL: https://petals.dev/
Submission: On September 19 via api from CA — Scanned from CA

Form analysis 1 forms found in the DOM

<form class="d-flex col-xl-8 mx-auto">
  <input class="form-control me-2" type="email" name="email" id="email" placeholder="Leave your email" style="font-size: 11pt;">
  <button class="btn btn-outline-dark my-0" type="submit">Subscribe</button>
</form>

Text Content

PETALS

Run large language models at home, BitTorrent‑style

 * Generate text with Llama 3.1 (up to 405B), Mixtral (8x22B), Falcon (40B+) or
   BLOOM (176B) and fine‑tune them for your tasks — using a consumer-grade GPU
   or Google Colab.
 * You load a part of the model, then join a network of people serving its other
   parts. Single‑batch inference runs at up to 6 tokens/sec for Llama 2 (70B)
   and up to 4 tokens/sec for Falcon (180B) — enough for chatbots and
   interactive apps.
 * Beyond classic LLM APIs — you can employ any fine-tuning and sampling
   methods, execute custom paths through the model, or see its hidden states.
   You get the comforts of an API with the flexibility of PyTorch and
   🤗 Transformers.

THANKS FOR SUBSCRIBING!

We will email you only if we have really exciting updates.

Try now in Colab Docs on GitHub

Top contributors right now:

Loading... Oops, can't load the network status... • and more
Contribute my GPU

Follow development in Discord or via email:

Subscribe
We send updates once a few months. No spam.
Submitting...

We sent you an email to confirm your address. Click it and you're in!

Featured on:



This project is a part of the BigScience research workshop.