From 50fb8205deb599656f9f518fbda6d24fa548ccb1 Mon Sep 17 00:00:00 2001 From: Vadim Peretokin Date: Thu, 29 Dec 2022 14:46:23 +0100 Subject: [PATCH] Correct grammar in readme (#166) --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index cfaa27e..15bb1f5 100644 --- a/README.md +++ b/README.md @@ -52,7 +52,7 @@ Check out more tutorials: - Launching your own swarm: [tutorial](https://github.com/bigscience-workshop/petals/wiki/Launch-your-own-swarm) - Running a custom foundation model: [tutorial](https://github.com/bigscience-workshop/petals/wiki/Run-a-custom-model-with-Petals) -## How it works? +## How does it work? - Petals runs large language models like BLOOM-176B **collaboratively** — you load a small part of the model, then team up with people serving the other parts to run inference or fine-tuning. - Inference runs at ≈ 1 sec per step (token) — 10x faster than possible with offloading, enough for chatbots and other interactive apps. Parallel inference reaches hundreds of tokens/sec.