|
# BLOOM, a version for Petals |
|
|
|
This model is a version of [bigscience/bloom](https://huggingface.co/bigscience/bloom) |
|
post-processed to be run at home using the [Petals](https://github.com/bigscience-workshop/petals#readme) swarm. |
|
|
|
Please check out: |
|
|
|
- The [original model card](https://huggingface.co/bigscience/bloom) |
|
to learn about the model's capabilities, specifications, and terms of use. |
|
- The [Petals repository](https://github.com/bigscience-workshop/petals#readme) |
|
to learn how to install Petals and run this model over the Petals swarm. |
|
|
|
We provide minimal code examples below. |
|
|
|
## Using the model |
|
|
|
```python |
|
from petals import DistributedBloomForCausalLM |
|
|
|
model = DistributedBloomForCausalLM.from_pretrained("bigscience/bloom-petals") |
|
# Embeddings & prompts are on your device, BLOOM blocks are distributed across the Internet |
|
|
|
inputs = tokenizer("A cat sat", return_tensors="pt")["input_ids"] |
|
outputs = model.generate(inputs, max_new_tokens=5) |
|
print(tokenizer.decode(outputs[0])) # A cat sat on a mat... |
|
``` |
|
|
|
## Serving the model blocks |
|
|
|
```bash |
|
python -m petals.cli.run_server bigscience/bloom-petals |
|
``` |
|
|