模型:
bigscience/bloom-7b1-petals
This model is a version of bigscience/bloom-7b1 post-processed to be run at home using the Petals swarm.
Note: Petals is developed to run 100B+ models like the full-scale BLOOM or BLOOMZ . This model is provided for testing purposes only. It may be more efficient to run the original version of it locally.
Please check out:
We provide minimal code examples below.
from petals import DistributedBloomForCausalLM model = DistributedBloomForCausalLM.from_pretrained("bigscience/bloom-7b1-petals") # Embeddings & prompts are on your device, BLOOM blocks are distributed across the Internet inputs = tokenizer("A cat sat", return_tensors="pt")["input_ids"] outputs = model.generate(inputs, max_new_tokens=5) print(tokenizer.decode(outputs[0])) # A cat sat on a mat...
python -m petals.cli.run_server bigscience/bloom-7b1-petals