diff --git a/README.md b/README.md index 3955835..35d5cba 100644 --- a/README.md +++ b/README.md @@ -12,7 +12,7 @@ from transformers import AutoTokenizer from petals import AutoDistributedModelForCausalLM model_name = "enoch/llama-65b-hf" # You can also use "bigscience/bloom" or "bigscience/bloomz" -tokenizer = AutoTokenizer(model_name) +tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoDistributedModelForCausalLM.from_pretrained(model_name) # Embeddings & prompts are on your device, transformer blocks are distributed across the Internet diff --git a/setup.cfg b/setup.cfg index 1e976a6..7341684 100644 --- a/setup.cfg +++ b/setup.cfg @@ -29,7 +29,7 @@ classifiers = package_dir = = src packages = find: -python_requires = >=3.8 +python_requires = >=3.8,<3.11 install_requires = torch>=1.12 bitsandbytes==0.40.1.post1