diff --git a/README.md b/README.md index 4f0f706..2c6b6f3 100644 --- a/README.md +++ b/README.md @@ -54,8 +54,8 @@ import requests from PIL import Image from transformers import BlipProcessor, Blip2ForConditionalGeneration -processor = BlipProcessor.from_pretrained("Salesforce/blip2-flan-t5-xxl") -model = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-flan-t5-xxl") +processor = BlipProcessor.from_pretrained("Salesforce/blip2-opt-2.7b") +model = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-opt-2.7b") img_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg' raw_image = Image.open(requests.get(img_url, stream=True).raw).convert('RGB') @@ -81,8 +81,8 @@ import requests from PIL import Image from transformers import Blip2Processor, Blip2ForConditionalGeneration -processor = Blip2Processor.from_pretrained("Salesforce/blip2-flan-t5-xxl") -model = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-flan-t5-xxl", device_map="auto") +processor = Blip2Processor.from_pretrained("Salesforce/blip2-opt-2.7b") +model = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-opt-2.7b", device_map="auto") img_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg' raw_image = Image.open(requests.get(img_url, stream=True).raw).convert('RGB') @@ -107,8 +107,8 @@ import requests from PIL import Image from transformers import Blip2Processor, Blip2ForConditionalGeneration -processor = Blip2Processor.from_pretrained("Salesforce/blip2-flan-t5-xxl") -model = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-flan-t5-xxl", torch_dtype=torch.float16, device_map="auto") +processor = Blip2Processor.from_pretrained("Salesforce/blip2-opt-2.7b") +model = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-opt-2.7b", torch_dtype=torch.float16, device_map="auto") img_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg' raw_image = Image.open(requests.get(img_url, stream=True).raw).convert('RGB') @@ -133,8 +133,8 @@ import requests from PIL import Image from transformers import Blip2Processor, Blip2ForConditionalGeneration -processor = Blip2Processor.from_pretrained("Salesforce/blip2-flan-t5-xxl") -model = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-flan-t5-xxl", load_in_8bit=True, device_map="auto") +processor = Blip2Processor.from_pretrained("Salesforce/blip2-opt-2.7b") +model = Blip2ForConditionalGeneration.from_pretrained("Salesforce/blip2-opt-2.7b", load_in_8bit=True, device_map="auto") img_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg' raw_image = Image.open(requests.get(img_url, stream=True).raw).convert('RGB')