ydshieh HF staff commited on
Commit
a62a272
1 Parent(s): 817d063

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -11
README.md CHANGED
@@ -9,43 +9,32 @@ The model can be used as follows:
9
  ```python
10
 
11
  import requests
12
-
13
  from PIL import Image
14
-
15
  from transformers import ViTFeatureExtractor, AutoTokenizer, FlaxVisionEncoderDecoderModel
16
 
17
  loc = "ydshieh/flax-vit-gpt2-coco-en"
18
 
19
  feature_extractor = ViTFeatureExtractor.from_pretrained(loc)
20
-
21
  tokenizer = AutoTokenizer.from_pretrained(loc)
22
-
23
  model = FlaxVisionEncoderDecoderModel.from_pretrained(loc)
24
 
25
  # We will verify our results on an image of cute cats
26
-
27
  url = "http://images.cocodataset.org/val2017/000000039769.jpg"
28
-
29
  with Image.open(requests.get(url, stream=True).raw) as img:
30
-
31
  pixel_values = feature_extractor(images=img, return_tensors="np").pixel_values
32
 
33
  def generate_step(pixel_values):
34
 
35
  output_ids = model.generate(pixel_values, max_length=16, num_beams=4).sequences
36
-
37
  preds = tokenizer.batch_decode(output_ids, skip_special_tokens=True)
38
-
39
  preds = [pred.strip() for pred in preds]
40
 
41
  return preds
42
 
43
  preds = generate_step(pixel_values)
44
-
45
  print(preds)
46
 
47
  # should produce
48
-
49
  # ['a cat laying on top of a couch next to another cat']
50
 
51
  ```
 
9
  ```python
10
 
11
  import requests
 
12
  from PIL import Image
 
13
  from transformers import ViTFeatureExtractor, AutoTokenizer, FlaxVisionEncoderDecoderModel
14
 
15
  loc = "ydshieh/flax-vit-gpt2-coco-en"
16
 
17
  feature_extractor = ViTFeatureExtractor.from_pretrained(loc)
 
18
  tokenizer = AutoTokenizer.from_pretrained(loc)
 
19
  model = FlaxVisionEncoderDecoderModel.from_pretrained(loc)
20
 
21
  # We will verify our results on an image of cute cats
 
22
  url = "http://images.cocodataset.org/val2017/000000039769.jpg"
 
23
  with Image.open(requests.get(url, stream=True).raw) as img:
 
24
  pixel_values = feature_extractor(images=img, return_tensors="np").pixel_values
25
 
26
  def generate_step(pixel_values):
27
 
28
  output_ids = model.generate(pixel_values, max_length=16, num_beams=4).sequences
 
29
  preds = tokenizer.batch_decode(output_ids, skip_special_tokens=True)
 
30
  preds = [pred.strip() for pred in preds]
31
 
32
  return preds
33
 
34
  preds = generate_step(pixel_values)
 
35
  print(preds)
36
 
37
  # should produce
 
38
  # ['a cat laying on top of a couch next to another cat']
39
 
40
  ```