Image-Text-to-Text
Transformers
Safetensors
English
Chinese
llava
pretraining
vision-language
llm
lmm
Inference Endpoints
bczhou commited on
Commit
b08ccca
1 Parent(s): 282a049

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -37,7 +37,7 @@ image = Image.open(requests.get(url, stream=True).raw)
37
  prompt = "USER: <image>\nWhat does the label 15 represent? (1) lava (2) core (3) tunnel (4) ash cloud\nASSISTANT:"
38
  outputs = pipe(image, prompt=prompt, generate_kwargs={"max_new_tokens": 200})
39
  print(outputs)
40
- >>> {"generated_text": "\nUSER: What does the label 15 represent? (1) lava (2) core (3) tunnel (4) ash cloud\nASSISTANT: Lava"}
41
  ```
42
 
43
  ### Using pure `transformers`:
 
37
  prompt = "USER: <image>\nWhat does the label 15 represent? (1) lava (2) core (3) tunnel (4) ash cloud\nASSISTANT:"
38
  outputs = pipe(image, prompt=prompt, generate_kwargs={"max_new_tokens": 200})
39
  print(outputs)
40
+ >>> {"generated_text': 'USER: \nWhat does the label 15 represent? (1) lava (2) core (3) tunnel (4) ash cloud\nASSISTANT: The label 15 represents lava, which is a type of volcanic rock."}
41
  ```
42
 
43
  ### Using pure `transformers`: