fashxp commited on
Commit
d634727
1 Parent(s): 301404d

fix readme

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -1,19 +1,19 @@
1
  ---
2
  tags:
3
  - vision
4
- - text-to-image
5
  - endpoints-template
6
  inference: true
7
- pipeline_tag: text-to-image
8
  base_model: Salesforce/blip-image-captioning-base
9
  library_name: generic
10
  ---
11
 
12
- # Fork of [Salesforce/blip-image-captioning-base](https://huggingface.co/openai/clip-vit-base-patch32) for a `text-to-image` Inference endpoint.
13
 
14
- > Based on https://huggingface.co/sergeipetrov/blip_captioning
15
 
16
- This repository implements a `custom` task for `text-to-image` for 🤗 Inference Endpoints to allow image capturing.
17
  The code for the customized pipeline is in the handler.py.
18
 
19
  To use deploy this model an Inference Endpoint you have to select `Custom` as task to use the `handler.py` file.
 
1
  ---
2
  tags:
3
  - vision
4
+ - image-to-text
5
  - endpoints-template
6
  inference: true
7
+ pipeline_tag: image-to-text
8
  base_model: Salesforce/blip-image-captioning-base
9
  library_name: generic
10
  ---
11
 
12
+ # Fork of [Salesforce/blip-image-captioning-base](https://huggingface.co/openai/clip-vit-base-patch32) for a `image-to-text` Inference endpoint.
13
 
14
+ > Inspired by https://huggingface.co/sergeipetrov/blip_captioning
15
 
16
+ This repository implements a `custom` task for `image-to-text` for 🤗 Inference Endpoints to allow image capturing.
17
  The code for the customized pipeline is in the handler.py.
18
 
19
  To use deploy this model an Inference Endpoint you have to select `Custom` as task to use the `handler.py` file.