matthewlyleolson commited on
Commit
b5674bf
1 Parent(s): d51c5d1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -51,7 +51,7 @@ pipeline_tag: image-text-to-text
51
 
52
  | Model Details | Description |
53
  | ----------- | ----------- |
54
- | Authors | Intel: [Musashi Hinck](https://huggingface.co/musashihinck), [Matthew Olson](https://huggingface.co/matthewlyleolson), [David Cobbley](https://huggingface.co/djcobble), [Shao-Yen Tseng](https://huggingface.co/shaoyent), [Vasudev Lal](https://huggingface.co/vasudevlal) |
55
  | Date | March 2024 |
56
  | Version | 1 |
57
  | Type | Large multimodal model (LMM) |
@@ -71,7 +71,7 @@ This model card was created by [Benjamin Consolvo](https://huggingface.co/bconso
71
 
72
  ### How to use
73
 
74
- Currently, using `llava-gemma` requires a [modified preprocessor](./processing_llavagemma.py). _We are currently working on modifying the `LlavaProcessor` class to streamline usage (see [PR #30030](https://github.com/huggingface/transformers/pull/30030)). Expect updates soon._
75
 
76
  For current usage, see [`usage.py`](./usage.py) or the following code block:
77
 
@@ -83,7 +83,8 @@ from transformers import (
83
  AutoTokenizer,
84
  CLIPImageProcessor
85
  )
86
- from processing_llavagemma import LlavaGemmaProcessor # This is in this repo
 
87
 
88
  checkpoint = "Intel/llava-gemma-2b"
89
 
 
51
 
52
  | Model Details | Description |
53
  | ----------- | ----------- |
54
+ | Authors | Intel: [Musashi Hinck*](https://huggingface.co/musashihinck), [Matthew L. Olson*](https://huggingface.co/matthewlyleolson), [David Cobbley](https://huggingface.co/djcobble), [Shao-Yen Tseng](https://huggingface.co/shaoyent), [Vasudev Lal](https://huggingface.co/vasudevlal) |
55
  | Date | March 2024 |
56
  | Version | 1 |
57
  | Type | Large multimodal model (LMM) |
 
71
 
72
  ### How to use
73
 
74
+ Using `llava-gemma` requires a [modified preprocessor](./processing_llavagemma.py) if your transformers version is < 4.41.1
75
 
76
  For current usage, see [`usage.py`](./usage.py) or the following code block:
77
 
 
83
  AutoTokenizer,
84
  CLIPImageProcessor
85
  )
86
+ #In this repo, needed for version < 4.41.1
87
+ #from processing_llavagemma import LlavaGemmaProcessor
88
 
89
  checkpoint = "Intel/llava-gemma-2b"
90