File size: 1,304 Bytes
5d6f895
dea4744
 
5d6f895
dea4744
5d6f895
dea4744
5d6f895
 
dea4744
5d6f895
 
dea4744
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
---
title: PaliGemma Demo
emoji: 🤲
colorFrom: green
colorTo: yellow
sdk: gradio
sdk_version: 4.22.0
app_file: app.py
pinned: false
license: apache-2.0
---

# PaliGemma Demo

See [Blogpost] and [`big_vision README.md`] for details about the model.


[Blogpost]: https://huggingface.co/blog/paligemma

[`big_vision README.md`]: https://github.com/google-research/big_vision/blob/main/big_vision/configs/proj/paligemma/README.md

## Development

Local testing (CPU, Python 3.12):

```bash
pip -m venv env
. env/bin/activate
pip install -qr requirements-cpu.txt
python app.py
```

Environment variables:

- `MOCK_MODEL=yes`: For quick UI testing.
- `RAM_CACHE_GB=18`: Enables caching of 3 bf16 models in memory: a single bf16
  model is about 5860 MB. Use with care on spaces with little RAM. For example,
  on a `A10G large` space you can cache five models in RAM, so you would set
  `RAM_CACHE_GB=30`.
- `HOST_COLOCATION=4`: If host RAM/disk is shared between 4 processes (e.g. the
  Huggingface `A10 large` Spaces).


Loading models:

- The set of models loaded is defined in `./models.py`.
- You must first acknowledge usage conditions to access models.
- When testing locally, you'll have to run `huggingface_cli login`.
- When running in a Huggingface Space, you'll have to set a `HF_TOKEN` secret.