paligemma / README.md
mjlm's picture
Initial commit.
dea4744
|
raw
history blame
1.3 kB
---
title: PaliGemma Demo
emoji: 🤲
colorFrom: green
colorTo: yellow
sdk: gradio
sdk_version: 4.22.0
app_file: app.py
pinned: false
license: apache-2.0
---
# PaliGemma Demo
See [Blogpost] and [`big_vision README.md`] for details about the model.
[Blogpost]: https://huggingface.co/blog/paligemma
[`big_vision README.md`]: https://github.com/google-research/big_vision/blob/main/big_vision/configs/proj/paligemma/README.md
## Development
Local testing (CPU, Python 3.12):
```bash
pip -m venv env
. env/bin/activate
pip install -qr requirements-cpu.txt
python app.py
```
Environment variables:
- `MOCK_MODEL=yes`: For quick UI testing.
- `RAM_CACHE_GB=18`: Enables caching of 3 bf16 models in memory: a single bf16
model is about 5860 MB. Use with care on spaces with little RAM. For example,
on a `A10G large` space you can cache five models in RAM, so you would set
`RAM_CACHE_GB=30`.
- `HOST_COLOCATION=4`: If host RAM/disk is shared between 4 processes (e.g. the
Huggingface `A10 large` Spaces).
Loading models:
- The set of models loaded is defined in `./models.py`.
- You must first acknowledge usage conditions to access models.
- When testing locally, you'll have to run `huggingface_cli login`.
- When running in a Huggingface Space, you'll have to set a `HF_TOKEN` secret.