gliner / extract-entities.py
davanstrien's picture
davanstrien HF Staff
extract-entities: support local parquet/jsonl paths for mounted-bucket workflows
7a8f001 verified
#!/usr/bin/env python3
# /// script
# requires-python = ">=3.10"
# dependencies = [
# "gliner>=0.2.16",
# "datasets>=3.0",
# "huggingface-hub",
# "tqdm",
# ]
# ///
"""
Extract named entities from a text column of a HuggingFace dataset using GLiNER.
GLiNER is a zero-shot NER model: you pass a list of entity types at inference
time (e.g., "Person", "Organization", "Dataset"), no fine-tuning needed.
Examples:
# Local CPU run on a small sample
uv run extract-entities.py \\
librarian-bots/model_cards_with_metadata \\
my-username/model-cards-entities \\
--text-column card \\
--entity-types Person Organization Dataset Model Framework \\
--max-samples 100
# On HF Jobs with a cheap GPU
hf jobs uv run --flavor t4-small \\
--secrets HF_TOKEN \\
https://huggingface.co/datasets/uv-scripts/gliner/raw/main/extract-entities.py \\
librarian-bots/model_cards_with_metadata \\
my-username/model-cards-entities \\
--text-column card \\
--entity-types Person Organization Dataset Model Framework \\
--max-samples 5000
Output schema: original dataset columns + new `entities` column, where each
row's `entities` is a list of dicts:
{"start": int, "end": int, "text": str, "label": str, "score": float}
"""
import argparse
import logging
import os
import sys
import time
from typing import Any, Dict, List
import torch
from datasets import Dataset, Features, Sequence, Value, load_dataset
from huggingface_hub import DatasetCard
logging.basicConfig(
level=logging.INFO,
format="%(asctime)s - %(levelname)s - %(message)s",
datefmt="%H:%M:%S",
)
log = logging.getLogger(__name__)
def parse_args():
p = argparse.ArgumentParser(
description="Bootstrap NER labels with GLiNER over a HuggingFace dataset",
formatter_class=argparse.RawDescriptionHelpFormatter,
epilog=__doc__,
)
p.add_argument(
"input_dataset",
help=(
"Input. Either a HF dataset ID (e.g. 'org/dataset') or a local path "
"to parquet/jsonl file(s) — useful when running in HF Jobs with a mounted "
"bucket: '-v hf://buckets/<ns>/<bucket>:/input' then pass '/input/cards.parquet'."
),
)
p.add_argument(
"output_dataset",
help=(
"Output HF dataset ID (e.g. 'user/output'). The script always pushes results "
"to a HF dataset repo regardless of where input came from."
),
)
p.add_argument(
"--text-column",
default="text",
help="Name of the text column to run NER over (default: 'text')",
)
p.add_argument(
"--entity-types",
nargs="+",
required=True,
help="Space-separated entity types to extract (e.g., Person Organization Date)",
)
p.add_argument(
"--gliner-model",
default="urchade/gliner_multi-v2.1",
help="GLiNER model ID (default: urchade/gliner_multi-v2.1, multilingual)",
)
p.add_argument(
"--threshold",
type=float,
default=0.5,
help="Confidence threshold for entity extraction (default: 0.5)",
)
p.add_argument(
"--split",
default="train",
help="Dataset split to process (default: train)",
)
p.add_argument(
"--max-samples",
type=int,
default=None,
help="Process at most N samples (default: all)",
)
p.add_argument(
"--batch-size",
type=int,
default=8,
help="Batch size for inference (default: 8)",
)
p.add_argument(
"--device",
choices=["auto", "cpu", "cuda"],
default="auto",
help="Device for inference (default: auto — uses CUDA if available)",
)
p.add_argument(
"--max-text-chars",
type=int,
default=8000,
help="Truncate texts longer than this many characters (default: 8000)",
)
p.add_argument(
"--private",
action="store_true",
help="Push the output dataset as private",
)
return p.parse_args()
def resolve_device(arg: str) -> str:
if arg == "cpu":
return "cpu"
if arg == "cuda":
if not torch.cuda.is_available():
log.warning("--device cuda requested but CUDA not available; falling back to CPU")
return "cpu"
return "cuda"
return "cuda" if torch.cuda.is_available() else "cpu"
def normalize_entity(ent: Dict[str, Any]) -> Dict[str, Any]:
return {
"start": int(ent["start"]),
"end": int(ent["end"]),
"text": str(ent["text"]),
"label": str(ent["label"]),
"score": float(ent.get("score", 0.0)),
}
def build_card(args, n_processed: int, n_entities: int, elapsed_s: float, device: str) -> str:
return f"""---
license: apache-2.0
tags:
- ner
- gliner
- zero-shot
- bootstrap
- uv-script
size_categories:
- n<10K
---
# {args.output_dataset}
Bootstrap NER dataset produced by [`{args.gliner_model}`](https://huggingface.co/{args.gliner_model}) over [`{args.input_dataset}`](https://huggingface.co/datasets/{args.input_dataset}).
Generated using [`uv-scripts/gliner/extract-entities.py`](https://huggingface.co/datasets/uv-scripts/gliner).
## Provenance
| | |
|---|---|
| Source dataset | `{args.input_dataset}` (split `{args.split}`) |
| Text column | `{args.text_column}` |
| Bootstrap model | `{args.gliner_model}` |
| Entity types | `{', '.join(args.entity_types)}` |
| Confidence threshold | {args.threshold} |
| Samples processed | {n_processed} |
| Total entities extracted | {n_entities} |
| Inference device | `{device}` |
| Wall clock | {elapsed_s:.1f}s ({n_processed / max(elapsed_s, 1e-9):.2f} samples/s) |
## Schema
Original `{args.input_dataset}` columns plus an `entities` column:
```python
entities: list of {{
"start": int, # character offset, inclusive
"end": int, # character offset, exclusive
"text": str, # the matched span
"label": str, # one of {args.entity_types}
"score": float, # GLiNER confidence in [0, 1]
}}
```
## Caveats
- These are **bootstrap labels**, not human-reviewed. Treat low-confidence (< 0.7) entities as candidates for review.
- GLiNER is zero-shot: changing `--entity-types` changes what it extracts, but quality varies by entity type.
- Long texts were truncated at {args.max_text_chars} characters before inference.
"""
def is_local_path(s: str) -> bool:
"""Heuristic: treat as local path if it starts with / or ./ or contains a known data extension."""
if s.startswith(("/", "./", "../")):
return True
if any(s.endswith(ext) for ext in (".parquet", ".jsonl", ".json", ".csv")):
return True
if "*" in s and any(ext in s for ext in (".parquet", ".jsonl", ".json", ".csv")):
return True
return False
def load_input(spec: str, split: str):
"""Load either a HF dataset by ID, or a local parquet/jsonl path (incl. globs)."""
if is_local_path(spec):
ext = ".parquet" if ".parquet" in spec else \
".jsonl" if spec.endswith(".jsonl") else \
".json" if spec.endswith(".json") else \
".csv" if spec.endswith(".csv") else ".parquet"
loader = {".parquet": "parquet", ".jsonl": "json", ".json": "json", ".csv": "csv"}[ext]
log.info("Loading local %s file(s): %s", loader, spec)
return load_dataset(loader, data_files=spec, split="train")
log.info("Loading HF dataset '%s' split=%s ...", spec, split)
return load_dataset(spec, split=split, streaming=False)
def main():
args = parse_args()
device = resolve_device(args.device)
ds = load_input(args.input_dataset, args.split)
if args.text_column not in ds.column_names:
sys.exit(
f"--text-column '{args.text_column}' not in {ds.column_names}"
)
if args.max_samples is not None:
n = min(args.max_samples, len(ds))
ds = ds.select(range(n))
log.info("Selected %d samples", n)
else:
log.info("Processing full split: %d samples", len(ds))
log.info("Loading GLiNER %s on %s ...", args.gliner_model, device)
from gliner import GLiNER
model = GLiNER.from_pretrained(args.gliner_model)
if device == "cuda":
model = model.to("cuda")
model.eval()
n_entities = 0
started = time.time()
def add_entities(batch: Dict[str, List]) -> Dict[str, List]:
nonlocal n_entities
texts = [
(t or "")[: args.max_text_chars] for t in batch[args.text_column]
]
entities_per_row = []
for text in texts:
if not text.strip():
entities_per_row.append([])
continue
try:
ents = model.predict_entities(
text, args.entity_types, threshold=args.threshold
)
except Exception as e:
log.warning("predict_entities failed: %s; returning []", e)
ents = []
normalized = [normalize_entity(e) for e in ents]
n_entities += len(normalized)
entities_per_row.append(normalized)
return {"entities": entities_per_row}
log.info("Running inference (batch_size=%d) ...", args.batch_size)
ds = ds.map(
add_entities,
batched=True,
batch_size=args.batch_size,
desc="Extracting entities",
)
elapsed = time.time() - started
log.info(
"Done. %d entities across %d samples in %.1fs (%.2f samples/s)",
n_entities, len(ds), elapsed, len(ds) / max(elapsed, 1e-9),
)
log.info("Pushing to %s ...", args.output_dataset)
ds.push_to_hub(args.output_dataset, private=args.private)
DatasetCard(build_card(args, len(ds), n_entities, elapsed, device)).push_to_hub(
args.output_dataset, repo_type="dataset"
)
log.info("Done: https://huggingface.co/datasets/%s", args.output_dataset)
if __name__ == "__main__":
main()