PaddleOCR / app.py
02alexander's picture
create description
cbcc4bf
raw
history blame
1.99 kB
from __future__ import annotations
import os
from pathlib import Path
import gradio as gr # type: ignore
import rerun as rr
import rerun.blueprint as rrb
from gradio_rerun import Rerun # type: ignore
from ocr import detect_and_log_layout # type: ignore
@rr.thread_local_stream("PaddleOCR")
def log_to_rr(img_path: Path):
print(img_path)
stream = rr.binary_stream()
blueprint = rrb.Blueprint(
rrb.Vertical(
rrb.Spatial2DView(name="Input", contents=["Image/**"]),
),
collapse_panels=True,
)
rr.send_blueprint(blueprint)
detect_and_log_layout(img_path)
yield stream.read()
DESCRIPTION = """
This space demonstrates the ability to visualize and verify the document layout analysis and text detection using [PaddleOCR](https://github.com/PaddlePaddle/PaddleOCR).
The [PP-Structure](https://github.com/PaddlePaddle/PaddleOCR/tree/main/ppstructure) used for this task, which is an intelligent document analysis system developed by the PaddleOCR team, which aims to help developers better complete tasks related to document understanding such as layout analysis and table recognition.
"""
with gr.Blocks() as demo:
gr.Markdown(DESCRIPTION)
with gr.Row():
with gr.Column(scale=1):
with gr.Row():
input_image = gr.Image(label="Input Image", image_mode="RGBA", sources="upload", type="filepath")
with gr.Row():
button = gr.Button()
with gr.Row():
gr.Examples(
examples=[os.path.join("examples", img_name) for img_name in sorted(os.listdir("examples"))],
inputs=[input_image],
label="Examples",
cache_examples=False,
examples_per_page=12,
)
with gr.Column(scale=4):
viewer = Rerun(streaming=True, height=900)
button.click(log_to_rr, inputs=[input_image], outputs=[viewer])
demo.launch()