| | import typer |
| | from inference_cli.lib import check_inference_server_status, start_inference_container |
| | from typing_extensions import Annotated |
| |
|
| | server_app = typer.Typer( |
| | help="""Commands for running the inference server locally. \n |
| | Supported devices targets are x86 CPU, ARM64 CPU, and NVIDIA GPU.""" |
| | ) |
| |
|
| |
|
| | @server_app.command() |
| | def start( |
| | port: Annotated[ |
| | int, |
| | typer.Option( |
| | "--port", |
| | "-p", |
| | help="Port to run the inference server on (default is 9001).", |
| | ), |
| | ] = 9001, |
| | rf_env: Annotated[ |
| | str, |
| | typer.Option( |
| | "--rf-env", |
| | "-rfe", |
| | help="Roboflow environment to run the inference server with (default is roboflow-platform).", |
| | ), |
| | ] = "roboflow-platform", |
| | ): |
| | start_inference_container("", port=port, project=rf_env) |
| |
|
| |
|
| | @server_app.command() |
| | def status(): |
| | print("Checking status of inference server.") |
| | check_inference_server_status() |
| |
|
| |
|
| | if __name__ == "__main__": |
| | server_app() |
| |
|