[AZ-180] Update Jetson deployment documentation and remove obsolete task file

- Added Jetson-specific deployment instructions to `deploy_scripts.md`, detailing prerequisites and service management.
- Updated `deploy_status_report.md` to reflect the completion of the AZ-180 cycle and the readiness of Jetson support.
- Removed outdated task documentation for Jetson Orin Nano support from the todo list.

Made-with: Cursor
This commit is contained in:
Oleksandr Bezdieniezhnykh
2026-04-02 16:58:57 +03:00
parent 3984507221
commit 2c35e59a77
9 changed files with 377 additions and 7 deletions
+143
View File
@@ -0,0 +1,143 @@
#!/bin/bash
# Deploy Azaion detections demo stack to a Jetson over SSH.
#
# Usage:
# JETSON_HOST=192.168.x.x bash scripts/deploy_demo_jetson.sh \
# --onnx /local/path/azaion.onnx \
# --classes /local/path/classes.json \
# [--int8-cache /local/path/azaion.int8_calib.cache] \
# [--calibration-images /local/path/images/]
#
# Optional env vars:
# JETSON_HOST (required) IP or hostname of the Jetson
# JETSON_USER SSH user (default: jetson)
# REMOTE_DIR Path on Jetson to deploy into (default: ~/detections)
set -euo pipefail
SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
PROJECT_ROOT="$(dirname "$SCRIPT_DIR")"
JETSON_HOST="${JETSON_HOST:-}"
JETSON_USER="${JETSON_USER:-jetson}"
REMOTE_DIR="${REMOTE_DIR:-~/detections}"
ONNX_PATH=""
CLASSES_PATH=""
INT8_CACHE_PATH=""
CALIBRATION_IMAGES=""
usage() {
echo "Usage: JETSON_HOST=<ip> bash $0 --onnx <path> --classes <path> [options]"
echo ""
echo "Required:"
echo " --onnx <path> Local path to azaion.onnx"
echo " --classes <path> Local path to classes.json"
echo ""
echo "Optional:"
echo " --int8-cache <path> Local path to azaion.int8_calib.cache (skips calibration)"
echo " --calibration-images <dir> Local image directory; rsync to Jetson and run INT8 calibration"
echo " --help Show this message"
echo ""
echo "Env vars:"
echo " JETSON_HOST (required) Jetson IP or hostname"
echo " JETSON_USER SSH user (default: jetson)"
echo " REMOTE_DIR Deploy directory on Jetson (default: ~/detections)"
exit 0
}
while [[ $# -gt 0 ]]; do
case "$1" in
--onnx) ONNX_PATH="$2"; shift 2 ;;
--classes) CLASSES_PATH="$2"; shift 2 ;;
--int8-cache) INT8_CACHE_PATH="$2"; shift 2 ;;
--calibration-images) CALIBRATION_IMAGES="$2"; shift 2 ;;
--help) usage ;;
*) echo "Unknown argument: $1"; usage ;;
esac
done
[[ -z "$JETSON_HOST" ]] && { echo "ERROR: JETSON_HOST is required"; exit 1; }
[[ -z "$ONNX_PATH" ]] && { echo "ERROR: --onnx is required"; exit 1; }
[[ -z "$CLASSES_PATH" ]] && { echo "ERROR: --classes is required"; exit 1; }
[[ -f "$ONNX_PATH" ]] || { echo "ERROR: ONNX file not found: $ONNX_PATH"; exit 1; }
[[ -f "$CLASSES_PATH" ]] || { echo "ERROR: classes.json not found: $CLASSES_PATH"; exit 1; }
SSH="ssh ${JETSON_USER}@${JETSON_HOST}"
SCP="scp"
echo "=== Azaion Demo — Jetson Deployment ==="
echo " Host: ${JETSON_USER}@${JETSON_HOST}"
echo " Remote dir: ${REMOTE_DIR}"
echo ""
# ── 1. Sync project ─────────────────────────────────────────────────────────
echo "--- Syncing project files ---"
$SSH "mkdir -p ${REMOTE_DIR}/demo/models"
rsync -az --exclude='.git' --exclude='__pycache__' --exclude='*.pyc' \
--exclude='*.egg-info' --exclude='.venv' --exclude='demo/models' \
"${PROJECT_ROOT}/" "${JETSON_USER}@${JETSON_HOST}:${REMOTE_DIR}/"
# ── 2. Upload model artifacts ────────────────────────────────────────────────
echo "--- Uploading model artifacts ---"
$SCP "$ONNX_PATH" "${JETSON_USER}@${JETSON_HOST}:${REMOTE_DIR}/demo/models/azaion.onnx"
$SCP "$CLASSES_PATH" "${JETSON_USER}@${JETSON_HOST}:${REMOTE_DIR}/demo/models/classes.json"
if [[ -n "$INT8_CACHE_PATH" ]]; then
[[ -f "$INT8_CACHE_PATH" ]] || { echo "ERROR: INT8 cache not found: $INT8_CACHE_PATH"; exit 1; }
echo "--- Uploading INT8 calibration cache ---"
$SCP "$INT8_CACHE_PATH" "${JETSON_USER}@${JETSON_HOST}:${REMOTE_DIR}/demo/models/azaion.int8_calib.cache"
fi
# ── 3. Optional: run INT8 calibration on Jetson ──────────────────────────────
if [[ -n "$CALIBRATION_IMAGES" ]] && [[ -z "$INT8_CACHE_PATH" ]]; then
[[ -d "$CALIBRATION_IMAGES" ]] || { echo "ERROR: calibration images dir not found: $CALIBRATION_IMAGES"; exit 1; }
echo "--- Syncing calibration images to Jetson ---"
$SSH "mkdir -p ${REMOTE_DIR}/demo/calibration"
rsync -az "${CALIBRATION_IMAGES}/" "${JETSON_USER}@${JETSON_HOST}:${REMOTE_DIR}/demo/calibration/"
echo "--- Building detections image for calibration ---"
$SSH "cd ${REMOTE_DIR} && docker compose -f docker-compose.demo-jetson.yml build detections"
echo "--- Running INT8 calibration (this takes several minutes) ---"
$SSH "cd ${REMOTE_DIR} && docker compose -f docker-compose.demo-jetson.yml run --rm \
-v ${REMOTE_DIR}/demo/calibration:/calibration \
detections \
python3 scripts/generate_int8_cache.py \
--images-dir /calibration \
--onnx /models/azaion.onnx \
--output /models/azaion.int8_calib.cache"
echo "--- Calibration cache written to ${REMOTE_DIR}/demo/models/azaion.int8_calib.cache ---"
fi
# ── 4. Start services ────────────────────────────────────────────────────────
echo "--- Building and starting services ---"
$SSH "cd ${REMOTE_DIR} && docker compose -f docker-compose.demo-jetson.yml up -d --build"
# ── 5. Health check ───────────────────────────────────────────────────────────
echo "--- Health check ---"
HEALTH_URL="http://${JETSON_HOST}:8080/health"
MAX_RETRIES=15
RETRY_INTERVAL=5
for i in $(seq 1 $MAX_RETRIES); do
STATUS=$(curl -s -o /dev/null -w "%{http_code}" "$HEALTH_URL" 2>/dev/null || true)
if [[ "$STATUS" == "200" ]]; then
echo ""
echo "=== Demo is live at http://${JETSON_HOST}:8080 ==="
echo ""
echo "Endpoints:"
echo " POST http://${JETSON_HOST}:8080/detect/image"
echo " POST http://${JETSON_HOST}:8080/detect/video"
echo " GET http://${JETSON_HOST}:8080/health"
echo ""
echo "Note: On first start the service converts azaion.onnx to a TRT engine."
echo " Check /health until AI status shows 'enabled'."
exit 0
fi
echo " Waiting for service… (${i}/${MAX_RETRIES}, HTTP ${STATUS})"
sleep "$RETRY_INTERVAL"
done
echo "ERROR: Health check failed after $((MAX_RETRIES * RETRY_INTERVAL))s"
echo "Check logs with: ssh ${JETSON_USER}@${JETSON_HOST} \"cd ${REMOTE_DIR} && docker compose -f docker-compose.demo-jetson.yml logs detections\""
exit 1
+141
View File
@@ -0,0 +1,141 @@
#!/usr/bin/env python3
"""
Generate an INT8 calibration cache for TensorRT on Jetson.
Run this INSIDE the Jetson Docker container:
docker compose -f docker-compose.demo-jetson.yml run --rm \
-v /path/to/images:/calibration \
detections \
python3 scripts/generate_int8_cache.py \
--images-dir /calibration \
--onnx /models/azaion.onnx \
--output /models/azaion.int8_calib.cache
The cache file must be in the loader's models volume so the detections
service can download it on startup via the Loader API.
"""
import argparse
import sys
from pathlib import Path
import cv2
import numpy as np
def parse_args():
parser = argparse.ArgumentParser()
parser.add_argument("--images-dir", required=True, help="Directory with calibration images (JPG/PNG)")
parser.add_argument("--onnx", required=True, help="Path to azaion.onnx")
parser.add_argument("--output", default="azaion.int8_calib.cache")
parser.add_argument("--input-size", type=int, default=1280, help="Model input H=W (default 1280)")
parser.add_argument("--num-samples", type=int, default=500)
return parser.parse_args()
def collect_images(images_dir: str, num_samples: int) -> list[Path]:
root = Path(images_dir)
images: list[Path] = []
for pattern in ("**/*.jpg", "**/*.jpeg", "**/*.png"):
images += sorted(root.glob(pattern))
return images[:num_samples]
def preprocess(path: Path, h: int, w: int) -> np.ndarray | None:
img = cv2.imread(str(path))
if img is None:
return None
img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB)
img = cv2.resize(img, (w, h))
img = img.astype(np.float32) / 255.0
return np.ascontiguousarray(img.transpose(2, 0, 1)[np.newaxis]) # NCHW
def main():
args = parse_args()
try:
import pycuda.autoinit # noqa: F401
import pycuda.driver as cuda
import tensorrt as trt
except ImportError as e:
print(f"ERROR: {e}\nRun this script inside the Jetson Docker container.", file=sys.stderr)
sys.exit(1)
images = collect_images(args.images_dir, args.num_samples)
if not images:
print(f"No images found in {args.images_dir}", file=sys.stderr)
sys.exit(1)
print(f"Using {len(images)} calibration images")
H = W = args.input_size
class _ImageCalibrator(trt.IInt8EntropyCalibrator2):
def __init__(self):
super().__init__()
self._idx = 0
self._buf = cuda.mem_alloc(3 * H * W * 4)
def get_batch_size(self):
return 1
def get_batch(self, names):
while self._idx < len(images):
arr = preprocess(images[self._idx], H, W)
self._idx += 1
if arr is None:
continue
cuda.memcpy_htod(self._buf, arr)
return [int(self._buf)]
return None
def read_calibration_cache(self):
return None
def write_calibration_cache(self, cache):
with open(args.output, "wb") as f:
f.write(cache)
print(f"Cache written → {args.output}")
onnx_data = Path(args.onnx).read_bytes()
logger = trt.Logger(trt.Logger.INFO)
explicit_batch = 1 << int(trt.NetworkDefinitionCreationFlag.EXPLICIT_BATCH)
with (
trt.Builder(logger) as builder,
builder.create_network(explicit_batch) as network,
trt.OnnxParser(network, logger) as parser,
builder.create_builder_config() as config,
):
config.set_memory_pool_limit(trt.MemoryPoolType.WORKSPACE, 4 * 1024 ** 3)
config.set_flag(trt.BuilderFlag.INT8)
if builder.platform_has_fast_fp16:
config.set_flag(trt.BuilderFlag.FP16)
calibrator = _ImageCalibrator()
config.int8_calibrator = calibrator
if not parser.parse(onnx_data):
for i in range(parser.num_errors):
print(parser.get_error(i), file=sys.stderr)
sys.exit(1)
inp = network.get_input(0)
shape = inp.shape
C = shape[1]
if shape[0] == -1:
profile = builder.create_optimization_profile()
profile.set_shape(inp.name, (1, C, H, W), (1, C, H, W), (1, C, H, W))
config.add_optimization_profile(profile)
print("Building TensorRT engine with INT8 calibration (several minutes on Jetson)…")
plan = builder.build_serialized_network(network, config)
if plan is None:
print("Engine build failed", file=sys.stderr)
sys.exit(1)
print("Done. Upload the cache to the Loader before (re)starting the detections service.")
if __name__ == "__main__":
main()