Files
detections-semantic/_docs/00_research/02_fact_cards.md
T
Oleksandr Bezdieniezhnykh 8e2ecf50fd Initial commit
Made-with: Cursor
2026-03-26 00:20:30 +02:00

7.6 KiB

Fact Cards

Fact #1

  • Statement: Jetson Orin Nano Super has 7.6GB total unified memory, but only ~3.7GB free GPU memory after OS/system overhead in a Docker container.
  • Source: Source #21
  • Phase: Assessment
  • Target Audience: Edge AI multi-model deployment on Orin Nano Super
  • Confidence: High
  • Related Dimension: Memory contention

Fact #2

  • Statement: A single TensorRT engine (YOLOv8-OBB) consumes ~2.6GB on Jetson Orin Nano. cuDNN/CUDA binary loading adds ~940MB-1.1GB overhead per engine initialization.
  • Source: Source #20
  • Phase: Assessment
  • Target Audience: TRT multi-engine memory planning
  • Confidence: High
  • Related Dimension: Memory contention

Fact #3

  • Statement: Running VLA + YOLO detection concurrently on Orin Nano Super is described as "mostly theoretical" in 2025 surveys. GPU sharing causes 10-40% latency jitter.
  • Source: Source #18
  • Phase: Assessment
  • Target Audience: Multi-model concurrent inference
  • Confidence: ⚠️ Medium (survey, not primary benchmark)
  • Related Dimension: Memory contention, performance

Fact #4

  • Statement: NVIDIA recommends using a single TRT engine with async CUDA streams over multiple separate engines for GPU efficiency. Multiple engines need CUDA context push/pop management.
  • Source: Source #19
  • Phase: Assessment
  • Target Audience: TRT engine management
  • Confidence: High
  • Related Dimension: Memory contention, architecture

Fact #5

  • Statement: YOLO26 exhibits bounding box drift and inaccurate confidence scores when deployed via TensorRT on Jetson Orin Nano in C++. This is an architecture-specific export issue not present in YOLOv8.
  • Source: Source #22
  • Phase: Assessment
  • Target Audience: YOLO26/YOLOE-26 TRT deployment
  • Confidence: High
  • Related Dimension: YOLOE-26 viability, deployment risk

Fact #6

  • Statement: YOLO26n INT8 TensorRT export fails during calibration graph optimization on Jetson Orin with TensorRT 10.3.0 / JetPack 6. ONNX export succeeds but TRT build crashes.
  • Source: Source #23
  • Phase: Assessment
  • Target Audience: YOLO26 edge deployment
  • Confidence: High
  • Related Dimension: YOLOE-26 viability, deployment risk

Fact #7

  • Statement: YOLOE supports multimodal fusion of text + visual prompts with two modes: concat (zero overhead) and weighted-sum (fuse_alpha). This can improve robustness over text-only or visual-only prompts.
  • Source: Source #30
  • Phase: Assessment
  • Target Audience: YOLOE prompt strategy
  • Confidence: High
  • Related Dimension: YOLOE-26 accuracy

Fact #8

  • Statement: YOLOE text prompts are trained on LVIS (1203 categories) and COCO. Military concealment classes (dugouts, branch camouflage, FPV hideouts) are far out-of-distribution from training data. No published benchmarks for this domain.
  • Source: Sources #2, #3 (inferred from training data descriptions)
  • Phase: Assessment
  • Target Audience: YOLOE-26 zero-shot accuracy
  • Confidence: ⚠️ Medium (inference from known training data)
  • Related Dimension: YOLOE-26 accuracy

Fact #9

  • Statement: Smaller YOLO models (commonly used on edge devices) are more vulnerable to adversarial patch attacks than larger counterparts, creating a latency-security trade-off.
  • Source: Source #26
  • Phase: Assessment
  • Target Audience: Edge AI security
  • Confidence: High
  • Related Dimension: Security

Fact #10

  • Statement: PatchBlock is a lightweight CPU-based preprocessing module that recovers up to 77% of model accuracy under adversarial patch attacks with minimal clean accuracy loss.
  • Source: Source #24
  • Phase: Assessment
  • Target Audience: Edge AI adversarial defense
  • Confidence: High
  • Related Dimension: Security

Fact #11

  • Statement: TensorRT-LLM developers explicitly stated they "do not aim to support models on edge devices/platforms" when asked about VLM support on Orin NX.
  • Source: Source #37
  • Phase: Assessment
  • Target Audience: VLM runtime selection
  • Confidence: High
  • Related Dimension: VLM integration

Fact #12

  • Statement: vLLM can deploy 2B models on Jetson Orin Nano 8GB. Shared memory must be increased to 8GB. Memory management is critical. Bottleneck is memory bandwidth (68 GB/s), not compute (67 TOPS).
  • Source: Sources #35, #36
  • Phase: Assessment
  • Target Audience: VLM runtime on Jetson
  • Confidence: High
  • Related Dimension: VLM integration

Fact #13

  • Statement: Cosmos-Reason2-2B achieves 4.7 tok/s on Jetson Orin Nano Super with W4A16 quantization. Llama-3.1-8B W4A16 achieves 44.19 tok/s (text-only). VLMs are significantly slower due to vision encoder overhead.
  • Source: Sources #5, #16
  • Phase: Assessment
  • Target Audience: VLM inference speed estimation
  • Confidence: High
  • Related Dimension: VLM integration, performance

Fact #14

  • Statement: A 1.5B Q4 model on Jetson Orin Nano Super failed to load because KV cache temp buffer required 10.7GB while only 6.5GB was available. Model weights alone were only 876MB.
  • Source: Source #21
  • Phase: Assessment
  • Target Audience: VLM memory management
  • Confidence: High
  • Related Dimension: Memory contention, VLM integration

Fact #15

  • Statement: Morphological skeletonization suffers from noise-induced boundary variations causing spurious skeletal branches. Recent methods (2025) use scale-space hierarchical simplification for controllable robustness.
  • Source: Source #31 (related search results)
  • Phase: Assessment
  • Target Audience: Path tracing robustness
  • Confidence: High
  • Related Dimension: Path tracing

Fact #16

  • Statement: GraphMorph (2025) operates at branch-level using Graph Decoder + SkeletonDijkstra, producing topology-aware centerline masks. Reduces false positives vs pixel-level segmentation approaches.
  • Source: Source #32
  • Phase: Assessment
  • Target Audience: Path extraction algorithms
  • Confidence: High
  • Related Dimension: Path tracing

Fact #17

  • Statement: Kalman filter + coordinate transformation in UAV gimbal systems eliminates attitude and mounting errors that PID controllers alone cannot compensate for during flight.
  • Source: Source #34
  • Phase: Assessment
  • Target Audience: Gimbal control algorithm
  • Confidence: High
  • Related Dimension: Gimbal control

Fact #18

  • Statement: Synthetic data generation for camouflage detection is a validated approach: GenCAMO (2026) uses scene graphs + generative models; CamouflageAnything (CVPR 2025) uses controlled out-painting. Both improve detection baselines.
  • Source: Sources #28, #29
  • Phase: Assessment
  • Target Audience: Training data strategy
  • Confidence: High
  • Related Dimension: Training data

Fact #19

  • Statement: Usable VRAM on Jetson Orin Nano Super is approximately 5.2GB after OS overhead (not the advertised 8GB). The 8GB is shared between CPU and GPU.
  • Source: Source #36
  • Phase: Assessment
  • Target Audience: Memory budget planning
  • Confidence: High
  • Related Dimension: Memory contention

Fact #20

  • Statement: FP8 quantization for Qwen2-VL-2B performs worse than FP16 on vLLM. INT8/W4A16 are the recommended quantization formats for 2B VLMs on constrained hardware.
  • Source: vLLM Issue #9992
  • Phase: Assessment
  • Target Audience: VLM quantization strategy
  • Confidence: High
  • Related Dimension: VLM integration