add clarification to research methodology by including a step for solution comparison and user consultation

This commit is contained in:
Oleksandr Bezdieniezhnykh
2026-03-17 18:43:57 +02:00
parent d764250f9a
commit b419e2c04a
35 changed files with 6030 additions and 0 deletions
@@ -0,0 +1,137 @@
# Fact Cards
## Fact #1
- **Statement**: Draft04 uses SuperPoint+LightGlue for VO (150-200ms/frame) while draft03 used XFeat (15ms/frame). This 10x speed regression was NOT listed in draft04's assessment findings — it appears to be an unintentional change.
- **Source**: [Source #3] XFeat paper, draft03 vs draft04 comparison
- **Phase**: Assessment
- **Target Audience**: GPS-denied UAV system
- **Confidence**: ✅ High
- **Related Dimension**: VO Matcher Selection
## Fact #2
- **Statement**: XFeat outperforms SuperPoint on Megadepth: AUC@10° 65.4 vs 50.1, with more inliers (892 vs 495). For high-overlap consecutive frames (60-80%), XFeat quality is sufficient.
- **Source**: [Source #3] XFeat paper Table 1
- **Phase**: Assessment
- **Target Audience**: UAV VO pipeline
- **Confidence**: ✅ High
- **Related Dimension**: VO Matcher Quality
## Fact #3
- **Statement**: SatLoc-Fusion (2025) validates XFeat for UAV VO in a similar setup: nadir camera, 100-300m altitude, <15m error, >90% trajectory coverage, >2Hz on 6 TFLOPS edge hardware.
- **Source**: [Source #4] SatLoc-Fusion
- **Phase**: Assessment
- **Target Audience**: UAV VO pipeline
- **Confidence**: ✅ High
- **Related Dimension**: VO Matcher Selection
## Fact #4
- **Statement**: LiteSAM's 77.3% Hard hit rate is on the authors' SELF-MADE dataset (Harbin/Qiqihar, 100-500m altitude), NOT UAV-VisLoc. On UAV-VisLoc Hard, LiteSAM achieves 61.65% hit rate with RMSE@30=17.86m.
- **Source**: [Source #2] LiteSAM paper
- **Phase**: Assessment
- **Target Audience**: Satellite-aerial matching
- **Confidence**: ✅ High
- **Related Dimension**: Satellite Matching Accuracy
## Fact #5
- **Statement**: LiteSAM GitHub repo has 5 stars, 0 forks, 4 commits, no releases, no license, no issues. Single maintainer. Very low community adoption.
- **Source**: [Source #1] LiteSAM GitHub
- **Phase**: Assessment
- **Target Audience**: Production readiness evaluation
- **Confidence**: ✅ High
- **Related Dimension**: LiteSAM Maturity
## Fact #6
- **Statement**: LiteSAM weights are hosted on Google Drive as a single .ckpt file (mloftr.ckpt) with no checksum, no mirror, no alternative download source.
- **Source**: [Source #1] LiteSAM GitHub
- **Phase**: Assessment
- **Target Audience**: Supply chain security
- **Confidence**: ✅ High
- **Related Dimension**: Security
## Fact #7
- **Statement**: CVE-2025-32434 allows RCE even with weights_only=True in torch.load() (PyTorch ≤2.5.1). CVE-2026-24747 shows memory corruption in the weights_only unpickler (fixed in PyTorch ≥2.10.0).
- **Source**: [Source #5, #6] NVD
- **Phase**: Assessment
- **Target Audience**: All PyTorch-based systems
- **Confidence**: ✅ High
- **Related Dimension**: Security
## Fact #8
- **Statement**: EfficientLoFTR (LiteSAM's base) has 964 stars, HuggingFace integration, CVPR 2024 publication. 15.05M params. Much more mature and proven than LiteSAM.
- **Source**: [Source #10] EfficientLoFTR GitHub
- **Phase**: Assessment
- **Target Audience**: Satellite-aerial matching fallback
- **Confidence**: ✅ High
- **Related Dimension**: LiteSAM Maturity
## Fact #9
- **Statement**: LiteSAM has no ONNX or TensorRT export path. EfficientLoFTR also lacks official ONNX support. Custom conversion work would be required.
- **Source**: [Source #1, #10] GitHub repos
- **Phase**: Assessment
- **Target Audience**: Performance optimization
- **Confidence**: ✅ High
- **Related Dimension**: Performance
## Fact #10
- **Statement**: LiteSAM was benchmarked on RTX 3090. Performance on RTX 2060 is estimated at ~140-210ms but not measured. RTX 2060 has ~22% of RTX 3090 FP32 throughput.
- **Source**: [Source #2] LiteSAM paper + GPU specs
- **Phase**: Assessment
- **Target Audience**: RTX 2060 deployment
- **Confidence**: ⚠️ Medium (extrapolated)
- **Related Dimension**: Performance
## Fact #11
- **Statement**: DINOv2 ViT-S/14 uses ~300MB VRAM; ViT-B/14 uses ~900-1100MB VRAM (3-4x more). ViT-B provides +2.54pp recall improvement over ViT-S.
- **Source**: [Source #7] Nature Scientific Reports
- **Phase**: Assessment
- **Target Audience**: VRAM budget
- **Confidence**: ⚠️ Medium (extrapolated from classification task)
- **Related Dimension**: DINOv2 Model Selection
## Fact #12
- **Statement**: Google Maps intentionally does not publish recent satellite imagery for conflict areas in Ukraine. Imagery is typically 1-3 years old. Google stated: "These satellite images were taken more than a year ago."
- **Source**: [Source #8] AIN.ua, Google statements
- **Phase**: Assessment
- **Target Audience**: Satellite imagery freshness
- **Confidence**: ✅ High
- **Related Dimension**: Satellite Imagery Quality
## Fact #13
- **Statement**: GTSAM iSAM2.update() can throw IndeterminantLinearSystemException with certain factor configurations. Long chains (3000 frames) should work via Bayes tree structure but need error handling.
- **Source**: [Source #9] GTSAM GitHub #561
- **Phase**: Assessment
- **Target Audience**: Factor graph robustness
- **Confidence**: ✅ High
- **Related Dimension**: GTSAM Robustness
## Fact #14
- **Statement**: No independent reproduction of LiteSAM results exists. Search results often confuse LiteSAM (feature matcher) with Lite-SAM (ECCV 2024 segmentation model).
- **Source**: [Source #1] Research verification
- **Phase**: Assessment
- **Target Audience**: Production readiness
- **Confidence**: ✅ High
- **Related Dimension**: LiteSAM Maturity
## Fact #15
- **Statement**: With XFeat for VO (~200MB VRAM) instead of SuperPoint+LightGlue (~900MB), peak VRAM drops from ~1.6GB to ~900MB (XFeat 200 + DINOv2 300 + LiteSAM 400).
- **Source**: Calculated from Sources #1, #3, #7
- **Phase**: Assessment
- **Target Audience**: VRAM budget
- **Confidence**: ⚠️ Medium (estimated)
- **Related Dimension**: VRAM Budget
## Fact #16
- **Statement**: Maxar restored satellite imagery access for Ukraine in March 2025 (was suspended). Commercial, paid-only. 31-50cm resolution (WorldView, GeoEye).
- **Source**: [Source #12] Defense Express
- **Phase**: Assessment
- **Target Audience**: Alternative satellite providers
- **Confidence**: ✅ High
- **Related Dimension**: Satellite Imagery Quality
## Fact #17
- **Statement**: Tracasa offers free super-resolved Sentinel-2 imagery for Ukraine at 2.5m resolution (500,000 km²). Deep learning upscale from 10m. Could serve as emergency fallback but resolution is insufficient for primary matching.
- **Source**: [Source #11] Tracasa
- **Phase**: Assessment
- **Target Audience**: Alternative satellite sources
- **Confidence**: ⚠️ Medium (2.5m resolution vs required 0.3-0.5m)
- **Related Dimension**: Satellite Imagery Quality