CASE STUDY: Optical Target Tracker

Portable Camera-Based Object Tracking: Mastering Simulation-Trained AI at the Edge

Building a portable optical tracking system using digital twin simulation, custom machine learning models, and edge AI deployment.

Systematic Consulting Group Logo

Camera-Based Aerial Object Tracking

Services
Hardware Engineering, Software Development,AI Development

Project Timeline
12 months

Tech Stack
NVIDIA Omniverse, Custom ML Training Pipeline, Embedded Linux, DJI Gimbal Integration, Camera-Based Tracking, Compact Edge Computing Platform

The Challenge

Drones have become an asymmetric threat across defense, security, and public safety applications. Traditional counter-drone systems rely on radar technology, which provides excellent range but requires significant power and infrastructure. We knew there had to be a better solution — one that could be deployed quickly and with minimal resources. 

“We could have waited for a client to fund this exploration, but that would have made us followers, not leaders. By investing our own resources into ambitious R&D, we master technologies before our clients need them—which means we can deliver faster, more innovative solutions when opportunities arise.”

Jonathan Pardeck, President, Systematic Consulting Group

Define & Discovery

Camera-Based Tracking

Camera-based optical tracking offers a compelling alternative to radar-based technology. It dramatically lowers power consumption, offers smaller form factors, and leverages rapidly advancing computer vision capabilities.

However, camera-based tracking at distance requires solving complex technical challenges: detecting 1-2 meter wide targets at 2+ kilometers, maintaining focus across varying zoom levels while tracking moving objects, deploying AI on resource-constrained hardware consuming less than 20W, and achieving this in a man-portable package under 50 pounds with 24+ hour unattended runtime. The system also needed rapid deployment (<10 minutes setup) and integration with established Command & Control platforms including TAK and Anduril Lattice.

Architect & Design

Simulation-First Development Strategy

Traditional machine learning development for computer vision follows a resource-intensive cycle: capture thousands of real-world images, manually annotate each frame with bounding boxes, train models, test in field conditions, and iterate. For drone tracking specifically, this means coordinating drone flights, camera equipment, various weather conditions, and hours of manual labeling.

We identified an alternative approach: leveraging digital twin simulation to generate synthetic training data. Recent advances in tools like NVIDIA Omniverse have made it possible to create photorealistic environments where virtual drones fly realistic patterns while virtual cameras capture training data with perfect automatic labeling.

Synthetic Training Data Generation

We created a virtual environment where drones fly realistic flight patterns while a virtual camera captures images. Because the simulation knows exactly what’s in each frame, it automatically generates perfect training labels—eliminating weeks of manual annotation work.

Automated Dataset Creation

Instead of spending months capturing and labeling thousands of drone images, we generated comprehensive training datasets in days. We could systematically vary lighting conditions, backgrounds, drone types, distances, and angles—creating more diverse training data than would be practical to capture in the real world.

Hardware Selection

We paired this software innovation with carefully selected commercial components: a DJI RS2 gimbal providing smooth pan-tilt operation with CAN bus control, a Blackmagic 4K camera paired with a Panasonic 45-175mm varifocal zoom lens providing the optical range needed to detect Group 1-3 UAS, and an NVIDIA Jetson Orin NX edge computing platform (executing 21 FP16 TOPS at just 20W) in a deck-of-cards form factor. 

Total prototype cost at commercial rates: approximately $3,550—demonstrating that sophisticated edge AI systems don’t require massive capital investment.

The goal wasn’t to create a polished commercial product—it was to validate the simulation-based development approach and understand the optical physics that govern camera-based tracking at distance.

Develop & Iterate

From Simulation to Reality

The development phase validated both our simulation approach and revealed critical lessons about real-world implementation.

Model Training Success

Our simulation-based training pipeline produced a model that successfully transitioned from virtual environments to real-world drone tracking. This proved the viability of synthetic data for this class of computer vision problems—a finding that’s directly applicable to our client work in sports technology, robotics, and industrial automation.

Tracking Performance and Technical Constraints

Field testing validated reliable detection and tracking of DJI M30T (Group 1 UAS) drones at 100-150 meters. While this fell well short of our 2+ kilometer target range, it revealed critical insights about the practical limits of commercial optical components. Three specific COTS hardware limitations impacted extended range performance:

  1. Autofocus quality and speed – The varifocal lens behavior required custom focus calibration across zoom levels to achieve parfocal-like performance
  2. Low-frequency joint position updates – 660ms command latency and 2Hz position updates from the gimbal API limited tracking responsiveness during rapid target maneuvers
  3. Detection model reliability – Unfocused targets at extended distances reduced AI detection confidence, even with our simulation-trained YOLO 11 model

 

Despite these range limitations, the system successfully achieved its integration objectives—publishing real-time tracks with geospatial coordinates (latitude, longitude, altitude) and thumbnail images to both Anduril Lattice and TAK/ATAK Command & Control platforms.

Edge AI Implementation

We successfully deployed the trained model to run locally on compact embedded hardware, demonstrating that sophisticated AI doesn’t require cloud connectivity or powerful GPUs. This edge-first architecture is increasingly critical for applications requiring low latency, privacy preservation, or operation in network-constrained environments.

Simulation-Trained AI Model Performance

The YOLO 11-based detection model, trained entirely on NVIDIA Replicator synthetic data, exhibited qualitatively strong real-world performance despite never seeing actual drone footage during training. This validated the simulation-first approach for rapid model development—generating comprehensive, automatically-annotated training datasets in days rather than the months typically required for field data collection and manual annotation.

The model architecture supported hot-swappable capabilities, allowing the system to be re-tasked on the fly to track different object types. At SOFWEEK 2025, we demonstrated this flexibility by simultaneously tracking both boats and people in Tampa Bay using the same hardware platform with dynamically loaded detection models.

The Results

The twelve-month project delivered more than a working prototype. It built organizational capabilities that benefit every subsequent project.

Simulation Infrastructure

We now have proven workflows for using digital twins to accelerate machine learning development. This infrastructure has already been deployed for commercial clients in sports technology, where we’re using similar techniques to track objects in complex visual environments.

Optical System Design Knowledge

We now know what’s possible with the physics of long-range camera systems and can prevent expensive mistakes when designing vision systems for clients and accelerate specification of appropriate components for any application.

Rapid Prototyping Methodology

The project validated our ability to move from concept to functional prototype quickly by leveraging simulation, commercial components, and agile development practices. 

Ready to Push Technical Boundaries with Your Product?

We’re ready to help you bring your vision to life. Request a consult with our product engineers.