Back to Resources
    Guide
    CV
    Edge
    Jetson

    Edge Computer Vision to Production: PoC → Pilot → Scale

    A practical blueprint to move from a PoC to stable edge CV—covering datasets, accuracy gates, latency budgets, and operational review loops.

    14 min readJanuary 20, 2025Marcus Rodriguez, Principal Engineer, Computer Vision
    Edge computing devices processing real-time camera feeds in manufacturing facility

    Choosing sites and datasets that generalize

    PoCs often succeed on carefully selected test data but fail in production when conditions vary. To build models that generalize:

    Site selection strategy

    Choose pilot sites that represent your operational diversity:

  1. Varied lighting conditions (natural, artificial, mixed)
  2. Different camera angles and mounting heights
  3. Range of product/object variations
  4. Typical background clutter and occlusions
  5. Don't pick your "easiest" site for pilot. Pick representative sites that will stress-test your models.

    Dataset requirements

    Collect training data that covers:

  6. All expected object classes and variations
  7. Edge cases and failure modes
  8. Seasonal and temporal variations (if relevant)
  9. Multiple sites and camera positions
  10. Aim for 1,000-5,000 labeled examples per class minimum. More is better, but quality trumps quantity—ensure diverse, representative samples.

    Accuracy & latency gates (and how to measure)

    Set quantitative acceptance criteria before deployment. Don't settle for "it looks good"—measure precisely.

    Accuracy gates

    Define per-class thresholds:

  11. Precision: ≥95% (minimize false positives)
  12. Recall: ≥90% (minimize false negatives)
  13. F1 score: ≥92% (balanced measure)
  14. Measure on held-out test sets from production sites. Track confusion matrices to identify systematic errors.

    Latency budgets

    Define end-to-end latency requirements:

  15. Camera capture → inference → alert: <500ms for real-time use cases
  16. Batch processing: <1 hour for overnight jobs
  17. Model load time: <30 seconds for edge device startup
  18. Measure in production conditions, not just on development machines. Account for network latency, concurrent workloads, and thermal throttling on edge devices.

    | Pipeline Stage | Budget | Measurement |

    |---------------|--------|-------------|

    | Frame capture | 33ms (30fps) | Camera specs |

    | Preprocessing | 10ms | Profiler |

    | Inference | 100ms | Triton metrics |

    | Post-processing | 20ms | Profiler |

    | Alert dispatch | 50ms | Network monitor |

    | Total | 213ms | End-to-end test |

    Edge pipelines (DeepStream/Triton) with batching

    NVIDIA DeepStream and Triton Inference Server provide production-grade edge inference:

    Pipeline architecture

    1. Capture: RTSP streams from IP cameras

    2. Decode: Hardware-accelerated video decode (NVDEC)

    3. Batch: Accumulate frames from multiple streams

    4. Inference: Run batched inference on GPU (Triton)

    5. Track: Multi-object tracking across frames

    6. Alert: Detect events and dispatch to upstream systems

    Use DeepStream's gst-launch pipelines for low-latency streaming or custom Python/C++ applications for complex logic.

    Batching strategy

    Batch size trades off latency vs. throughput:

  19. Batch size 1: Lowest latency (~50ms), lower GPU utilization
  20. Batch size 8: Moderate latency (~120ms), high throughput
  21. Batch size 32: High latency (~300ms), maximum throughput
  22. Choose based on your use case. Real-time alerting needs small batches; overnight analysis can use large batches.

    Drift detection and retraining hooks

    Models degrade over time as conditions change. Detect drift and trigger retraining:

    Monitoring signals

    Track these metrics continuously:

  23. Confidence distribution: Falling average confidence indicates drift
  24. Prediction entropy: Rising entropy suggests uncertainty
  25. Human corrections: Increased override rate signals model mismatch
  26. Performance metrics: Declining accuracy on validation sets
  27. Set thresholds and alert when metrics cross into red zones.

    Retraining workflow

    1. Detect drift signal

    2. Sample recent edge cases (low confidence, human corrections)

    3. Label and add to training set

    4. Retrain model with augmented dataset

    5. Validate on test set (must meet original accuracy gates)

    6. Deploy to edge devices via OTA update

    Automate steps 1-2 and 6. Keep humans in the loop for steps 3-5 until you have high confidence in automated pipelines.

    Reviewer tooling and evidence packaging

    Edge CV isn't fully autonomous—humans review edge cases, validate alerts, and provide ground truth for retraining.

    Review UI requirements

    Build tooling that enables efficient review:

  28. Queue of flagged items (low confidence, alerts, samples)
  29. Side-by-side comparison (model prediction vs. ground truth)
  30. Quick annotation actions (approve, reject, correct)
  31. Keyboard shortcuts for power users
  32. Progress tracking and quotas
  33. Measure reviewer throughput (items/hour) and tune UI to maximize efficiency.

    Evidence packaging

    When CV detects events, package evidence for downstream consumers:

  34. Video clip: 5-10 seconds surrounding event
  35. Metadata: Timestamp, camera ID, confidence scores
  36. Annotations: Bounding boxes, class labels
  37. Context: Related events, historical patterns
  38. Export in standard formats (JSON + MP4) for integration with MES, ERP, or analyst tools.

    Scaling: health telemetry, upgrades, and costs

    Deploying to dozens or hundreds of edge devices requires operational discipline.

    Health telemetry

    Monitor every device:

  39. System: CPU, GPU, memory, disk, temperature
  40. Pipeline: Frame rate, inference latency, queue depth
  41. Model: Prediction counts, confidence distribution
  42. Network: Bandwidth usage, packet loss, latency
  43. Aggregate metrics in central dashboard. Alert on anomalies (thermal throttling, memory leaks, network issues).

    OTA upgrade strategy

    Deploy model and software updates safely:

    1. Canary: Deploy to 1-2 devices, monitor for 24 hours

    2. Pilot: Expand to 10% of fleet, monitor for 48 hours

    3. Rollout: Deploy to remaining devices in waves

    4. Rollback: Maintain previous version as fallback

    Use device management platforms (Balena, AWS IoT, custom) to orchestrate deployments.

    Cost model

    Edge CV costs include:

  44. Hardware: $500-$5,000 per device (Jetson Orin, industrial PCs)
  45. Cameras: $200-$1,000 per camera (IP cameras, lenses, mounts)
  46. Connectivity: $50-$200/month per site (network, VPN)
  47. Maintenance: 10-20% of hardware cost annually
  48. Factor in total cost of ownership over 3-5 year lifespan when comparing to cloud-based alternatives.

    Frequently asked questions

    Ready to get started?

    Let's discuss how these patterns apply to your deployment.