Large retail environments face a persistent challenge. Inventory systems know what should be on the shelf, but not what is actually there. Sales data, replenishment events, and periodic audits provide indirect signals, yet physical shelves change continuously throughout the day. This gap leads to out of stock events, phantom inventory, and heavy reliance on manual audits.
To address this, a large retail organization successfully deployed an edge AI driven perpetual inventory system. The platform was architected and built by Jitender Jain, a software engineer working on applied AI and distributed systems for stores and distribution centers. The solution combines computer vision, IoT, and cloud scale data pipelines to process unstructured imagery. Rather than treating inventory as a batch reconciliation problem, the system treats it as a continuously evolving signal derived directly from shelf reality.
The core architectural decision was to push intelligence closer to where data is generated. Shelf facing cameras capture images at regular intervals, but raw video is never streamed to the cloud. Instead, lightweight inference modules run on in store edge devices with strict compute and memory limits. These modules filter frames, enforce privacy constraints, and extract only the metadata required to reason about inventory state.
Privacy enforcement is handled at the edge itself. Any frame containing customers or associates is discarded before it can leave the store. This simplifies compliance and reduces downstream risk, while also cutting bandwidth usage dramatically. Only structured events such as product presence, confidence scores, and shelf identifiers are transmitted upstream.
On the cloud side, the system relies on event driven ingestion and streaming pipelines. Inventory is modeled as state over time rather than a single snapshot. Signals from the edge are aggregated, reconciled, and scored to converge on an accurate shelf level view. This probabilistic approach avoids over reacting to noise and allows the system to remain stable even when inputs are imperfect.
Machine learning models are trained centrally using historical data and curated ground truth, then deployed incrementally back to edge devices through configuration driven rollouts. This separation between training and inference enables continuous improvement without disrupting store operations.
At scale, the system has proven effective. It improves shelf inventory accuracy, reduces manual audit effort, and enables faster detection of availability issues. More importantly, it operates continuously and autonomously, which is essential in environments spanning thousands of locations.
The key takeaway is that successful applied AI systems in retail are less about model sophistication and more about system design. Edge constraints, privacy requirements, and operational realities shape every architectural decision. Treating inventory as a real time signal rather than a batch calculation unlocks a fundamentally different level of visibility.
Below is a simplified view of the architecture
A small example of what edge level filtering logic might look like is shown below. The intent is not accuracy but fast rejection under tight constraints.
def process_frame(frame):
if contains_human(frame):
return None
products = detect_products(frame)
if not products:
return None
return {
"shelf_id": frame.shelf_id,
"products": products,
"timestamp": frame.timestamp
}
Perpetual inventory at retail scale is ultimately a distributed systems problem. When edge intelligence, streaming data, and ML lifecycle management work together, the result is not just better data, but a more resilient and operationally viable platform.

Top comments (0)