All case studies
AutomationManufacturingApril 10, 2026

AI Fault Tracing Agent Cuts Conveyor Downtime from 40 Minutes to 28 Seconds

Result

40 min fault traces resolved in 28 seconds

AI Fault Tracing Agent Cuts Conveyor Downtime from 40 Minutes to 28 Seconds

A mid-size manufacturer was losing production time to manual fault tracing across disconnected SCADA, PLC, and network layers. We built an AI agent that traces faults through the full control stack in under a minute.

A mid-size discrete manufacturer running 14 production lines was averaging 28 minutes of downtime per fault event — not because the fault was hard to fix, but because tracing the root cause through disconnected system layers was. Controls engineers were manually bridging Ignition SCADA alarms, OPC-UA tag mappings, Rockwell Studio 5000 ladder logic, and Cisco IE network diagnostics by hand. Every minute spent on that trace was a minute the line sat idle.

The Real Cost of Manual Fault Tracing

At this facility, unplanned downtime cost approximately $4,200 per hour across all lines — a figure derived from their own OEE tracking, accounting for lost throughput, labor standing by, and downstream scheduling impact. With 40 fault events per week averaging 28 minutes each, the facility was absorbing roughly $78,000 per week in downtime cost attributable not to the faults themselves, but to the time it took to diagnose them.

The controls team was four engineers responsible for 14 lines. They were experienced — two had Rockwell-certified TechConnect training, one had an ISA certification in functional safety. The problem wasn't competence. The problem was architecture: the tools they used to diagnose faults were designed independently, maintained independently, and provided no unified view of a fault as it propagated across system layers.

When a line goes down, a controls engineer is working against two pressures simultaneously: the production manager asking for an ETA and the fault itself, which may be masking a secondary issue. The fastest engineers at this facility had developed mental shortcuts — pattern matching from experience to guess the most likely root cause without fully tracing the dependency chain. That worked most of the time. When it didn't, it added another 20 minutes to the fault event.

The Diagnostic Problem in Industrial Control Systems

Most industrial environments are built in layers that work well individually but communicate poorly across boundaries when something breaks. Alarms surface in Ignition SCADA. The tags behind those alarms map to addresses in an OPC-UA server. The logic driving those tags lives inside Rockwell Studio 5000 routines — ladder logic, Function Block Diagrams, and Add-On Instructions (AOIs) layered several levels deep. The hardware state that's blocking the logic sits somewhere on the Cisco IE switch network or in a DeviceNet or EtherNet/IP I/O module. No single interface shows all of it at once.

When a fault fires, a controls engineer typically has to:

  • Locate the alarm source in Ignition and determine whether it's a process alarm or a device fault
  • Identify the underlying OPC-UA tag path and map it back to a PLC data element
  • Open Studio 5000, navigate to the relevant routine, and find the rung or FBD block that drives the tag
  • Trace upstream permissives — which may span multiple routines, multiple AOIs, or even multiple PLCs if the line uses a distributed architecture
  • Check DeviceNet or EtherNet/IP device status for hardware faults on the network
  • Verify I/O module status, CIP connection health, and any pending inhibits or faults on the safety system if applicable

For a straightforward fault — a single failed proximity sensor on a conveyor — that process takes 15 to 20 minutes for an experienced engineer. For a fault with multiple contributing conditions or one that sits inside a deeply nested AOI structure with indirect addressing, it can take over an hour. At 40 events per week, the cumulative impact is enormous.

What We Built

We built an AI reasoning agent with structured, read-only access to three data sources: the Ignition 8.1 SCADA project (alarm configuration, tag paths, OPC-UA device mappings), the Rockwell Studio 5000 L5X project exports for all 14 PLCs (full logic, routine cross-references, AOI definitions, and tag cross-reference tables), and live CIP and EtherNet/IP device diagnostics pulled through the existing Rockwell FactoryTalk Diagnostics infrastructure.

The agent's job is diagnostic traversal. When a fault fires, an operator or engineer enters a plain-language description of the symptom — "Line 4 E-stop reset isn't latching," "Conveyor 7 jog mode not responding," "Filler head 3 fault on every cycle start" — and the agent traverses the full dependency graph from that symptom to its root cause.

The system keeps the intelligence layer inside the plant network boundary. The agent runs on a hardened edge server on the OT network segment. No PLC configuration data, no tag data, and no production records leave the facility. The interface is a terminal accessible from the main control room and from any engineering workstation on the plant floor network. There is no cloud dependency in the diagnostic path.

The agent is read-only at the control system level by design and by architecture. It has no write access to any PLC, no ability to issue commands through Ignition, and no connection to any actuator or drive system. The only output it produces is a diagnostic report.

The Demonstrating Event

During the commissioning period, a conveyor on Line 7 faulted overnight and wouldn't clear on operator reset. The floor operators attempted the SCADA reset sequence six times across two shifts — the standard procedure for that fault code — without success. By the time the day-shift controls engineer arrived, the line had been down for nearly four hours. The preliminary assessment was a wiring issue somewhere in the safety circuit, which could mean pulling conduit.

The agent was given a single input: the alarm text from the Ignition event log.

It traced from the alarm through Ignition's alarm journal to the OPC-UA tag, followed the tag into the Studio 5000 ladder logic for PLC-07, located the rung holding the reset output in a latched-off state, traced the blocking permissive upstream through two nested AOI layers — a conveyor zone controller AOI and a safety interlock AOI that sat one level above it — and identified a GuardLogix Safety I/O module (1791DS-IB12) with a dropped CIP connection as the upstream condition preventing the reset.

It also identified two additional conveyors on Lines 8 and 9 that were running on borrowed time: the same Safety I/O module was part of their interlock permissive chain, and both were one CIP timeout event away from faulting under load.

The engineer confirmed a red NET1 fault LED on the suspect module — one confirming observation. The agent returned the exact root cause and the corrective action sequence.

Total diagnostic trace time: 28 seconds. Estimated manual trace time for the same path, given the nested AOI structure: 35 minutes.

The root cause was a loose fiber SFP connection at the module backplane — a physical issue that required a physical fix. The fix took four minutes. The two downstream conveyors were inspected proactively before the shift ended, preventing two additional fault events.

Results

Three months after deployment, the facility's average fault-to-resolution time dropped from 28 minutes to under 6 minutes. The agent handles the diagnostic trace; the engineer handles the physical fix. For the three most common fault categories at this facility — I/O hardware faults, safety system permissive failures, and PLC logic-level inhibits — the agent's diagnostic accuracy against confirmed root cause is 94%.

  • Average fault trace time: 28 min baseline to under 60 seconds
  • Mean time to resolution (fault to line restart): 28 min to 6 min
  • Estimated weekly production time recovered: 14+ hours across all lines
  • Estimated weekly downtime cost reduction: ~$40,000 based on facility's own OEE figures
  • Secondary faults identified proactively before failure: 11 in first 90 days
  • Controls engineers redirected to capital project work and preventive maintenance: ~8 hrs/week recovered

The client is currently scoping the system for their second facility, a similar discrete manufacturing operation with 18 lines and a Siemens TIA Portal / WinCC SCADA architecture.

How the Agent Handles Complexity

The most valuable cases are not straightforward single-sensor faults — an experienced engineer can find those quickly with or without the agent. The high-value cases are faults with multi-layer contributing conditions: a process alarm that's actually caused by a permissive buried in an AOI that was originally written for a different line and copied without full review, or an intermittent fault that only fires under specific timing conditions and has no clear alarm trail.

In those cases, the agent's advantage is completeness. It doesn't take shortcuts or pattern-match from experience. It traverses the full dependency graph every time, which means it finds the second and third contributing conditions that an experienced engineer might not have checked because the first one looked sufficient.

The agent also doesn't forget context. When a fault on Line 7 was traced to a Safety I/O module, the agent cross-referenced that module's tag ownership against all 14 PLCs to identify every affected permissive chain — not just the one associated with the active alarm. That cross-facility awareness requires a level of systematic indexing that a human engineer simply can't replicate in real time.

The OT Security Posture

Industrial AI tools raise legitimate OT security questions. The architecture here was designed with those questions in mind from the start, not as an afterthought.

The agent runs on an isolated edge server inside the plant's OT network with no external connectivity. PLC access is read-only, using a dedicated read-only tag server credential scoped to the diagnostic tag set only — no write paths exist in the configuration. Studio 5000 L5X exports are loaded as static files rather than live PLC connections, so there's no persistent socket connection to any controller. The Cisco IE network diagnostic access uses SNMP read-only community strings.

The security review was conducted with the client's IT/OT security team before deployment. The architecture passed their review without requiring exemptions or compensating controls.

Why This Works

Industrial fault tracing is not primarily a reasoning problem — it's a retrieval and graph traversal problem. The information needed to resolve 80 percent of faults already exists inside the control system. The bottleneck is access speed: getting the right person to the right place in the right system in the right order, under time pressure, while the production manager is asking for an ETA.

An agent with structured access to the full control system graph doesn't need to search. It traverses. The answer that takes an engineer 20 minutes to locate manually takes the agent seconds to retrieve.

The remaining 20 percent of faults — novel failure modes, hardware degradation with ambiguous diagnostic signatures, environmental causes — still require an experienced controls engineer. The agent doesn't replace that judgment. It eliminates the time that engineer would otherwise have spent on everything that didn't need their judgment, so they arrive at the hard part faster.

Ready to see results like these?

Tell us what you're working on. We'll scope it and tell you how we'd approach it.

Start a project