← Case Studies

A security camera feed from a rainy day at a fenced industrial facility, showing the T-DA system's object detection capabilities. The feed, labeled "UK - FACILITY UGV CAM 04" with the date "12/01/2026" and time "05:39:15 PM," displays four detected objects with bounding boxes and confidence scores: a security guard in a yellow high-visibility jacket near the fence (green box, "SECURITY GUARD (95%)"), a person in a dark hoodie walking behind the fence (red box, "PERSON (88%)"), a multi-rotor drone flying above the fence (blue box, "DRONE (98%)"), and a white van on the wet road outside the fence (yellow box, "VEHICLE (99%)").

Project T-DA: Securing the Shadows with Temporal-Aware AI

Sector: Critical Infrastructure Security
Origin: Decision Lab’s AI Innovation Lab 

Executive Summary

  • Delivers 24/7 autonomous situational awareness on standard edge hardware, removing the need for costly server infrastructure refits.

In the high-stakes world of critical infrastructure protection, the gap between a routine patrol and a security breach is often measured in seconds. Yet, traditional surveillance systems are manipulated by intelligent behaviour. Standard algorithms rely on invariant detections (classification of known objects) and often lack the ability to determine action and intent. While these algorithms can effectively detect and classify inanimate objects that might be inherently harmful or broken, they falter when detecting complex actions such as abnormal or harmful patterns of behaviour.  

Emerging from our AI Innovation Lab, the Threat-Detection using Autoencoders (T-DA) programme was designed to close this gap. By combining state-of-the-art Computer Vision with novel temporal awareness, we delivered an unsupervised Deep Learning solution capable of learning the ‘pattern of life’ aspects of behaviour. The result is a system that doesn’t just see movement but understands context—differentiating between a scheduled patrol and an unscheduled intrusion without requiring massive, labelled datasets

The Challenge: The Signal in the Noise 

Our client, responsible for the security of high-sensitivity sites, faced a strategic pain point: operational blindness caused by data overload. Their existing surveillance infrastructure relied on simple motion detection and rule-based triggers. 

These legacy systems suffered from two critical failures relating to Pattern of Life (PoL): 

Pattern of Life anomalies diagram.Action: suspicious, harmful
Time: unusual, after hours
Object: weapon, mask
Role: insider, impersonator

High False Positive Rate: Innocent environmental changes (e.g., wind-blown debris) or routine scheduled events triggered constant alarms, leading to operator fatigue and desensitisation. 

Contextual Blindness: The systems could not distinguish between visually similar but contextually different events. A guard walking a perimeter at 14:00 is routine; a person walking the same path at 03:00 could be a threat. Standard models saw only ‘person walking’. 

The client required a solution that could autonomously detect anomalies in real-time, operate on resource-constrained edge devices, and—crucially—learn what normal looks like without needing thousands of manually labelled “threat” examples. 

The Solution: Temporal-Aware Deep Learning

Decision Lab deployed a cutting-edge unsupervised anomaly detection pipeline that fundamentally changes how machines perceive security footage.

Traditional systems require training on thousands of examples of threats (which are rare and varied). Instead, we taught the model what normality looks like. By learning the standard pattern of life, the system can autonomously flag any event that deviates—whether it is a known threat type or an entirely new anomaly.

A flowchart diagram illustrating an anomaly detection pipeline using an autoencoder. The flow moves from left to right: Input: "Raw sensor datapoints" and "Temporal factors" feed into a "Datapoint fusion" block. Processing: This feeds into an "Autoencoder" block, which splits into two paths: a training loop and an inference path. Training Pipeline: A lower section labeled "Autocoder training pipeline" details a cycle: "Encoder" points to "Latent representation" (highlighted in green), which flows to "Decoder," then "Loss (MSE/Binary Cross-entropy)," "Back propagation," "Update weights," and finally loops back to "Encoder." Output: The Autoencoder also connects to an "Inference" block (in blue). "Inference" points to an "Anomaly detection (flag)" block. Model State: An arrow points from "Inference" up to a "Trained model" block, which then points down to the "Anomaly detection (flag)" block.

1 . The Core Architecture

To ensure our solution remained modal-agnostic, we experimented with various encoder models, ranging from standard LSTMs to Gaussian Mixture Models. For the Proof of Concept (PoC), we implemented a Vision Transformer (ViT)-based Variational Autoencoder (VAE).

  • The Encoder (ViT): Unlike standard CNNs that look at localised pixels, the ViT uses self-attention mechanisms to capture global contextual information from video frames.
  • The Decoder: This component attempts to reconstruct the video frames from the encoder’s summary.
  • The Trigger: If the model cannot accurately reconstruct a scene (resulting in a high reconstruction error), it indicates the event is not in its learned database of normal behaviours, instantly triggering an anomaly alert.

2. Innovation: Temporal Integration

Standard computer vision models are time-blind; they see a person walking but do not know if it is 14:00 (routine) or 03:00 (suspicious). To solve this, we engineered a novel Cyclic Time Encoding mechanism.

  • Cyclic Encoding: We encoded timestamps using sine and cosine functions. This captures the periodic nature of time (24-hour cycles) more effectively than linear numbers.
  • Contextual Conditioning: This time vector modulates the model’s latent space. effectively teaching the AI that Activity A is normal at Time X, but anomalous at Time Y.

This approach yielded two critical capabilities:

  1. Temporal Anchoring: We successfully introduced a temporal factor—contextual metadata that anchors the model in time, rather than relying solely on visual pixel data.
  2. Scalable Context: While this PoC used timestamps, the architecture can ingest any form of metadata. The model can be conditioned on geographical data (weather, pressure), or solution-specific constraints (security clearance levels, job titles), making T-DA highly adaptable across defence, supply chain, and rail verticals.

3. Edge Deployment (SWaP Optimised)

Meeting the strict requirements for defence operations, we optimised the model using FP16 (16-bit floating point) precision.

  • Size Reduction: This compressed the model size by 50%.
  • Performance: The system runs efficiently on resource-constrained edge devices (e.g., drones, remote sentries) without sacrificing detection accuracy.
  • Security: Data is processed locally, reducing bandwidth requirements and closing the attack surface associated with cloud transmission.

Results & Impact 

The T-DA project successfully demonstrated that autonomous systems can reduce the cognitive load on human operators while enhancing threat detection. 

  • High Precision: The ViT-based VAE achieved an ROC AUC of 0.855 on general visual anomaly detection, validating the unsupervised approach. 
  • Operational Efficiency: The move to FP16 precision resulted in a 50% reduction in model size and significant runtime memory savings, enabling deployment on standard edge hardware rather than expensive server racks. 

Reduced Fatigue: By automating the detection of contextually specific anomalies, the system significantly reduced the time security personnel spent reviewing false alarms, allowing them to focus on genuine threats.

FOCUS: AI TRiSM (Trust, Risk, and Security Management) 

As part of Decision Lab’s commitment to Responsible AI (read our full series here), the T-DA project was developed in strict alignment with the AI TRiSM framework. In high-stakes defence environments, an AI model must be as trustworthy as the officers using it. 

1. Trust: Explainability beyond the Black Box 

A security operator cannot act on an alert they don’t understand. We moved beyond simple “anomaly scores” by integrating Explainable AI (XAI) techniques. 

  • Heatmaps: The system provides real-time reconstruction error heatmaps, visually highlighting exactly where in the frame the anomaly is occurring (e.g., highlighting a specific backpack or unauthorised vehicle). 
  • Contextual Logic: We explored the integration of LLMs to generate natural language explanations, translating complex vector data into clear summaries: ‘Unusual activity detected: Person running at 02:45 AM (high deviation from routine).’ 

2. Risk: Proactive Reliability 

Unsupervised models can drift over time if the environment changes. We mitigated this risk through: 

  • Synthetic Anomaly Injection: To rigorously test the system before deployment, we developed a methodology to inject synthetic temporal anomalies into the data, ensuring the model could catch threats that hadn’t happened yet. 
  • Bias Audits: We conducted formal audits of the training data to ensure the normal baseline didn’t inherently bias the model against specific demographics or harmless behaviours. 

3. Security: ModelOps & Data Integrity 

Security is paramount not just in the physical site, but in the digital pipeline. 

  • Data Protection: We implemented encrypted channels for all video streams and strict access controls for training data. 
  • ModelOps: A robust lifecycle management framework was established, including version control for model weights and automated drift detection to trigger retraining. This ensures the model adapts to new patterns of life securely and transparently. 

Learn more in our AI TRiSM blog series.

Conclusion 

The T-DA project illustrates the power of the Decision Lab Innovation Lab to translate theoretical AI advances into robust, deployable security solutions. By treating time as a critical feature of reality, we moved surveillance from reactive monitoring to proactive threat detection. 

Project T-DA Key Facts:

  • Developer: Decision Lab
  • Primary Tech: Vision Transformer (ViT) & Variational Autoencoder (VAE)
  • Innovation: Cyclic Time Encoding (Temporal Awareness)
  • Use Case: Unsupervised anomaly detection for critical infrastructure.
  • Performance: 0.855 ROC AUC with 50% model compression via FP16.

Would you like to explore how Decision Lab can streamline your operations? Contact our Innovation Team today