Sensor Fusion for Indoor Localization and Positioning Systems

Sensor fusion for indoor localization combines data from multiple heterogeneous sensors to determine the position and orientation of a device or person within a building or enclosed structure — an environment where GPS signals are typically attenuated to the point of unusability. The discipline sits at the intersection of estimation theory, signal processing, and embedded systems engineering. Achieving sub-meter accuracy indoors requires resolving conflicting sensor outputs, compensating for multipath interference, and managing computational constraints in real time. This page describes the technical structure, operational scenarios, and classification boundaries that define this application domain within the broader sensor fusion landscape.

Definition and Scope

Indoor localization using sensor fusion refers to the continuous estimation of a mobile entity's position — expressed as 2D coordinates, 3D coordinates, or a full six-degree-of-freedom pose — inside environments where external reference signals are unavailable or insufficient. The scope includes pedestrian tracking in hospitals, asset tracking in warehouses, robot navigation in factories, and first-responder positioning in buildings.

The National Institute of Standards and Technology (NIST) has formally addressed this problem domain through its Public Safety Communications Research (PSCR) program, which sponsors annual indoor localization competitions measuring positioning accuracy in meters of error. NIST defines accuracy benchmarks for first-responder use cases, with target thresholds at or below 1 meter in 90% of floor-area coverage.

The sensor modalities involved fall into four primary categories:

  1. Inertial sensors — accelerometers and gyroscopes packaged as inertial measurement units (IMUs) provide dead-reckoning estimates but accumulate drift error over time
  2. Radio-frequency signals — Wi-Fi, Bluetooth Low Energy (BLE), and ultra-wideband (UWB) beacons provide range or fingerprint-based position estimates subject to multipath distortion
  3. Magnetic field sensors — magnetometers detect anomalies in the building's magnetic field to produce a structural fingerprint
  4. Optical and depth sensors — cameras and LiDAR provide feature-rich environmental maps but impose significant computational load

Each modality carries a distinct noise profile and failure mode, which is precisely the condition under which fusion algorithms outperform any single-sensor approach.

How It Works

The fusion pipeline for indoor localization typically operates in three phases: measurement collection, state estimation, and map anchoring.

Measurement collection aggregates raw readings from all active sensors at synchronized timestamps. Temporal alignment is non-trivial; IMUs may sample at 200 Hz while BLE RSSI scans update at 1–5 Hz, requiring interpolation or event-driven buffering.

State estimation applies a probabilistic filter to merge these measurements into a position estimate. The Extended Kalman Filter (EKF) is the industry standard for systems with nonlinear motion models, such as pedestrian locomotion. For environments with highly non-Gaussian noise or multimodal position hypotheses, particle filters offer superior robustness at higher computational cost. Bayesian fusion frameworks provide the theoretical envelope within which both approaches operate, treating position as a probability distribution rather than a point estimate.

Map anchoring corrects accumulated drift by matching observed features — Wi-Fi fingerprints, magnetic anomalies, or visual landmarks — against a pre-built map of the environment. This step, often called loop closure in robotics contexts, is what prevents IMU dead-reckoning from diverging over corridors longer than approximately 20 meters.

The IEEE 802.11az standard (Next Generation Positioning) formalizes fine-timing measurement procedures for Wi-Fi–based ranging, targeting centimeter-class accuracy in dense deployments. UWB implementations conforming to IEEE 802.15.4z achieve two-way ranging precision of ±10 centimeters in line-of-sight conditions.

Common Scenarios

Warehouse and logistics — autonomous mobile robots (AMRs) fuse wheel odometry, LiDAR scan matching, and UWB anchors to navigate shelving aisles with clearances as narrow as 0.5 meters. Position errors exceeding 15 centimeters can cause collision events.

Hospital asset tracking — BLE beacon networks combined with IMU-equipped tags track medical equipment across multi-floor structures. The Joint Commission's environment-of-care standards create compliance pressure to locate crash carts and infusion pumps within defined response windows.

Public safety and first responder — firefighters and tactical teams use body-worn IMU-plus-barometer units, with barometric pressure providing floor-level disambiguation when stairwell traversal interrupts radio contact. NIST PSCR's 2023 competition results documented leading-team accuracies below 0.5 meters mean absolute error in multi-story test buildings.

Smart buildings — occupancy sensing for HVAC optimization fuses passive infrared (PIR) arrays with Wi-Fi probe detection. This application tolerates zone-level accuracy (±3 meters) rather than sub-meter precision, enabling lower-cost deployments. Smart-home and building applications increasingly adopt this tiered accuracy model.

Decision Boundaries

Selecting a fusion architecture for indoor localization requires matching algorithmic complexity to accuracy requirements, latency constraints, and infrastructure cost.

EKF vs. particle filter: EKF is appropriate when position uncertainty remains unimodal and the environment model is known. Particle filters are preferred when position hypotheses are multimodal — for example, when a device could be on any of 3 floors — but require 10× to 100× more computation per update cycle.

Infrastructure-dependent vs. infrastructure-free: UWB and BLE systems require pre-installed anchor hardware but achieve consistent sub-meter accuracy. IMU-only pedestrian dead reckoning requires no infrastructure but accumulates 2–5% of distance traveled as error, becoming unusable beyond corridors of 50 meters without correction. Noise and uncertainty management techniques such as zero-velocity updates (ZUPTs) extend IMU-only range by resetting velocity estimates at detected footfall events.

Centralized vs. decentralized processing: A centralized fusion architecture routes all sensor data to a single processor for joint estimation — optimal for accuracy but vulnerable to single-point failure. Decentralized architectures distribute estimation across nodes, trading a fraction of accuracy for resilience and reduced communication bandwidth, which aligns with edge-computing deployment models in large facilities.

Accuracy metrics for indoor systems are standardized in ISO/IEC 18305:2016, which defines test methodologies for indoor positioning systems including cumulative distribution function (CDF) error reporting — the format used in NIST PSCR competition scorecards.

References