pix4dcatch indoor 4

Indoor mapping workflow with Autotags on your phone

Smartphone scanning poses a challenge in large indoor spaces due to "drift." We address this with Autotag detections. See our full workflow!

Mapping large indoor spaces with a smartphone can be a real challenge. As you walk and scan, errors in position and orientation can build up over time with longer captures, a phenomenon known as "drift." This drift can accumulate and cause the final model to be inaccurate, with warped or duplicated surfaces that make the data unusable for professional purposes. This is a challenge for anyone trying to capture large or complex indoor environments with a mobile device.

To address this, we tested a workflow using an iPhone with PIX4Dcatch, Pix4D Autotags, and the GeoFusion algorithm. Autotags are an effective method to correct drift on your phone; the Autotags are detected in real time during data acquisition and are used with tracking information to correct drift immediately after data collection is completed. After data acquisition, the GeoFusion algorithm plays a crucial role: it uses all the data to figure out the exact location and angle of the camera for every picture, ensuring a highly accurate result.

This combination allows users to capture accurate data in real-time, even in large spaces where traditional methods fail.

The article is based on a research paper by Christoph Strecha, Martin Rehak, and Davide Cucci.

pix4dcatch indoor mapping
iPhone with PIX4Dcatch.The phone’s acquired LiDAR data can be overlaid on the current image, providing real-time feedback on the completeness of acquisition. Autotags are detected in real-time (indicated in blue)

PIX4Dcatch's GeoFusion algorithm

First, we wanted to ascertain the effect of the GeoFusion algorithm when used with Autotags by doing a comparative test. We did this by walking in a loop indoors with PIX4Dcatch on two phones, with identical settings, except one phone had “Tag Detection” enabled and the other did not. The latter relied on the app's inbuilt “minimalistic mode”, using live video feed, phone sensors such as GPS and compass — a technology similar to the well-known Simultaneous Localization and Mapping (SLAM).

After the walk, we compared the results to see how accurately our phones' cameras had "seen" the Autotags. The "re-projection error" is a measurement of this accuracy—the lower the number, the more accurate the camera's position.

The results showed a huge difference. The phone using Tag Detection had a much lower average error (7 pixels) compared to the phone without Tag Detection enabled (41 pixels). This means the GeoFusion method was much better at keeping track of the camera's position and significantly reduced the problem of "drift," which is when the phone's position becomes less accurate over time. This was especially clear at the beginning of the test, where drift is usually the worst.

Merging multiple scans for large spaces

Using this method, we successfully mapped a 2000 m² office building. For scenarios like this, you can capture multiple individual scans and merge them later. The Autotags are essential for ensuring accurate alignment of the merged datasets.

pix4dcatch indoor 2
Screenshots from PIX4Dmatic show the 2000 m2 office space from a top view across five different scans. In each image, one can observe the Autotags marked in green and the iPhone LiDAR point cloud, which is colored with the images.

Since there's no GPS indoors, each scan is initially in its own coordinate system. By strategically placing Pix4D Autotags, the software can use them as reference points to automatically align and merge multiple scans. These fixed reference points help the program accurately place all the photos and build the 3D model. Using the shared Autotags, PIX4Dmatic stitched all five of our projects together into one single, massive model.

After a final check and optimization of the whole merged project, the result was a detailed 3D digital reconstruction of the 60 × 42 m office space, created by combining the data from about 5,000 photos and LiDAR scans taken with an iPhone.

pix4dcatch indoor 3
Screenshots from PIX4Dmatic of the merged datasets from the image above are shown. One can observe the Autotags in green, the 5000 camera positions as small green dots, the iPhone LiDAR point cloud colored with the images, and the dense photogrammetry point cloud.

Interior and exterior: a unified 3D model

The second project involved a combined scan of a family house, where data from both the exterior and various floors were collected. A total of 15 PIX4Dcatch scans were captured, each processed individually before being merged in PIX4Dmatic. The final result is a single 3D model of the entire house, demonstrating the ability to unify both indoor and outdoor data from a multi-level structure using PIX4Dcatch and PIX4Dmatic. The final point cloud is shown in various cutting planes below.

pix4dcatch indoor 4
Screenshots from PIX4Dmatic of a house from 15 different PIX4Dcatch scans that have been merged with Autotags. Autotags facilitate the modeling for both indoor and outdoor environments

We successfully combined exterior and multi-floor interior data, proving that our mobile scanning method with PIX4Dcatch, PIX4D Autotags, and GeoFusion is robust enough to accurately merge everything into a single, comprehensive interior and exterior 3D model in PIX4Dmatic.

Indoor mapping workflow with Pix4D Autotags

Our workflow is designed to be simple and accessible for everyone. Here’s how it works:

  1. Place the Autotags: First, you strategically place the Autotags in the area you want to scan; their position does not need to be known in advance.
  2. Scan the area: Use PIX4Dcatch to walk around the area, capturing images and LiDAR data. The software automatically detects the Autotags in real-time.
  3. Let the software do the work: After you finish the data acquisition, the GeoFusion algorithm runs on the phone in seconds, correcting for drift and optimizing the camera positions.
  4. Merge the scans (for large projects): For very large projects, you can combine multiple datasets together to create a full 3D model of the entire space.

Indoor reality capture: affordable workflow for challenging environments

This workflow is a significant advancement for anyone who needs to capture accurate 3D data of indoor spaces. For a long time, professionals had to choose between two options: expensive, specialized laser scanners that provide high accuracy, or simple mobile apps that are fast and cheap but suffer from major accuracy issues.

Our solution changes that. It bridges the gap by providing a professional-grade tool that uses a device you already have—your smartphone. Ultimately, this workflow empowers you to get high-quality, actionable 3D data from challenging environments quickly and affordably.


High-quality data capture with PIX4Dcatch

Related articles