Blog article

Navigating the Era of Synthetic Imagery: Why Trust in Geospatial Data Matters

Discover how OGC’s Integrity, Provenance, and Trust (IPT) framework ensures data reliability, transparency, and accuracy across industries.

In an era where synthetic imagery, deepfakes, and increasingly complex AI workflows dominate our digital landscape, the question of trust in data has never been more critical. How can we rely on the data that powers our most important decisions—from flood and wildfire management to national security, emergency response, and infrastructure planning?

Consider the catastrophic flooding in Germany in 2021. Despite advanced warning systems, questions emerged about the accuracy and timeliness of the data behind flood prediction models—contributing to delays in evacuations and the loss of over 180 lives. In Australia, during the 2020 bushfires, gaps and inaccuracies in satellite imagery made it harder for emergency responders to plan and deploy resources effectively.

The stakes are just as high in national security and public information. In 2022, a deepfake video of a prominent political leader went viral, briefly shaking public confidence and prompting urgent questions about authenticity and attribution. During the COVID-19 pandemic, misinformation and inconsistent data reporting sowed confusion and complicated efforts to coordinate an effective public health response. These are not fringe scenarios—they are a preview of what happens when data integrity, provenance, and trust aren’t built into our systems by design.

That’s why the Open Geospatial Consortium (OGC) is developing a comprehensive framework for Integrity, Provenance, and Trust (IPT) in geospatial data. This work aims to create a foundation that organizations around the world can rely on—and build upon—to ensure their data is not only useful, but trustworthy.

This is the start of a year-long initiative to define what trustworthy geospatial data really looks like—and how we, as a global community, can ensure it underpins the systems we all depend on.

Why It Matters Now

Organizations today rely on an increasingly vast and interconnected data ecosystem—often ingesting data from external providers or processed via black-box AI models. But this convenience comes at a cost: can we be sure that the data is what it claims to be?

We’ve already seen fabricated satellite imagery, AI-generated 3D urban models, and synthetic road networks. These aren’t just theoretical risks—they’re real challenges to public trust, operational accuracy, and accountability.

In high-stakes environments, uncertainty about data integrity or origin can lead to faulty analyses, missed warnings, or even policy failures. It’s time to shift from blind trust to evidence-based trust.

Let’s start with some definitions and an example to set the perspective of OGC in building this framework.

Integrity refers to how data has been handled throughout its lifecycle, including its characteristics such as content, accuracy, and completeness from collection to processing and distribution. Integrity can be compromised by tampering, errors, or undocumented changes, making safeguards essential to ensure reliability.

Provenance refers to the origin of the data, how it has been modified, and who or what performed those modifications. Provenance is compromised when operations that alter the data go unrecorded, affecting both traceability and integrity.

When Integrity and Provenance are well-documented, consistent, and unalterable, there is a foundation for Trust in data. Trust does not simply mean that the user is comfortable that the data meets their expectations, but also that the source, evolution, and suitability of the data can be unambiguously described so that others can also use the data.

For instance, a satellite with known camera parameters captures an image. That image is passed directly to an organization that adjusts it to fit a terrain model using a well-defined algorithm. The adjusted image is then used to calculate the size of a lake knowing that the source has a certain pixel resolution and that the processing included manipulation of the data to retain that original resolution. If each of these transactions—from collection to processing to delivery—is documented using standardized and unambiguous parameters, then users can trust that the data are suitable for their needs and that analyses can be validated and reproduced.

BUILDING TRUST IN GEOSPATIAL DATA: PROGRESS AND NEXT STEPS

OGC and others have made significant progress in developing Standards, tools, and frameworks that enhance IPT in geospatial data. These efforts provide a foundation for ensuring data reliability, traceability, and usability across industries. Some aspects of the IPT framework are already in place or soon to emerge, such as the following.

Integrity

Metadata Standards define the original and transformed data. There are well-established Standards for recording metadata, but guidance on effective use of common terminology to allow “apples-to-apples” comparisons needs to be developed.

OGC’s sensor model registry ensures consistent descriptions of sensor capabilities. This registry solves one part of the problem, but similar sets of definitions to describe other sources and capture methods for data need to be developed or integrated.

Provenance

OGC’s Training Data Markup Language for AI standardizes training and validation datasets. New work is being proposed in OGC to develop specialized metadata to describe the AI models used in geospatial data processing.

OGC and ISO’s data quality measures registry establishes quantifiable, consistent descriptions of data quality. These quality measures are applicable to describing the integrity of the data and the impact of processing that makes up the provenance of the information.

Provenance building blocks are being implemented by OGC on multiple European Union projects. OGC API Standards are constructed of a number of independently implementable building blocks of functionality that collectively can be assembled to create an implementation. In the same vein, descriptors of operations that impact integrity and that are part of the provenance of the data can be modular and inserted into each step in the lifecycle of data.

What Comes Next

Despite these advancements, a comprehensive Trust model—built from interoperable, standards-based IPT components—has yet to be completed.

Over the next year, OGC will work with members and partners to:

  • Integrate IPT building blocks into operational systems and workflows
  • Ensure descriptors are machine-readable, tamper-evident, and actionable
  • Package IPT parameters with datasets for end-to-end transparency
  • Reflect real-world user needs through open collaboration and shared testing

Be Part of the Future of Trusted Data

This blog kicks off a deeper dive into the Integrity, Provenance, and Trust framework and how we, as a community, can build something meaningful—and scalable—together.

We’ll be releasing technical spotlights, real-world use cases, and engagement opportunities throughout the year. Whether you’re a standards developer, data user, AI practitioner, or public servant, your insight is essential.

Join the movement:

Let’s make data trust visible, measurable, and shared—so that everyone can make better decisions with confidence.