The observational roots of reference of the semantic web
Shared reference is an essential aspect of meaning. It is also indispensable for the semantic web, since it enables to weave the global graph, i.e., it allows different users to contribute to an identical referent. For example, an essential kind of referent is a geographic place, to which users may contribute observations. We argue for a human-centric, operational approach towards reference, based on respective human competences. These competences encompass perceptual, cognitive as well as technical ones, and together they allow humans to inter-subjectively refer to a phenomenon in their environment. The technology stack of the semantic web should be extended by such operations. This would allow establishing new kinds of observation-based reference systems that help constrain and integrate the semantic web bottom-up.
💡 Research Summary
The paper tackles a foundational problem for the Semantic Web: how to achieve truly shared reference to real‑world entities so that disparate contributors can reliably point to the same thing. While current Semantic Web practice relies heavily on ontologies and globally unique identifiers (URIs), this approach often abstracts away the concrete, observable phenomena that users actually encounter. Consequently, the link between a digital identifier and the physical or social reality it is meant to denote remains tenuous, especially when multiple users describe the same entity from different perspectives.
To bridge this gap, the authors propose a human‑centric, operational model of reference grounded in three complementary competencies: perceptual, cognitive, and technical. Perceptual competence concerns the acquisition of raw sensory data—through sensors, photographs, audio recordings, or direct human observation—about a phenomenon in the environment. Cognitive competence involves the mental processes that classify, abstract, and linguistically label that data, mapping it onto concepts such as “park,” “river,” or “historical monument.” Technical competence then translates these cognitively derived labels into Semantic Web artifacts (RDF triples, OWL classes, SKOS concepts) while simultaneously attaching rich provenance and contextual metadata.
The central construct introduced is “observation‑based reference.” Rather than treating a URI as a static pointer, each reference is accompanied by a structured description of the observation that generated it: who observed, when, with what instrument, under which conditions, and what methodological steps were taken. This meta‑layer enables automatic alignment of multiple observations that refer to the same underlying entity, even when the surface identifiers differ.
Implementation-wise, the authors suggest extending the existing Semantic Web stack with three orthogonal layers:
- Observation Layer – a formal model for representing the act of observation itself, including raw data pointers and measurement parameters.
- Provenance Layer – a detailed record of the agents, timestamps, devices, and processes involved, building on W3C PROV standards but enriched for the specific needs of reference.
- Context Layer – a capture of situational factors (geographic coordinates, cultural setting, temporal granularity) that influence how an observation should be interpreted.
These layers sit alongside traditional URI‑based identification, creating a hybrid system where a place like “Central Park” is not just a single URI but a constellation of observations: GPS traces, photographs, citizen‑science notes, and sensor readings, each annotated with provenance and context. Machine‑learning techniques can then cluster these observations, discover emergent concepts, and dynamically evolve the underlying ontology in a bottom‑up fashion.
The authors argue that this operational, observation‑driven approach mitigates the rigidity of top‑down ontology engineering. By allowing real‑world data and user contributions to shape the reference structure, the Semantic Web can grow organically, improving interoperability, data quality, and trustworthiness. Moreover, the model supports fine‑grained access control and quality assessment because each observation carries its own credibility indicators.
In the discussion, the paper highlights several implications: (a) enhanced inter‑subjectivity, as multiple users can converge on the same referent through shared observational evidence; (b) better alignment with existing standards for provenance (PROV-O) and geospatial data (GeoSPARQL); (c) a pathway for integrating heterogeneous sensor networks, citizen‑science platforms, and cultural heritage databases into a unified graph.
The conclusion reiterates that embedding human perceptual and cognitive processes into the technical fabric of the Semantic Web is essential for robust, scalable reference. Future work is outlined, including the development of standardized vocabularies for observation metadata, algorithms for automatic observation clustering, and pilot deployments in domains such as environmental monitoring, smart cities, and digital humanities. By grounding reference in observable reality, the Semantic Web can achieve the promised vision of a globally shared, machine‑understandable knowledge graph that remains faithful to the world it describes.