provenance

Provenance of the incipient passive margin of NW Laurentia (Neoproterozoic): detrital zircon from continental slope and basin floor deposits of the Windermere Supergroup, southern Canadian Cordillera

Hadlari, T; Arnott, R W C; Matthews, W A; Poulton, T P; Root, K; Madronich, L I. Lithosphere vol. 2021, no. 1, 8356327, 2021 p. 1-10, https://doi.org/10.2113/2021/8356327
<a href="https://geoscan.nrcan.gc.ca/images/geoscan/20210465.jpg"><img src="https://geoscan.nrcan.gc.ca/images/geoscan/20210465.jpg" title="Lithosphere vol. 2021, no. 1, 8356327, 2021 p. 1-10, https://doi.org/10.2113/2021/8356327" height="150" border="1" /></a>




provenance

Timing and provenance of Paleoproterozoic supracrustal rocks in the central Thelon tectonic zone, Canada: implications for the tectonic evolution of western Laurentia from ca. 2.1 to 1.9 Ga

Davis, W J; Sanborn-Barrie, M; Berman, R G; Pehrsson, S. Canadian Journal of Earth Sciences vol. 58, issue 4, 2021 p. 378-395, https://doi.org/10.1139/cjes-2020-0046
<a href="https://geoscan.nrcan.gc.ca/images/geoscan/20200302.jpg"><img src="https://geoscan.nrcan.gc.ca/images/geoscan/20200302.jpg" title="Canadian Journal of Earth Sciences vol. 58, issue 4, 2021 p. 378-395, https://doi.org/10.1139/cjes-2020-0046" height="150" border="1" /></a>




provenance

The age and provenance of the Lay Range assemblage provides an indirect record of basement to north-central Quesnellia, British Columbia

Ootes, L; Ferri, F; Milidragovic, D; Wall, C. Geological fieldwork 2021: a summary of field activities and current research; British Columbia Geological Survey Geological Fieldwork Paper 2022-01, 2022 p. 31-44
<a href="https://geoscan.nrcan.gc.ca/images/geoscan/20210442.jpg"><img src="https://geoscan.nrcan.gc.ca/images/geoscan/20210442.jpg" title="Geological fieldwork 2021: a summary of field activities and current research; British Columbia Geological Survey Geological Fieldwork Paper 2022-01, 2022 p. 31-44" height="150" border="1" /></a>




provenance

Using X.509 Certs for DID Provenance

Summary: The abundance of X.509 certificate authorities who already perform identity proofing for businesses provides a rich resource that can be leveraged to boot the verifiable data ecosystem.

When you used a verifiable credential to prove something about yourself, the verifier can know cryptographically: (1) the identifiers for the issuer, (2) the credential hasn't been tampered with, (3) the credential was issued to you, and (4) the credential hasn't been revoked. These four checks are important because their establish the fidelity of the data being transferred. They don't, however, tell them whether they can trust the issuer. For that, they need to take the issuer's decentralized identifier (DID) that they got from credential presentation and determine who it belongs to.

At the most recent Internet Identity Workshop, Drummond Reed gave a session on how X.509 certificates could help with this. The first step, like always, is to resolve the DID and retrieve the DIDDoc that associates keys and endpoints with the DID. The endpoint can be an HTTP server and, of course, should have an X.509 certificate providing TLS security. That certificate, at the very least, has a a domain name to bind that to the certificate's public key. It can, if you pay for the feature, also include information about the entity that applied for the certificate. The certificate authority proofs that information and is vouching for it when they sign the certificate.

The key to making the X.509 certificate useful for checking the provenance of a DID lies in one key change. X.509 certificates can contain and extended field called a Subject Alternative Name. This following figure shows how it can help.

Using X.509 Certificates to establish the owner of a DID (click to enlarge)

In this figure:

  1. The issuer (Attestor) creates the DID they will use to issue the certificate along with its associated DIDDoc, including an HTTP endpoint for DID verification.
  2. Attestor applies for a X.509 certificate for that endpoint, including in the application the DID they created in (1).
  3. The certificate authority does it's usual proofing of the application and issues a certificate that includes the DID in the Subject Alternative Name field.
  4. The issuer creates a credential definition in the usual way that includes their DID and writes it to whatever Verifiable Data Registry their DID method dictates.
  5. Attestor issues a credential to a holder (Alice) using that credential definition.
  6. At some later time, Alice presents the credential to the verifier (Certiphi).
  7. Certiphi resolves the DID to get the DIDDoc and retrieves the verfication endpoint from the DIDDoc
  8. Certiphi retrieves the certificate for that endpoint1.
  9. Certiphi verifies the certificate by checking it's signature and ensures that the DID in the DIDDoc for the credential matches the one in certificate.2

The issuer's DID has now been tied in a verifiable way to whatever information is in the certificate. Provided the certificate includes information about the entity beyond the domain name, the verifier can use that information to determine whether or not the credential is authentic (i.e., issued by who the credential definition purports issued it). That might be all the evidence they need to determine whether to trust the entity. Certificate authorities could also issue verifiable credentials to the customer attesting the same verified claims—after all, it's one more product they can offer.

The benefit of doing issuer validation using X.509 certificates is that there are already many trusted X.509 certificate authorities in business who already do proofing of attributes about businesses. That's a huge chunk of the verifiable data ecosystem that doesn't need to be built because it can be leveraged. To make this work, digital certificate authorities would need to start offering to validate DIDs and include them in a certificate as a Subject Alternative Name. I don't discount that this will take some bureaucratic maneuvering. Certificate authorities will need to see a business opportunity. I'd love to see Digitcert or someone do a pilot on this.

Notes

  1. Note that this step might be combined with the previous step if the Verifiable Data Registry is the same server as the endpoint, but that's not necessarily going to be the case for a number of reasons.
  2. Note that this does not create a call back wherein Attestor can determine which credential was used, preserving the privacy of the presentation. Attestor does know one of its credentials has been presented to Certiphi. If this information leakage bothers you, then any web-based DID method is potentially a problem.

Tags:




provenance

Why provenance of SPARQL 1.1 queries

In this paper, we study and provide algorithms for source-provenance of answers to extended SPARQL queries. Extended SPARQL queries are an extension of SPARQL 1.1 queries which support not only a single dataset but multiple datasets, each in a particular context. For example, normal subqueries, aggregate subqueries, (NOT) EXISTS filter subqueries may (optionally) have their own dataset. Additionally, GRAPH patterns can query multiple RDF graphs from the local FROM NAMED dataset and not just one. For monotonic queries, the source why provenance sets that we derive for an answer mapping are each the minimal set of sources appearing in the query that if we consider as they are while the rest of the sources are considered empty, we derive the same answer mapping. We show that this property does not hold for non-monotonic queries. Among others, knowing source why provenance is of critical importance for judging confidence on the answer, allow information quality assessment, accountability, as well as understanding the temporal and spatial status of information.




provenance

Sheer Curation of Experiments: Data, Process, Provenance

This paper describes an environment for the “sheer curation” of the experimental data of a group of researchers in the fields of biophysics and structural biology. The approach involves embedding data capture and interpretation within researchers' working practices, so that it is automatic and invisible to the researcher. The environment does not capture just the individual datasets generated by an experiment, but the entire workflow that represent the “story” of the experiment, including intermediate files and provenance metadata, so as to support the verification and reproduction of published results. As the curation environment is decoupled from the researchers’ processing environment, the provenance is inferred from a variety of domain-specific contextual information, using software that implements the knowledge and expertise of the researchers. We also present an approach to publishing the data files and their provenance according to linked data principles by using OAI-ORE (Open Archives Initiative Object Reuse and Exchange) and OPMV.




provenance

Automatic Generation of Temporal Data Provenance From Biodiversity Information Systems

Aim/Purpose: Although the significance of data provenance has been recognized in a variety of sectors, there is currently no standardized technique or approach for gathering data provenance. The present automated technique mostly employs workflow-based strategies. Unfortunately, the majority of current information systems do not embrace the strategy, particularly biodiversity information systems in which data is acquired by a variety of persons using a wide range of equipment, tools, and protocols. Background: This article presents an automated technique for producing temporal data provenance that is independent of biodiversity information systems. The approach is dependent on the changes in contextual information of data items. By mapping the modifications to a schema, a standardized representation of data provenance may be created. Consequently, temporal information may be automatically inferred. Methodology: The research methodology consists of three main activities: database event detection, event-schema mapping, and temporal information inference. First, a list of events will be detected from databases. After that, the detected events will be mapped to an ontology, so a common representation of data provenance will be obtained. Based on the derived data provenance, rule-based reasoning will be automatically used to infer temporal information. Consequently, a temporal provenance will be produced. Contribution: This paper provides a new method for generating data provenance automatically without interfering with the existing biodiversity information system. In addition to this, it does not mandate that any information system adheres to any particular form. Ontology and the rule-based system as the core components of the solution have been confirmed to be highly valuable in biodiversity science. Findings: Detaching the solution from any biodiversity information system provides scalability in the implementation. Based on the evaluation of a typical biodiversity information system for species traits of plants, a high number of temporal information can be generated to the highest degree possible. Using rules to encode different types of knowledge provides high flexibility to generate temporal information, enabling different temporal-based analyses and reasoning. Recommendations for Practitioners: The strategy is based on the contextual information of data items, yet most information systems simply save the most recent ones. As a result, in order for the solution to function properly, database snapshots must be stored on a frequent basis. Furthermore, a more practical technique for recording changes in contextual information would be preferable. Recommendation for Researchers: The capability to uniformly represent events using a schema has paved the way for automatic inference of temporal information. Therefore, a richer representation of temporal information should be investigated further. Also, this work demonstrates that rule-based inference provides flexibility to encode different types of knowledge from experts. Consequently, a variety of temporal-based data analyses and reasoning can be performed. Therefore, it will be better to investigate multiple domain-oriented knowledge using the solution. Impact on Society: Using a typical information system to store and manage biodiversity data has not prohibited us from generating data provenance. Since there is no restriction on the type of information system, our solution has a high potential to be widely adopted. Future Research: The data analysis of this work was limited to species traits data. However, there are other types of biodiversity data, including genetic composition, species population, and community composition. In the future, this work will be expanded to cover all those types of biodiversity data. The ultimate goal is to have a standard methodology or strategy for collecting provenance from any biodiversity data regardless of how the data was stored or managed.




provenance

Flow-Based Provenance

Aim/Purpose: With information almost effortlessly created and spontaneously available, current progress in Information and Communication Technology (ICT) has led to the complication that information must be scrutinized for trustworthiness and provenance. Information systems must become provenance-aware to be satisfactory in accountability, reproducibility, and trustworthiness of data. Background: Multiple models for abstract representation of provenance have been proposed to describe entities, people, and activities involved in producing a piece of data, including the Open Provenance Model (OPM) and the World Wide Web Consortium. These models lack certain concepts necessary for specifying workflows and encoding the provenance of data products used and generated. Methodology: Without loss of generality, the focus of this paper is on OPM depiction of provenance in terms of a directed graph. We have redrawn several case studies in the framework of our proposed model in order to compare and evaluate it against OPM for representing these cases. Contribution: This paper offers an alternative flow-based diagrammatic language that can form a foundation for modeling of provenance. The model described here provides an (abstract) machine-like representation of provenance. Findings: The results suggest a viable alternative in the area of diagrammatic representation for provenance applications. Future Research: Future work will seek to achieve more accurate comparisons with current models in the field.




provenance

74% of online products in UK feature green claims: BRC & Provenance

Seventy-four per cent of online products in the UK feature green claims, averaging 2.9 claims per product, according to BRC and Provenance research. Nature-related claims were most prevalent, while carbon-related claims were few. To support compliance, BRC and Provenance are launching Retailer Green Claims Forum, promoting transparency in green marketing.




provenance

We Have Urgent Questions About the Unholy Provenance of Netflix’s Hot Frosty

Who built this ripped, anatomically graphic snowman? Is there a world of snowmen offscreen waiting to be turned into sex objects for widows?




provenance

Provenance for the Description Logic ELHr. (arXiv:2001.07541v2 [cs.LO] UPDATED)

We address the problem of handling provenance information in ELHr ontologies. We consider a setting recently introduced for ontology-based data access, based on semirings and extending classical data provenance, in which ontology axioms are annotated with provenance tokens. A consequence inherits the provenance of the axioms involved in deriving it, yielding a provenance polynomial as an annotation. We analyse the semantics for the ELHr case and show that the presence of conjunctions poses various difficulties for handling provenance, some of which are mitigated by assuming multiplicative idempotency of the semiring. Under this assumption, we study three problems: ontology completion with provenance, computing the set of relevant axioms for a consequence, and query answering.