You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 4 Next »

Here we track the items in the backlog for architecture:

Backlog


DescriptionDecisionsAdoptionSpike/POC
Data VerificationHow do we verify the data is available for a data call / regulatory report?
  • The verification that the data is available is made at the time of the extraction
  • The data format(s) are given a specific name and version.  This includes the "level" of data expected.
  • The extraction identifies the format and version that is required for the extraction to succeed.
  • By consenting to the extraction, the data owner asserts that the data meets the format requirements.



Data ValidationWhat validation of the data is provided?  Is it all part of openIDL?  Is it shared across the community?  Is there some provided by the carrier/member?
  • For stat reporting, the stat-agent is required to validate the data for accuracy.
  • The validation rules are common and can be shared across all data owners.



Edit PackageSDMA provides a way to identify and fix errors as they occur before submission of the data.  Does this belong in the openIDL as a community component? If so, how do we provide that?
  • A reference implementation is made available for the etl that updates the HDS
  • The data owner is responsible for the implementation of the ETL
  • The data must meet certain levels of correctness (see SDMA for guidance)



Adapter Hosting ApproachThe adapter runs multiple components required by participants in openIDL.  How can we host these in member environments?
  • Software running in the Adapter is provided as Docker containers
  • The images are maintained in open source.  Members are free to vet and bring the images internally.
  • A mechanism for updates must be agreed upon.  This includes an SLA that describes the timeframe for absorbing updates.



Adapter ComponentsWhat are the components in the Adapter?



Extraction TechnologyWhat is the technology used to execute the extractions in the member environment?



Data Standard FormatWhat is the format for the data?  The data at rest in the HDS is the same or different from the data as it is being validated?



Data Standard LevelsWhat are the levels of the data standard?  How are they identified by extractions?  What do they cover?  Timeframe, data?



Discussion

Data Verification

When data is inserted into the harmonized data store, the insertion is registered on ledger as passing all validation and supporting some level of extraction.

Is there a need to capture a hash of the data?

The data inserted into the HDS is captured per policy.  The history of the transactions is captured, so the "state" of the policy at a given date may change between the time data is initially inserted and when it is extracted.

The data extraction will state what level of data is required.  How does the member attest that the data meets that expectation?  Does the consent constitute that attestation?


Data Validation


  • No labels