Guide: Governance interoperability design and conformance testing

Track

Test bed setup

About this guide

This guide focuses on a solution’s governance, discussing design guidelines to enable interoperability at a governance level across its stakeholders. For each considered topic the guide also offers tips on covering such governance aspects as part of your conformance testing setup.

What you will achieve

At the end of this guide you should have an improved awareness over a solution’s governance aspects, and the agreements involved in ensuring interoperability at a governance level for the involved implementing parties. For each discussed governance topic you will be presented with design guidelines sourced from best-in-class standards and specifications, and see which parts of such agreements you can practically cover as part of your conformance testing design.

If you are interested in more details on the background and sources behind this guide, you will also find a summary of the European Interoperability Reference Architecture (EIRA©) and the EIRA© Library of Interoperability Specifications (ELIS) that you can proceed to further explore.

Note

The discussed topics and guidelines in this guide are not meant to be exhaustive. They represent common elements per topic sourced from the standards and specifications included in the ELIS.

What you will need

  • About 2 hours to go through all points.

  • A high-level understanding of IT concepts such as encryption and APIs. Note that no technical details are discussed, nor is there need to have a technical profile to follow the guide.

  • (Optional) Check the Test Bed’s developer onboarding guide to get a high-level understanding of Test Bed technical concepts. Although not necessary to follow the guide, such concepts are often mentioned when discussing conformance testing guidelines.

How to complete this guide

The first chapter of this guide covers background concepts, and explains the guide’s rationale and approach. Although interesting, you may choose to skip this and go directly to the subsequent sections discussing legal, organisational, semantic and technical governance topics.

Each of these chapters is further split into topics to group together similar aspects. Within each topic section you will find the same structure, presenting first design guidelines for the topic in question, followed by conformance testing tips. To make each of these distinct and clear, they are presented in separate highlight boxes.

Design guidelines

Listing of design guidelines for the discussed topic

Conformance testing tips

Proposals on what can be tested and how to implement the resulting tests

Background

The sections in this chapter provide background information on governance interoperability, and the approach followed by this guide to extract design and testing guidelines.

The European Interoperability Reference Architecture (EIRA©)

The European Interoperability Reference Architecture (EIRA©) is a reference architecture to implement interoperable European digital public services. It extends the ArchiMate® enterprise architecture modelling language, reusing ArchiMate® elements to define the key architecture building blocks (ABBs) needed to build interoperable e-Government systems. These building blocks are furthermore organised in four views as defined by the European Interoperability Framework.

../_images/eif.png

Figure 1: Interoperability views defined by the European Interoperability Framework

These views define the structure of the EIRA©, serving to classify a digital public service’s architecture building blocks based on their purpose. Specifically:

  • The Legal view building blocks cover its legal basis and motivation.

  • The Organisational view building blocks capture its organisational structure.

  • The Semantic view building blocks define its aspects pertaining to data and metadata.

  • The Technical view building blocks define its technical application and infrastructure elements.

../_images/eira.png

Figure 2: Organisational view excerpt showing ABBs related to service delivery (EIRA© v6.1.0)

Architecture building blocks are supported by rich definitions, rendering them a common terminology that can be used by administrators, architects and developers. Furthermore, the building blocks’ interconnection through the Legal, Organisational, Semantic and Technical views results in an implicit narrative that links all elements, and enables traceability within and across different layers. As an example, one can use the EIRA© to trace the legal basis for specific application components, and the reason why they are deployed in a specific hosting environment.

Besides its views and building blocks, the EIRA© provides insight and knowledge on different perspectives of the core architecture. Specific topics are addressed by means of viewpoints, diagrams that complement the four core views by taking a specific perspective considering the topic in question. Among these, and of particular interest for this guide, is the Key Interoperability Enablers viewpoint, which shows the key elements needed to achieve interoperability for the four core views, but based on different interoperability dimensions.

../_images/dimensions.png

Figure 3: Interoperability dimensions and the Key Interoperability Enablers viewpoint (EIRA© v6.1.0)

These dimensions provide an interesting way to consider interoperability, as they define the nature of building blocks alongside their purpose. Three such dimensions are defined as follows:

  • Structural interoperability, pertaining to a solution’s structure, as manifested by the various components it consists of.

  • Behavioural interoperability, pertaining to a solution’s behaviour, in terms of data, information and knowledge exchanges with its environment.

  • Governance interoperability, pertaining to peer collaboration measures, agreements and shared policies, with stakeholders from external services.

Combining views, building block definitions, and interoperability dimensions, allows us to express and better understand a solution’s elements in terms of what they do and how they do it. This gives us interesting guidance when it comes to the solution’s conformance testing.

Conformance testing and interoperability dimensions

Conformance testing is the process of verifying that a solution correctly realises its requirements. In other words it helps us apply quality control over a solution’s design and implementation.

Addressing a solution’s design, meaning validating that its “blueprint” matches the requirements, the EIRA© proposes the EIRA© validator, a service taking as input an ArchiMate® model and producing a validation report. This report covers deviations from architecture blueprints (termed solution architecture templates), and highlights cases where the EIRA© modelling guidelines have not been correctly applied.

Note

The EIRA© validator is based on the Test Bed’s XML validator, validating ArchiMate® models represented in XML, using XML Schema and Schematron rules.

Moving onto conformance testing of a solution’s implementation is where the Test Bed comes in. Conformance testing here takes the form of scenario-based test cases that trigger and make assertions on the system’s perceived behaviour. This is typically achieved by sending and receiving messages to and from the system under test, using its APIs as defined by the requirements, and validating produced results.

Considering the previously mentioned interoperability dimensions, this falls under behavioural interoperability as we make assertions based on the system’s observed behaviour. In doing so however, we also implicitly test for structural interoperability given that the exhibited behaviour is only made possible by the components that constitute the system. It could of course be the case that certain expected behaviours can not be validated by interacting with the system from outside its boundaries, in which case we can foresee a manual submission and validation of evidences to reach a degree of assurance that the implementation is complete.

When we look at structural interoperability, and in particular the semantic building blocks used such as data models and ontologies, we can also consider the use of validators to complement conformance tests. The Test Bed defines validators as components that are configured with a specification’s validation artefacts, and serve in validating content over several APIs. Besides using validators as standalone services, these can also be integrated into conformance test cases to cover validation of exchanged data.

With respect to governance interoperability, conformance testing becomes less obvious given that we focus on collaboration rules and agreements between stakeholders. Even in this case however, we can typically extract common agreement elements that manifest as implementations within the systems under test, that can moreover be observed by interacting with them. As an example, a service level agreement stipulating minimum API response times can be tested by making calls, recording response times, and then asserting them against the required thresholds. Such an approach obviously can only be applied to a subset of topics, but nonetheless remains interesting to consider as part of a solution’s conformance testing design.

This guide focuses on governance interoperability to highlight the involved agreements and policies as a valid source of conformance test cases. When designing a solution’s conformance testing setup, we typically focus on testing the foreseen business processes and core technical elements. This guide, aims to highlight that governance-level requirements can also be partly tested for, allowing us to increase further our testing coverage and quality assurance.

The EIRA© Library of Interoperability Specifications (ELIS)

The EIRA© Library of Interoperability Specifications (ELIS) is a catalogue of best-in-class standards and specifications, categorised based on their relevant EIRA© architecture building blocks. When designing a solution, one can use the EIRA© to determine and express the building blocks needed, and then refer to the ELIS to determine the specifications to consider when implementing each building block (either through reuse or new development).

The selection of specifications is based on rigorous assessments by applying the Common Assessment Method for Standards and Specifications (CAMSS), whereby only the highest scoring specifications are included. These assessments are published in the ELIS alongside the specifications to provide transparency, and also to allow interested parties to go into further details if desired. Both the CAMSS and the ELIS are curated and maintained by the CAMSS support team, as part of the Interoperable Europe initiative.

As part of providing a curated library of interoperability specifications, the ELIS also implicitly becomes a valuable body of knowledge on numerous interoperability aspects. By examining the specifications proposed by the ELIS for a given EIRA© architecture building block, we can extrapolate common elements, requirements and guidelines that can help us design for interoperability. Considering these guidelines we can determine what is typically expected per topic, but also what we can most likely be able to test for as part of our conformance testing.

This guide follows this exact approach from the perspective of governance interoperability, helping us to identify common governance aspects we can include in our conformance testing. Specifically:

  1. We use the EIRA© to determine the building blocks enabling governance interoperability.

  2. For each building block we collect from the ELIS the proposed specifications.

  3. We analyse the specifications to identify common aspects to treat as design guidelines.

  4. We consider these guidelines to determine what can be included in conformance tests.

../_images/flow.png

Figure 4: Extracting design guidelines and testing tips for governance interoperability

The following sections result from this work, proposing guidelines and conformance testing tips for Legal, Organisational, Semantic and Technical governance interoperability.

Organisational governance

Organisational governance pertains to the agreements in place to define and facilitate collaboration within and across organisations.

Agreements on data sharing

Agreements on data sharing formalise the information requirements, syntax bindings, protocols and semantic artefacts that must be used for the exchange of data.

Topic: Data sharing

Design guidelines

  • Data quality standards must be met, including accuracy, completeness, and timeliness. Metadata must be provided in accordance with agreed metadata standards.

  • Use secure protocols to protect data during transmission, providing encryption and ensuring data integrity.

  • Define the use of APIs used for data exchange.

  • Specify the protocols for transferring large data files or batch data.

  • Foresee messaging and queuing protocols for asynchronous data exchange.

  • Manage communication protocol versioning to ensure compatibility and manage updates. Stakeholders must be informed of any changes to agreed protocols.

  • Data must be formatted and structured according to agreed syntax standards.

Conformance testing tips

Use of data and metadata standards are prime testing candidates, for which validators can be configured depending on the syntax in question (see for XML, RDF, JSON and CSV). Such validators can be used as standalone tools provided to a project’s community, but also integrated in test cases by means of verify steps to validate exchanged data.

Regarding use of secure protocols, involved APIs, asynchronous messaging and other data exchange aspects, this should be the key focus of test cases. Scenarios should be foreseen based both on sending and receiving messages, validating multi-step exchanges for consistency, and validating edge cases resulting from “unhappy flows”.

In terms of messaging design for test cases, the best practice is to consider entire business processes involving multiple exchange steps between the Test Bed and the system under test. Besides validating messages exchanged at each step, this allows you also to check that each message in the overall sequence is coherent with respect to its predecessors, by correctly referring to selected data and correlation identifiers. It is also a good practice to try and make each test session unique, by generating random data or selecting randomly elements from an agreed test data set. Doing so ensures that the system under test not only produces valid messages, but that it does so taking into account the specific test data of each test session.

Having said this, nothing prevents you from also defining more focused or technical-oriented test cases for individual message types, where messages are tested in isolation. In doing so you will effectively be introducing a set of unit tests that complements the test cases addressing business processes as a whole.

Agreements on the use of common infrastructure

Agreements on the use on common infrastructure outline the terms and conditions for sharing and utilising a specific infrastructure or facility between multiple parties.

Topic: Infrastructure usage

Design guidelines

  • Specify the financial arrangements for using the shared infrastructure. Ensure cost-sharing models support the objective of creating economic value from data by enabling cost-efficient access to data infrastructure.

  • Ensure each party is explicitly entitled to access and utilise the infrastructure as per the agreed terms; and is made responsible for adhering to usage policies, maintenance, and reporting of issues.

  • Define the security protocols to protect the shared infrastructure that must be followed by all parties. Security measures should comply with requirements for safeguarding data, including the protection of sensitive and private data.

  • Outline the terms and conditions under which the infrastructure can be accessed and used. Access to the infrastructure will be granted upon request and under the condition that usage does not exceed allocated resources.

Conformance testing tips

Such points have no bearing on a system’s perceived behaviour and cannot be practically tested via conformance tests.

Privacy frameworks and agreements

Privacy frameworks enable the confidentiality aspects of data, information and knowledge assets; and the organisational resources handling them. Agreements on privacy enable a set of rules for the personal data of individuals’ collection, processing and transference by Public Administrations.

Topic: Privacy

Design guidelines

  • The service provider must comply with all relevant data protection regulations, ensuring that personal data is processed lawfully, fairly, and transparently.

  • Explicit consent must be obtained from data subjects before processing their personal data, and mechanisms for managing and withdrawing consent must be provided.

  • Personal data must be anonymised or pseudonymised where possible to enhance privacy and reduce the risk of identification.

  • Only the minimum amount of personal data required for the service will be collected and processed, in accordance with the principle of data minimisation.

  • Regular Data Privacy Impact Assessments (DPIAs) must be conducted to identify and mitigate risks to personal data.

  • The service provider must provide mechanisms for data subjects to exercise their rights, including access, rectification, erasure, and data portability.

Conformance testing tips

A solution’s data protection implementation can be partially tested based on its exhibited behaviour. We can foresee test cases involving message exchanges in which test data is used that is considered for testing purposes as being of a private nature. As part of these tests we can add assertions following exchanges to ensure that the data in question is removed when no longer needed. This ideally would be done in an automated manner through specific additional requests, but could otherwise be achieved manually by providing evidence via interact steps (e.g. a screenshot or log extract). Following evidence submission by the tester, a second administrator-level interact step would then follow, allowing an administrator to verify that processing was correct, before completing the test session.

The anonymisation and pseudonymisation of data can also be tested for through relevant exchanges. For example, tests can be based on an agreed test dataset with mock private information, for which we can add assertions following exchanges to ensure that anonymisation and pseudonymisation have taken place. It is interesting to note that anonymisation and pseudonymisation can also take place on data managed by the Test Bed, by means of a custom extension that is used before data is processed. Depending on what is more meaningful, this could be done as part of a send step using a custom messaging implementation, or considered as a separate utility used through a process step that delegates to a custom processing service.

Security frameworks and agreements

Security frameworks enable the protection of various aspects of data, information and knowledge assets and the organisational resources handling them. Agreements on interoperability security formalise governance rules and conditions to grant the identification, authorisation and transmission of the data, information and knowledge being exchanged between digital public services.

Topic: Security and trust

Design guidelines

  • Ensure compliance with all legal, regulatory, and internal security policies. Regular audits and assessments must be conducted to ensure compliance and identify areas for improvement.

  • Ensure allocation of sufficient resources (financial, technical and human) to implement, maintain and continuously improve security measures.

  • Define risk management processes to identify, assess, and mitigate risks associated with the use and sharing of digital resources. These processes should also be subject to regular risk assessments.

  • Involve all relevant internal and external stakeholders in security planning and decision-making processes. Regular communication and collaboration are required to ensure comprehensive security measures.

  • Adhere to established security standards and best practices. Regular reviews must be performed to ensure compliance is maintained.

  • Ensure all partners and suppliers undergo a thorough vetting process to ensure they meet security standards and can be trusted with sensitive data. Regular audits must be conducted to ensure trustworthiness is maintained.

Conformance testing tips

Although most topics here refer to organisational alignment on security and trust, we could foresee test cases covering the aspects that would be reflected in a solution’s implementation. The primary focus here would be the agreed security standards in so far as they pertain to the solution’s operational security. As an example one could envision test cases focusing on key points of the OWASP Application Security Verification Standard (ASVS) in which the solution’s behaviour is assessed through tailored message exchanges.

Given the extensive requirements of a standard such as the OWASP ASVS, it would most likely be overkill to implement each of these as a distinct test case. Keep in mind that the focus of most solutions’ conformance testing is the business processes it implements, which by themselves could result in hundreds of test cases without even considering aspects such as security. In addition, note that the Test Bed is not a penetration testing solution; there are several purpose-built specific tools available, that provide out-of-the-box testing and reporting features specifically targetting security.

Nonetheless, it is always beneficial to provide at least some coverage for non-functional aspects such as security, by means of a limited set of core test cases that cherry-pick key requirements to test. By requiring this test suite to be completed alongside business-oriented tests, we increase our assurance of the tested systems’ quality, and at least raise awareness on the importance of such aspects.

Semantic governance

Semantic governance pertains to the policies and agreements ensuring that all parties share a common understanding and treatment of semantic assets.

Data policies

Data policies aim to form the guiding framework in which data management can operate.

Topic: FAIR principles compliance

Design guidelines

  • Ensure data is Findable through standardised metadata, unique and persistent identifiers, and data indexing in searchable repositories.

  • Ensure data is Accessible through clear access protocols, authentication and authorisation procedures if necessary, and archiving for long-term availability.

  • Ensure data is Interoperable through use of open and standard data formats, shared vocabularies and ontologies, and cross-referencing with other datasets and standards.

  • Ensure data is Reusable by means of clear licencing, detailed documentation, and continuous quality assurance processes to ensure data is valid and reliable.

Conformance testing tips

Aspects such as the use of specific data and metadata standards, vocabularies, ontologies and syntaxes can all be tested through test cases. The Test Bed includes several built-in validation capabilities to do so but also offers the possibility of defining full validators that are configured with the target specifications’ validation artefacts (e.g. SHACL shapes configured in a RDF validator instance). Such validators are usable as standalone services, but also within test cases as the implementation of verify steps.

Concerning data retrieval channels we can test aspects such as access protocols, as well as authentication and authorisation, by having test cases simulate data consumers. Moreover test cases can be based on an agreed test dataset, from which specific data is requested depending on the scenario being tested.

Semantic interoperability agreements

Semantic interoperability agreements formalise governance rules enabling collaboration between digital public services with ontological value.

Topic: Data exchange

Design guidelines

  • Ensure that data flowing across systems, services, and devices is identified and classified based on types, formats, structures, and sensitivity.

  • Define clear data flow paths between the components realising the digital public service. Specify any transformations or processing steps that occur during data flow.

Conformance testing tips

The data flows foreseen by a solution represent the key focus point of any conformance testing campaign. In GITB TDL test cases these are tested by means of messaging steps, using either built-in capabilities, or custom extensions for protocols without built-in support or where finer control is needed.

You should ensure that all foreseen flows are covered by test cases, including also processes that entail multiple exchanges to ensure conversational consistency. In addition, don’t forget to add “unhappy flow” test cases that purposely make invalid exchanges, to ensure edge cases and error handling are correctly implemented.

Topic: Data modelling

Design guidelines

  • Use standardised semantic data models ensuring models are simplified, reusable, and extensible, capturing the fundamental characteristics of data exchanged by stakeholders.

  • Regularly check and ensure compliance with the adopted semantic data models. Keep the data models updated and maintain them as necessary to reflect changes in standards or requirements.

  • Provide support and training to involved stakeholders to ensure proper implementation of the data models.

Conformance testing tips

Validation of data models is extensively supported by the Test Bed’s validators. These follow a configuration-driven approach whereby validation artifacts are configured over reusable core components covering most popular syntaxes (XML, RDF, JSON and CSV). These are integrated in test cases using verify steps, and can even be supplied with custom artefacts generated during the course of test sessions to validate specific scenarios.

Versioning updates in the semantic standards used can also be tested by updating our test suites’ validation artefacts. When doing so we can create new versions of the test suites and their related specifications, to distinguish tests carried out for earlier releases. This allows us to maintain the systems’ testing history for reference, while having clarity over the testing progress for latest releases.

Topic: Metadata use and management

Design guidelines

  • Use descriptive core metadata elements common to all exchange domains.

  • Define and include additional extension metadata elements specific to particular exchanges.

  • Ensure metadata quality through validation processes. Focus on an automated testing approach to identify and report inconsistencies or errors.

  • Use specialised tools to effectively manage and integrate metadata standards, offering functionalities such as metadata cataloguing, search and discovery, and governance.

  • Establish governance frameworks to ensure metadata accuracy, consistency, and compliance.

  • Foresee automatic discovery and harvesting or metadata from various data sources.

  • Implement a change management process to handle updates and modifications to metadata schemas, and use version control systems to track changes.

Conformance testing tips

Metadata quality control should be a first class concern alongside the quality of data. Metadata would typically be expressed through RDF ontologies that can easily be validated using SHACL shapes via an RDF validator instance. Besides using such validators as standalone tools, they can also be integrated in test cases through verify steps. Moreover, validation in test cases can be extended with custom shapes, generated on-the-fly using templates and data from the current test session. Doing so ensures not only that metadata is correct but also that it matches each specific test scenario.

Besides validating metadata, test cases can also be defined to validate their discovery and harvesting. The relevant APIs and involved exchanges can be simulated by the Test Bed, ensuring that systems under test correctly implement their specifications. This can be achieved through messaging steps and potentially supported by a custom messaging service.

Finally, it is interesting to note automated quality assurance is well supported by the Test Bed. Validators expose machine-to-machine APIs and extensive configuration options making it simple to include them in automated quality control workflows. For more elaborate tests, the GITB Test Bed software itself also exposes a REST API that allows you to deploy test suites and execute tests in a “headless” manner as part of continuous integration processes.

Topic: Ontology design

Design guidelines

  • Ensure the use of established Ontology Design Patterns (ODPs) to promote semantic interoperability and consistency. Adopt ODPs from recognised sources such as the W3C and other standardisation bodies.

  • Conduct regular reviews of the ontologies to ensure ongoing compliance with the adopted ODPs.

  • Establish conformance checks to verify that ontologies adhere to the selected patterns and best practices.

  • Promote continuous improvement by updating ODPs and ontology development processes based on feedback and evolving standards.

  • Use well-established ontology languages to define and instantiate ODPs such as OWL (Web Ontology Language), RDFS (RDF Schema) and SKOS (Simple Knowledge Organisation System).

  • Ensure comprehensive documentation for each ODP including its purpose, structure, usage and reuse guidelines, and implementation workflow.

  • Use specialised tools compatible with standardised formats that include key functionalities (ontology editing, visualisation, reasoning and validation) for the proper and consistent development of ontologies and semantic datasets.

Conformance testing tips

Validating the correct use of ontologies is a key use case for the Test Bed’s RDF validator, making use of the ontology’s foreseen validation artefacts (typically SHACL shapes). Besides validating for an RDF graph’s basic correctness, a validator can also be configured to check best practices and design patterns, both specific to the ontology as well as extensions foreseen for a given solution.

A good example of this is the SEMIC Style Guide validator, that checks ontologies based on the eGovernment Core Vocabularies for alignment with the SEMIC Style Guide. With respect to the SHACL shapes defining conformance rules for an ontology, the Test Bed offers a SHACL shape validator, that can be used to validate the shapes themselves against the SHACL specification and curated best practices. Such validations checks would take place typically as part of the development of the ontologies in question, and could also be replicated in conformance tests to ensure different parties’ alignment.

Topic: Standards compliance

Design guidelines

  • Ensure compliance with recognised semantic standards such as the Core eGovernment Vocabularies to ensure uniformity and interoperability across systems.

  • Ensure compliance with recognised metadata standards such Schema.org.

Conformance testing tips

Testing conformance to such semantic standards is a classic Test Bed use case. Typically all test cases involving data exchanges should include verify steps that validate the data against the expected validation artefacts. This can be done by using the Test Bed’s built-in validation capabilities, or by delegating to custom validator instances.

Interestingly, when dealing with recognised data and metadata standards, there is also a good possibility that the standard is supported by an already existing public validator which can be integrated as-is in test cases. This is possible given that all the Test Bed’s validator instances by default also implement the GITB validation service API that allows them to be used in GITB TDL test cases as the implementation of verify steps. An example of such a validator is the public DCAT-AP validator maintained by the SEMIC team.

With regards to reusing the eGovernment Core Vocabularies in particular, the Test Bed hosts the SEMIC Style Guide validator. This is a validator also maintained by the SEMIC team, allowing ontology developers to test for alignment against the guidelines listed in the SEMIC Style Guide.

Technical governance

Technical governance pertains to the agreements on technical processes and operations, enabling knowledge sharing across Public Administrations to achieve public policy goals.

Service Level Agreements (SLAs)

Service Level Agreements (SLAs) formalise the agreement between a service provider and consumer, outlining expected service levels, as well as the metrics to measure the service’s quality.

Topic: Metrics and reporting

Design guidelines

  • Foresee performance metrics such as response times, throughput, and latency; defining altogether the solution’s expected performance characteristics.

  • Foresee scalability metrics such as load handling and elasticity; defining the solution’s expected behaviour when subjected to increasing or decreasing load.

  • Foresee reliability metrics such as error rates, mean time between failures, and mean time to repair; defining how the solution should handle errors.

  • Establish mechanisms for regular reporting and analysis of established metrics. Regular audits should also be foreseen to ensure continuous alignment.

Conformance testing tips

Expected response times and throughput could be good testing candidates by generating a batch of typical requests for the system, and observing the produced responses to measure processing times and completion within the expected time window. Similarly, SLA metrics pertaining to characteristics of individual requests (e.g. maximum supported payload sizes) could also be tested for by configuring appropriate data for such edge cases. A good approach to realise such tests would be to use a custom messaging or processing service, that the test case would signal to carry out the requests on its behalf. This service would then call the system under test as needed, and report a summary of the calls and observed responses as the relevant test step’s report.

Error and edge cases are also ideal candidates for “unhappy flow” test cases, as here you can benefit from the fact that you are using the Test Bed and not an actual implementation as the counterpart of the system under test. A real-world implementation of the specifications would likely be unable to “behave badly”, something that in contrast is perfectly achievable (and testable) through the Test Bed.

Having said this, keep in mind that metrics linked to load testing would ideally be validated by specialised stress testing tools, and not the Test Bed. In addition, metrics such as availability that relate to observing the system over a period of time, would also be impractical to test.

Topic: Review processes

Design guidelines

  • Specify the frequency of SLA reviews to ensure they remain relevant, effective, and aligned with evolving business needs, technological advancements, and regulatory requirements.

  • Define the procedure for SLA reviews, including stakeholder meetings and feedback collection to identify important service features and appropriate metrics.

  • Describe the SLA change management process, clarifying how amendments are proposed, approved, documented, and planned for implementation.

Conformance testing tips

The Test Bed would naturally not be used to validate how SLAs are reviewed and updated. However, this topic highlights the fact that SLAs evolve over time and when doing so, incur changes on the systems under test. Having defined test cases for the SLA metrics that are practically testable, will allow you to test that implementations conform to the requirements listed therein before SLA changes are brought into force.

Topic: Roles and responsibilities

Design guidelines

  • Identify the roles and responsibilities of all parties involved in the service’s delivery.

  • Define for all involved parties, their respective points of contact and communication channels.

  • Define the escalation procedures for addressing issues or concerns, and the manner with which efficient decision-making can be sought.

Conformance testing tips

Such aspects are more human-oriented and cannot be practically tested via the Test Bed.

Topic: Security and data protection

Design guidelines

  • Ensure data protection by specifying encryption standards, data backup procedures, and data access controls. Include specifications regarding security standards, vulnerabilities, and steps taken to protect digital solutions from breaches or unauthorised access.

  • Ensure compliance to relevant security regulations and standards.

  • Describe incident response procedures for detecting, reporting, and responding to security breaches.

Conformance testing tips

Anything related to the setup of the communication channel with the system under test can be easily tested. In fact using the Test Bed gives us the possibility to manipulate aspects such as encryption algorithms to use ones that are expected to be rejected. Moreover, doing this through a custom messaging service triggered from a test case, gives us the possibility to configure the channel per tested scenario, and in a manner that correctly implemented software would normally not do.

Similarly, requirements pertaining to access control can also be tested by including test cases where invalid authorisation attempts are triggered and are expected to be rejected by the system under test.

A set of tests focusing on selected security aspects like the ones mentioned above, could be defined in what the Test Bed terms a shared test suite. Such test suites can be linked to multiple business-oriented conformance statements as a set of additional tests to be completed at least once before claiming conformance.

Topic: Service availability

Design guidelines

  • Define uptime guarantees as the percentage of time the service is expected to be available.

  • Define clear periods for scheduled maintenance when the service may be unavailable to support maintenance work. Ensure advance notice for users about the scheduled maintenance windows.

  • Define how unplanned outages are managed, and how users are informed about the status and resolution of issues.

Conformance testing tips

Guarantees on uptime and planned or unplanned downtime, are not practically testable through the Test Bed. Such measures are typically listed as percentage values over time and would require constant monitoring of the system under test over an extended period.

Technical interoperability agreements

Technical interoperability agreements formalise the technical specifications for organisations, to ensure consistent operations under varying technical frameworks.

Topic: Joint processes improvement

Design guidelines

  • Ensure all parties agree to adhere to governance frameworks serving as tools to enhance processes, ensuring efficiency, effectiveness, and alignment with industry standards and best practices.

  • Ensure commitment to a continuous evaluation and enhancement of processes, including periodic reviews to assess the effectiveness of adopted frameworks and their adaptation for ongoing alignment with organisational goals and industry best practices.

Conformance testing tips

Such governance and process-oriented points cannot be practically tested via the Test Bed.

Topic: Security and trust

Design guidelines

  • Ensure all parties agree to ensure compliance with established security principles and best practices for secure data transmission, irrespective of specific security policies.

  • Enable trust verification by securing communications using certificates issued by trusted Certificate Authorities (CAs).

  • Under no circumstances tolerate the use of self-signed certificates of certificates issued from untrusted CAs.

  • The configuration of Transport-Layer Security (TLS) should be kept up-to-date with latest encryption and security standards, and never be downgraded to accommodate parties technical deficiencies.

  • Ensure the continuous monitoring and evaluation of all components to ensure ongoing compliance with established security principles.

Conformance testing tips

Testing of certificate trust and correct TLS configuration are perfectly achievable through the Test Bed. This can be done through test cases that will initiate communications using on purpose self-signed certificates or obsolete encryption protocols, expecting the communication to be rejected by the system under test.

Similarly, further tests cases can use valid certificates and focus on trust issues in exchanged payloads. A typical test would be to have a test case produce a valid message, sign it with a trusted certificate, but then adapt the message before being sent. The test case would then include verify steps as assertions to make sure that the system under test detected the manipulation and responded with the expected error message.

The simplest way to achieve such tests would be through a custom messaging service that will use an invalid configuration when requested to do so by the calling test case. This could be signalled to the service by means of a special input flag of the relevant send or receive step.

Topic: Technical collaboration

Design guidelines

  • Use agreements to establish collaborative frameworks between organisations to facilitate the exchange of knowledge, expertise, and resources.

  • Align priorities and coordination activities to prevent duplication of effort and maximise the impact of standardisation initiatives.

  • Include measures to enhance the traceability of adopted standards, ensuring alignment with international best practices and regulatory requirements.

Conformance testing tips

Such points pertaining to collaborative agreements cannot be practically tested via the Test Bed.

Summary

Congratulations! Having completed this guide you should now have a better awareness of governance aspects, and how to design involved agreements to enable interoperability. Besides this, you should now also be aware that such agreements can provide a good source for conformance test cases, and that doing so will increase your testing coverage beyond only your project’s business processes. Finally, the guide highlighted typical points you can practically test for, and initial ideas on how this can be achieved.

See also

If you are interested in more details on the background and foundation of this guide you can:

  • Check the EIRA© for more information and other EIRA© solutions not discussed here.

  • Check the CAMSS for more details on its assessments and related solutions.

  • Explore yourself the standards and specifications included in the ELIS.

Regarding conformance testing, if you are unfamiliar with the Test Bed you can start from its Joinup space for a first introduction. You can then follow-up with the Test Bed’s value proposition that presents its key services and use cases.

For more technical information on conformance testing your starting point should be the Test Bed’s developer onboarding guide. From here you can then check the more detailed documentation that was also referenced frequently during this guide:

At several points this guide also mentioned validators, components that can used as standalone services and as test case validation steps. If you would like to know more and see how to create such validators you can check: