---
tags:
- report section
title: "Section 1: Trustworthy Digital Mental Healthcare"
authors:
- Christopher Burr
draft: true
section: 1
---
# Section 1: Trustworthy Digital Mental Healthcare

:::info
**Section Overview**
This section introduces a model and procedure for enabling a more trustworthy ecosystem of digital mental healthcare through the responsible and ethical design, development, and deployment of digital technologies. It serves as an introduction for the analysis and recommendations for the following sections.
First, we introduce a model of a typical research or innovation lifecycle for a data science or AI project comprising activities of project design, model development, and system deployment.
Second, we discuss the methodology of trustworthy assurance that is at the centre of our project. We provide a simple overview of the methodology, focusing on the structure, elements, and purpose of an assurance case.
Finally, we look at argument patterns: reusable templates that can be developed to ensure a more consistent approach to trustworthy design, development, and deployment of digital mental healthcare.
:::
## Designing, Developing, and Deploying Trustworthy Digital Mental Health Technologies
Designing, developing, and deploying an AI system is not a one-person task! The stages and activities that comprise a typical AI project lifecycle involve a wide-ranging and far-reaching set of skills and capabilities. These skills are encapsulated within myriad roles, including project commissioner, product manager, data protection officer, data scientist, system architect and engineer. And, these roles are interwoven so that they create an irreducible and collective responsibility across the project lifecycle—a responsibility that may span multiple teams and organisations.
Figure 1 presents a simplified model of a typical research or innovation lifecycle for a data science or AI project lifecycle, to help gain a purchase on the interweaving roles, skills, and responsibilities. The model represents three over-arching stages of (project) design, (model) development, and (system) deployment. For each stage, there are corresponding activities, detailed in Box 1. The project lifecycle is depicted as a circular process to highlight the fact that responsibility is ongoing and does not end once a system has been implemented or put into deployment. Responsible research and innovation, by contrast, requires consideration of how a technological system may need to be removed and replaced once it reaches the end of its lifecycle.

*Figure 1: a model of a typical project lifecycle for a data-driven technology, detailing the overarching stages of design, development, and deployment and their constitutive activities (reprinted from Burr and Leslie, 2022)*
**Table 1.1: Project Lifecycle Activities**
| Stage | Description |
| --- | --- |
| **Project Planning** | Preliminary activities designed to help scope out the aims, objectives, and processes involved with the project, including potential risks and benefits. |
| **Problem Formulation** | The formulation of a clear statement about the over-arching problem the system or project addresses (e.g., a research statement or system specification) and a lower-level description of the computational procedure that instantiates it. |
| **Data Extraction or Procurement** | The design of an experimental method or decisions about data gathering and collection, based on the planning and problem formulation from the previous steps. |
| **Data Analysis** | Stages of exploratory and confirmatory data analysis designed to help researchers or developers identify relevant associations between input variables and target variables. |
| **Preprocessing and Feature Engineering** | A process of cleaning, normalising, and refactoring data into the features that will be used in model training and testing, as well as the features that may be used in the final system. |
| **Model Selection and Training** | The selection of a particular algorithm (or multiple algorithms) for training the model. |
| **Model Testing and Validation** | Testing the model against a variety of metrics, which may include those that assess how accurate a model is for different sub-groups of a population. This is important where issues of fairness or equality may arise. |
| **Model Documentation** | A process of documenting both the formal and non-formal properties of both the model and the processes by which it was developed (e.g., source of data, algorithms used, evaluation metrics). |
| **System Implementation** | The process of implementing the technological system into its intended environment or target domain to enable and structure interaction with the underlying model(s) (e.g. a recommender system that suggests possible treatment options for patients based on input data. |
| **User Training** | Training for those individuals or groups who are either required to operate a data-driven system (perhaps in a safety critical context) or who are likely to use the system (e.g. healthcare professionals, medical researchers). |
| **System Use and Monitoring** | Ongoing monitoring and feedback from the system, either automated or probed, to ensure that issues such as model drift have not affected performance or resulted in harms to individuals or groups.|
| **Model Updating or Deprovisioning** | An algorithmic model that that adapts its behaviour over time or context may require updating. Where no further updating can be carried out, and this results in a system being removed from production (i.e. deprovisioned), a new system may be required. This restarts the project lifecycle. |
To see how this model can help us understand the interwoven nature of responsible research and innovation, consider the following example. An organisation wants to implement a speech recognition algorithm within a service they are developing for online counselling. However, there is no one in the organisation with the relevant expertise to collect data and train a model from scratch. Therefore, they choose to procure a pre-trained model from [another company](https://aws.amazon.com/marketplace/solutions/machine-learning/pre-trained-models). This means that a significant portion of the project lifecycle—from `Data Extraction or Procurement` to `Model Documentation`—will have been carried out by a separate organisation.
Although the specifics of the relationship between the two organisations will complicate forms of responsibility, such as legal duties or obligations, this need not concern us here. Instead, we can focus on how the initial organisation who has chosen to undertake the project (e.g. the product owner or commissioner) can use the project lifecycle model to a) identify and analyse their own responsibility and how it intersects with the responsibilities of others, and b) how this supports a process of trustworthy assurance.
### Reflective and Anticipatory Deliberation
At the start of a project, while activities such as planning and initial evaluation of feasibility are being conducted, the project lifecycle model can be used to structure reflective and anticipatory processes of deliberation among the project team. For instance, the team could use the model to identify potential actions and decisions that are likely to emerge during specific activities, such as which data types may be required and whether stakeholders or users will consent to these data being collected and analysed. As this example suggests, the project team may carry out the preliminary deliberation, but additional stakeholders will need to be engaged to thoroughly evaluate the ethical, legal, and social permissibility and acceptability of the project.
Consider another example. A company have identified a risk associated with an AI system they have developed, which they claim is able to detect emotions. They have been approached by a healthcare provider who wish to implement the system into a video consultation service to help counsellors better understand the affective responses of their patients during initial assessment. However, the developers did not evaluate their model (during `Model Testing or Validation`) using an dataset that was representative of the patient population intended by the healthcare provider (i.e. individuals that are likely to be suffering from a mental health issue). Therefore, the company is unable to make any claims about the generalisability of their model to this new population. Moreover, neither the company nor the healthcare provider have engaged the relevant stakeholder groups during `Project Planning` to determine if this would be an acceptable use for their system. As such, additional activities would need to be carried out to determine the full scope of the risks and possible harms that could arise from the use of this technology. This would likely require the procuring organisation (i.e. the healthcare provider) to set clear requirements for what forms of evidence would be required from the developers (e.g., at `Model Documentation`), and to determine clear boundaries and thresholds for whether the project should proceed.
As this example illustrates, the project lifecycle structure can help support forms of reflective and anticipatory deliberation that help instantiate a responsible ecosystem of research and innovation. However, there is a secondary purpose for the project lifecycle model that will become clearer in the next sections: the identification of actions and decisions that generate forms of evidence that provide justificatory support for trustworthy assurance.
## What is Trustworthy Assurance?
> Trustworthy assurance is a procedure for developing a structured argument, which provides reviewable (and contestable) assurance that a set of claims about the ethical properties of a data-driven technology are warranted given the available evidence.
This definition captures three important and interlocking components of trustworthy assurance:
1. A structured *argument* comprising linked claims and evidence that collectively justify a top-level goal
2. A *procedure* for developing an assurance case, which represents the argument either formally and/or visually
3. Agreed upon *standards* for reviewing and evaluating the argument

*Figure 1.2: a schematic showing the three interlocking components that support trustworthy assurance.*
### Argument
Trustworthy assurance is a form of *argument-based assurance*. It uses a structured argument, known as an assurance case, as the primary means for providing assurance that a goal has been obtained, based on the claims and evidence presented.
There are three basic elements of an argument:
1. A claim about the goal to be established (supported by descriptions of the system and the context in which the system is intended to operate)
2. The claims that justify the goal
3. The evidence that establishes the validity of the claims
<!-- Optional box about the additional properties and elements of GSN (e.g. syntax, context and system descriptions) -->
For example, the following figure constitutes a simple argument.

*Figure 1.3: a portion of a simplified assurance case on respect for privacy*
However, this example is too simple to constitute a full-fledged assurance case. This is because it reduces the concept of 'respect for user privacy' to a single principle (i.e., data minimisation), which although relevant is insufficient on its own. We can, for example, consider an app that collects no personal data but still violates reasonable expectations of privacy by routinely notifying and disturbing users.
A procedure is, therefore, required for identifying relevant claims that a) specify the scope of a top-level goal and b) collectively justify the goal.
### Procedure
The procedure adopted in trustworthy assurance is the anticipatory and deliberative exercise [introduced above](#reflective-and-anticipatory-deliberation), which incorporates inclusive and accessible forms of stakeholder engagement.
By using the project lifecycle model as a scaffold for anticipatory reflection and stakeholder-informed deliberation, project teams are able to answer the following questions:
1. Which claims are necessary and sufficient to specify and justify the top-level goal?
2. How do these claims relate to one another?
3. What evidence is required to demonstrate the validity of the claims being made?
Consider the following example. A company wishes to develop an assurance case that shows how their system, which uses a ML algorithm to predict whether users of an online betting platform are "problem gamblers", can generate results that are explainable to their users. They decide this is an important ethical goal for an assurance case, because they want to be able to provide accessible forms of communication to any user that they contact on the basis of their algorithmically-generated prediction.
They start by formulating the following goal statement:
> "An explanation of how our system predicts whether a user is a "problem gambler" can be provided to all users of our platform."
Next, they consider whether any action or decision taken throughout the stages of the project lifecycle is relevant to the specification and justification of this goal. For example, they highlight results from workshops that were carried out with representative users during their `Project Planning` activities, which concluded that plain language explanations were preferable to detailed explanations of the algorithmic techniques, but that users were more trusting of these explanations when they knew they had been independently validated by a professional auditor. This result influenced which machine learning algorithm the team selected during the `Model Selection and Training` activities, and which features they reported on during `Model Documentation` for the independent audit.
Once the team have reflected on all the stages of the project lifecycle, they recognise that there are two broad sets of claims. One set of claims are about the *design choices* made during the project, which support accessible explanations for users. The second set are about the *interpretability of the system*, which are relevant for professional auditing and assessment of the system. Categorising the claims in this manner helps the company determine what evidence will be needed for each claim, and how best to structure the argument.
The following figure summarises the points made in the above example.

*Figure 1.4: a portion of the assurance case for this hypothetical project.*
The final component of trustworthy assurance relates to *standards*.
### Standards
Standards support the development and refinement of best practices and codes of conduct. There are standards for measurement (e.g., universal scales), procedures (e.g., manufacturing), and assessments (e.g., risk and impact assessment), and much more. Here, we are interested in standards as they pertain to evidence and claims.
Evidential standards can refer to both the identification and evaluation of evidence.
Standards for identifying evidence are common in areas such as law where rules exist to determine what constitutes *relevant*, *material*, and *admissible* evidence.
Similarly, standards for evaluating the quality of evidence are well-established in domains such as scientific research, where various procedures or methods are held to produce reliable forms of evidence (e.g. peer review or randomised controlled trials).
As communities of practice develop and emerge, we would expect standards to evolve to help with both the identification and evaluation of evidence. Subsequently, this would help support the development of trustworthy assurance in domains such as digital mental healthcare because specific types of claims or evidence would be recognised as more reliable forms of assurance.
It should be noted, however, that regulators and developers are not starting with a blank slate. There are many relevant standards that exist today, and new standards are emerging to support the procedure of constructing a trustworthy assurance case. We will consider some of these standards in section 3.
## Argument Patterns
### Claims as Reasons
<!-- this section could be shortened because of some repetition with the above content on standards -->
Trustworthy assurance is a process of giving and justifying claims about choices made during the design, development, and deployment of a digital mental health technology.
These claims can be viewed as a series of *reasons* for why a particular decision was made. To see why, let’s assume that an organisation is in the process of procuring an AI-enabled chatbot to provide therapeutic support to service members returning from deployment. As this technology is new and relatively untested, the organisation has a series of questions for the developers:
1. "Why should we license this digital system instead of investing in traditional forms of talk therapy?"
2. "Why have you chosen a female avatar as your virtual assistant?"
3. "How did you measure and validate the clinical efficacy of the system for different subgroups to ensure that it is fair?"
In answering these three questions, the developers would be *giving reasons* (and evidence) for their actions—reasons that would need to be accepted by the procuring organisation to be relevant and justifiable. This perspective on claims emphasises one of their most vital roles as *publicly contestable reasons*. That is, whether a claim (or set of interrelated claims) are valid in the context of an assurance case is, in part, conditional on whether they are accepted as *reasonable* justifications by those who are tasked with evaluating the assurance case.
Let’s look at another example. Consider the following section of an assurance case for the aforementioned chatbot, which the developers have produced for the procuring organisation:

*Figure 1.5: a portion of an assurance case for a chatbot*
[^claim]
[^claim]: Recall, that evidential claims are required to link property claims to their supporting evidential artefacts, and evidential claims, therefore, can also serve as reasons.
As we have discussed, the organisation responsible for evaluating the trustworthiness of the AI system has to determine whether the evidential claim (EC1.1) is a *reasonable* choice to justify its parent claim (SC1.1). They may, for instance, argue that EC1.1. is a *reasonable claim*, but nevertheless argue that it is insufficient on its own to justify the claim that the chatbot is an "accessible" alternative to human-led therapy. Alternatively, they may claim that it is *not reasonable* on the grounds that the ratings given by the service members are not relevant to establishing whether the chatbot is an “accessible form of therapy” but merely that the avatar is “friendly and realistic”. This example highlights a potential challenge associated with the development of assurance cases: determining what constitutes relevant, sufficient, and reasonable claims.
In the context of safety assurance, a large body of guidance has been established to help developers assess what claims they will need to establish and justify, and a key part of this guidance is the development of *argument patterns*.
### What are argument patterns?
Argument patterns are *starting templates* for building assurance cases. They identify the types of claims (or, the sets of reasons) that need to be established to justify the associated top-level normative goal. Figure 1 shows an example argument pattern for the technical goal of interpretability [@ward2020].

*Figure 1.6: a pattern for an interpretability case (reprinted from Ward and Habli, 2020)*
The pattern depicted in Figure 1.6 shows a template for an assurance case that serves to justify the following top-level goal:
> "The {ML Model} is sufficiently {interpretable} in the intended {context}".
Here, the curly brackets serve as placeholders for specific variables that are properly established when a full assurance case is developed. A notable contribution of this pattern is the identification of three essential aspects of interpretability:
1. Right Method - The right interpretability methods are implemented, i.e. the correct information is faithfully being explained.
2. Right Context:
- Time - Interpretations produced at the appropriate times.
- Setting - Interpretations are available in the right setting.
- Audience - Interpretations produced for the right audience.
3. Right Format - The interpretability methods are presented in the right format for the audience.
These three essential aspects subsequently serve to delineate the more detailed argument and evidential claims at the lowest levels.
Argument patterns, such as the one above, are helpful for the following reasons:
1. They provide a consistent and systematic approach for the reflective and deliberative activities carried out across a project’s lifecycle.
2. They speed up the process of developing assurance cases.
3. They provide reusable structures that, if used widely throughout a domain, could establish best practices.
But where do they come from?
### Generalisable Patterns
In the case of the argument pattern from Figure 1.6, this pattern was proposed by the authors as a means to address a gap in the safety assurance of ML systems. As experts in their domain, and as a peer-reviewed contribution, this is a valid source for an argument pattern. However, an alternative (though not entirely disconnected) source is to extract generalisable structures and patterns from existing assurance cases. This is the method that we have explored in the current project on Digital Mental Healthcare and subsequently propose as a procedure for Trustworthy Assurance.
Much like ML algorithms, humans have remarkably effective (but biased) pattern recognition capabilities. This capability may be said to be a key contributor to our internalisation of ethical, legal, and social norms. As an example, James W. Nickel says of human rights:
> "We can think of the emergence of a human right as the coming together of the *recognition* of a problem; the belief that the problem, is very severe; and optimism about the possibility of addressing it through social and political action at national and international levels." [@nickel2007, emphasis added]
Similarly, we can think of ethical and social norms as the shared recognition and subsequent externalisation of beliefs and attitudes towards events as diverse as acceptable etiquette during a dinner party through to permissible forms of punishment for various transgressions. This understanding of the emergence of norms is crucial to the ensuring legitimacy for trustworthy argument patterns. In terms of the emergence of patterns, the three elements that we have explored already are, again, important: the top-level normative goal, the property claims, and the evidential support. Let's take each of these in turn.
The phrase 'trustworthy assurance' creates a wide scope for top-level goals that may be deemed relevant to establishing trust (e.g. *sustainable* digital platforms, *accountable* methods of data governance, *fair* classifiers, and *explainable* decision support systems). As trustworthy assurance cases are developed for data-driven technologies, it is likely that we will see certain goals emphasised (and re-emphasised) over others. In turn, these normative goals will orient other projects and help cultivate best practices. In [related work](https://rm.coe.int/huderaf-coe-final-1-2752-6741-5300-v-1/1680a3f688), we have proposed a series of ethical principles that have been developed to provide actionable insights and safeguards on responsible research and innovation in data science and AI. They are known as the SAFE-D principles:
- Sustainability
- Accountability
- Fairness
- Explainability
- Data (Quality, Integrity, Protection and Privacy)
These principles have been refined and validated in a wide range of domains, and were originally based on a broad understanding of the typical harms and benefits associated with data-driven technology. Therefore, unlike alternative frameworks they are tailored to the specific needs and challenges of responsible, trustworthy, and ethical data science and AI, rather than, say, importing or revising existing frameworks such as biomedical ethics [cf. @floridi2019]. However, they were designed to be domain-neutral starting points. Therefore, we did not presume that these principles would capture the ethical, social, and legal values that are dominant in digital mental healthcare. Instead, this project undertook a process of exploratory engagement and participatory design to identify and develop a series of ethical values and principles that were relevant to the specific context of trustworthy digital mental healthcare. We will return to this point in Section 3 where we analyse our findings from the project's workshops.
Turning now to the property claims and supporting evidence, as assurance cases are communicated for specific goals we will likely see *sets of property claims and supporting evidence* used more than others as justifiable and accepted reasons for establishing the respective goal. For instance, as developers focus on goals such as 'accountability' core attributes of the system and project are likely to be emphasised as relevant targets (e.g. constructing traceable data pipelines, establishing mechanisms to support auditing processes, ensuring accessible documentation).
Returning to the SAFE-D principles once more, we have previously developed a set of core attributes for each of the principles, which a) identify the types of properties that need to be established in a project or a system to ensure the relevant goal is obtained and b) the stages of the project lifecycle where actions can be taken to implement the respective property. Table 1.2 shows an example of the core attributes for 'sustainability'.
| Core Attribute | Description |
| --- | --- |
| Safety | Safety is core to sustainability but goes beyond the mere operational safety of the system. It also includes an understanding of the long-term use context and impact of the system, and the resources needed to ensure the system continues to operate safely over time within its environment (i.e. is sustainable). For instance, safety may depend upon sufficient change monitoring processes that establish whether there has been any substantive drift in the underlying data distributions or social operating environment. Or, it could also involve engaging and involving users and stakeholders in the design and assessment of AI systems that could impact their human rights and fundamental freedoms. |
| Security | Security encompasses the protection of several operational dimensions of an AI system when confronted with possible adversarial attack. A secure system is capable of maintaining the integrity of its constitutive information. This includes protecting its architecture from the unauthorised modification or damage of any of its component parts. A secure system also remains continuously functional and accessible to its authorised users and keeps confidential and private information secure even under hostile or adversarial conditions. |
| Robustness | The objective of robustness can be thought of as the goal that an AI system functions reliably and accurately under harsh or uncertain conditions. These conditions may include adversarial intervention, implementer error, or skewed goal-execution by an automated learner (in reinforcement learning applications). The measure of robustness is, therefore, the strength of a system’s functional integrity and the soundness of its operation in response to difficult conditions, adversarial attacks, perturbations, data poisoning, or undesirable reinforcement learning behaviour. |
| Reliability | The objective of reliability is that an AI system behaves exactly as its designers intended and anticipated. A reliable system adheres to the specifications it was programmed to carry out. Reliability is therefore a measure of consistency and can establish confidence in the safety of a system based upon the dependability with which it conforms to its intended functionality. |
| Accuracy and Performance | The accuracy of a model is the proportion of examples for which it generates a correct output. This performance measure is also sometimes characterised conversely as an error rate or the fraction of cases for which the model produces an incorrect output. Specifying a reasonable performance level for the system may also require refining or exchanging the measure of accuracy. For instance, if certain errors are more significant or costly than others, a metric for total cost can be integrated into the model so that the cost of one class of errors can be weighed against that of another. |
*Table 1.2: a summary of the core attributes for the principle 'sustainability'*
Again, we are not proposing that these principles and core attributes be adopted in digital mental healthcare as the respective goals, claims, and evidence. However, they could provide a starting point for the refinement of domain-specific principles while argument patterns emerge and become crystallised. This process is (loosely) derived from the idea of 'reflective equilibrium', made famous by the political philosopher John Rawls [@rawls1971]. In short, the phrase 'reflective equilibrium' refers to a state of coherence among moral beliefs and attitudes, which emerges over time as a result of public deliberation and consensus building activities that focus on the relevant moral beliefs and attitudes (e.g. notions of 'justice'). Figure 1.7 offers a simple graphic to help visualise the rough idea as it relates to trustworthy assurance.
<!-- Nerea will need to re-design this graphic -->

*Figure 1.7: process of consensus formation for ethical principles as constraints on trustworthy assurance*
These preliminary remarks about trustworthy assurance serve as a foundation for understanding and contextualising our project's research and the recommendations we derive from our findings.
> Additional guidance on trustworthy assurance can be found in a more detailed article introducing the methodology.
<!-- add reference to paper -->