Abstract
Keywords
Introduction
The increased availability of data, in particular data collected routinely, provides a valuable opportunity for analysis with a view to supporting evidence based decision making. Analysing available data in support of a research question or hypothesis necessitates some statistical approach to ascertain whether any effect observed in the data is due to chance or not. In a simple example one may have access to data on the number of patients admitted or the number of students passing a test, and one may wish to answer the question: “Is there an increase between last year and this year?” Simply comparing the number of patients admitted each year or comparing the pass percentage of students on its own does not provide an answer to the question asked. A difference in the values, however large or small, may be due to random variations and as such provides no confident evidence that there is indeed a difference between the years.
It is at this point that the statistical model selection problem arises. Each of these situations can be supported through the use of an appropriate statistical model or approach. In some cases there will be more than one possible approach to statistically test the research question or hypothesis. A human statistician is trained and experienced at considering all the factors relevant to a specific research question and data in order to recommended an analysis model or approach.
As there is an ever increasing amount of data available to test hypotheses on, coupled with easily usable statistical functionality being offered in standard (off-the-shelf) spreadsheet products there is a need to offer a method to guide, support and justify the selection of one statistical approach over another. This can be done by consulting a statistician, but this involves additional effort and time. An automated recommendation and justification shortens the time required.
The choice of model in support of a research question involves knowledge on what model approaches achieve the type of objective specified by the research question and by the available data. Additionally, statistical theory dictates conditions under which models cannot be used and conditions under which models may not perform at their best. The former are known as critical assumptions and the latter are context domains. The extent to which these conditions are met can be tested either by querying the data or by eliciting information about the data from the domain expert. Utilising a model when its critical assumptions don’t hold can lead to erroneous conclusions being drawn, or effects being missed.
In previous work [23,24] we proposed an argumentation-based approach to providing decision support to a user when deciding which statistical model or approach is best suited to their research question and data. In this paper we build on our previous work by articulating a formalisation of the argumentation schemes, critical questions and knowledge base required to support the instantiation of an Extended Argumentation Framework (EAF) [20]. The application of EAFs and their instantiation in the context of statistical model selection is a novel approach. We also outline Z notation schemes [26] to describe this formalisation and outline the prototype development that was based on these Z schemes. To the best of our knowledge this approach, although used in Multi Agent Systems has not previously been applied to argumentation.
An initial step towards evaluating the contributions outlined in this paper has been achieved through the use of case studies, one of which is included herein. The next step in evaluation will involve user studies aimed both at ensuring the reasoning and recommendations made by implementing our approach is consistent with the recommendations a statistician would make, and to asses the end user experience.
The paper is structured as follows: Section 2 provides relevant background and introduces a motivating example. Section 3 formalises the argument schemes, knowledge base and critical questions. It illustrates the formalisms through the use of an example and also provides Z notation schemas. Section 4 describes the prototype implemented. Section 5 discusses related work. Section 6 provides a summary of the work proposed and articulates our plans for further research.
Background
In order to illustrate the process of selecting an appropriate statistical model we will be introducing an example based on a freely available data set. The data set is called
There is a difference in patient survival (
In order to test Hypothesis 1 there is a need to select and apply a statistical model. The target attribute for the hypothesis is survival time contained in
The presence of a significant difference in survival times between different groups of patients can be formally tested using
Prior to applying the
The other common Survival Analysis method is
There are some assumptions that must be met prior to the use of
A third approach to consider is
There are many additional models for survival analysis but for the purpose of this example we only chose to consider three. Of the three models considered in this example there were two models that were possible (
It is also possible to assess whether one model is more suitable that the other one by taking into account some additional conditions. A statistician may assess the situation by discussing the objective of the analysis with the clinician. If the objective of the analysis extends beyond testing Hypothesis 1 but also looks to produce a benchmark to be used to estimate survival time for future patients then the recommendation would be to use
Z notation [26] is a formalism created to represent interacting computational systems and it is based on elementary components such as set theory and first order predicate logic. There are examples of the use of Z notation to formalise multi agent systems ([6,17,19]). Luck
An additional implementation in a multi agent setting that makes use of Z is provided by D’inverno
We have expressed our argumentation-based method in a formal language in order to make the specification precise, and to aid its implementation. The expressiveness and accessibility of Z notation coupled with the need to facilitate a prototype implementation justified our use of it to formalise the contributions in this paper.
Method
In this section we articulate the elements that comprise our method to support statistical model selection through the use of Extended Argumentation Frameworks (EAF). We also provide the Z notation schemes representation of the elements introduced.
The statistical knowledge base
Our method relies on a statistical knowledge base (SKB) which includes all of the relations between the objectives of the research question or hypothesis (
The relations and contents of the SKB are derived from statistical theory and best practice, these relations are defined by an expert, not the end user. An example of the elements of the SKB is illustrated in Fig. 1, this is pertinent to the

The knowledge base contents relevant to the example.
The set of The set of The set of
The following relationships are defined in the SKB:
(Relations in the statistical knowledge base).
The Z notation basic types required to define the elements of the SKB (Definition 1) are:
[MODEL] – the set of all possible models
[OBJECTIVE] – the set of all possible objectives
[ASSUMPTION] – the set of all possible assumptions
There is also a need to strengthen the specification by setting up variables to account for potential input errors. Z schemas can be implemented (such as REPORT) so that when used in conjunction with the other schemas errors can be flagged. The state space for the SKB is:
The SKB contents relevant to the example introduced are in Fig. 1 and the relations as per Definitions 1 and 2 are:
The SKB contents in Z notation are:
Argument schemes and critical questions
When the
((AS1): Argument Scheme for model to consider on grounds of achieving the objective –
).
The premises for this AS1 (Definition 3) are statements that are verified against the SKB, given the specific
where the type REPORT will be defined to flag situations where the objective stated is not known.
AS1 (Definition 3) is subject to critical questions. These are used to test the assumptions of the scheme (such as CQ2) for potential undercuts or to highlight exceptions (CQ1) or rebuttals. The CQs identified and their respective argument schemes are:
CQ1: Are there alternative ways of answering
CQ2: Do any of the critical assumptions for
((AS2): CQ1: Argument for alternative objective:
).
((AS3): CQ2: Argument against the use of a Model for failed critical assumptions:
).
The premises for CQ1 (Definition 4) are statements extracted from the SKB once the initial research objective
Given the list of models, the critical questions need to be instantiated.
In order to set this relation up the schema is initialised:
The schema
The [AddAlternativeObjective] schema adds a relation between one objective (
Now the argument scheme AS2 in support of CQ1 (Definition 4) is instantiated through the following schema:
The aim of CQ2 (Definition 5) is to validate the critical assumptions for each of the models returned as part of
The resulting list of assumptions contain elements of the type:
Instantiation of the argument scheme and critical questions for the example
In order to illustrate the application of the ASs and CQs the example introduced in Section 2 is used. The objective of the research question is Survival Analysis
Table Analysis Fisher’s Exact non informative censoring proportional hazards table cell minimum
The contents of the SKB were outlined earlier in this section. The critical assumptions relevant to this example are:
Instantiating AS1 (Definition 3) in this example leads to the following:
Three arguments have been instantiated, these now need to be subject to the two critical questions. Instantiating CQ1 using AS2 (Definition 4) generates the following argument:
This results in the following set of arguments: {
Instantiating CQ2 using AS3 (Definition 5):
The instantiation of AS3 (Definition 5) generates two undercuts to two of the arguments in favour of the use of the respective models.
AS3 (Definition 5) can generate an undercut to AS1 (Definition 3) thereby undercutting some arguments in support of models from the set of ones that could be applied in this example. The resulting AF is illustrated in Fig. 2 and its arguments are

Argumentation Framework for the
The instantiation of all of these argument schemes (Definitions: 3, 4, 5) will produce a set of arguments in support of or against the use of a specific model. This set of arguments make up the argumentation framework (AF) of relevant arguments to the research question and data at hand. The attack relations within this AF are relevant and derived from the desire to run only the most appropriate models, implying that a decision to use one model (with arguments in support of its use) negates the use of other models with arguments supporting their use in the AF.
If it is acceptable to run all the models that have an argument supporting their use then the AF contains no attack relations per se. If it is desirable to run only the most suitable models the relative strength or preference of each argument in support of the use of a model should be considered. In order to pick the most suitable of the models, this approach is extended with a means to express and reason with preferences.
We identified three different sources of preferences in the context of statistical model selection and we formalise these preferences through an extended SKB and EAFs.
A preference expressed in the context of statistical model selection refers to an order of priority between models. If we have a set of models
The source of the preference orders are mapped to a priority or importance when leveraging the preferences to find the model most suitable to the situation.
To incorporate preferences into the approach, the SKB introduced in Definition 1 is extended:
(Extended statistical knowledge base).
The mapping of the model to the performance measure
In order to extend the SKB as per (Definition 6) context domain and performance measure are added to the specification.The state space for the extension of the SKB is defined as follows:
In order to populate this extended knowledge base, firstly the context domains are defined:
For each given context domain (defined with the previous two schemas) the models relevant to it can be added and mapped onto the existing defined performance measures. The schema to map a model to a context domain and performance measure is defined as follows:
The
The
Context Domains Performance Measure Mapping for (a)
light censoring and (b)
model intent explain
Context Domains Performance Measure Mapping for (a)
The definition of the context domain that is derived from clinician preference is also mapped using the Z schemas
To construct an EAF based on the extended SKB, first the set of contexts
Let
(Generating the EAF using the extended statistical knowledge base).
Intuitively, an attack relationship
Within an EAF that includes the preference arguments from one context domain the acceptable arguments supporting the use of a model to the preferred extension semantics provide justification to the suitability of the model. These preferred or more suitable models are the ones supported by an argument that is acceptable with respect to the set of preference arguments from the context domain in question. If the arguments in support of the use of more than one model are acceptable to the set of preference arguments generated by one context domain, another context domain (the next one in order of importance) can be introduced.
Instantiating the extended argumentation framework for the example
Table 1(a) contains a subset of the
In this situation we have an AF containing three arguments each supporting the use of a different model, this is illustrated in Fig. 2. If the overall aim is to recommend one model or aim to refine the list of models to apply then the AF is extended to include and account for the information that can assist in arguing what contexts are relevant to the situation and leverage them in order to recommend the most suitable model(s).
The AF from the
In order to generate the EAF for this scenario then the following additional inputs are required in order to generate the preference meta level arguments
The
The second context domain of relevance in this example is
Finally there is a clinician expressed preference, this will be
There are now seven meta level arguments derived from the preferences in

EAF for

EAF for
The most important context domain in this example is derived from censoring (
If we assume that the order over the context domains in
The formalisation and the Z notation schemes presented in this paper formed the basis for the development of the Small Data Analyst prototype [30]. This was developed and deployed on

The user interaction to confirm an assumption for a model.
Once all the assumptions are tested, and the context domains identified the model recommendations are presented both in a list and as an EAF. In Fig. 6 the models considered are

The model recommendation made in case of the

Graphical display of the resulting EAF to user (optional view for end user).
The administrator has access to functionality to populate the Extended SKB (Fig. 8(a)) and to modify the model definitions, such as assumptions required for a model (Fig. 8(b)).

Administrator screens for editing the Extended SKB.
Outside the context of statistical model selection there are examples of the application of argumentation to decision support, Fox
EIRA (Explaining, Inferencing, and Reasoning about Anomalies) is an argumentation-based clinical decision support system designed to flag anomalies in patients’ reactions to medication within the Intensive Care Unit [11,12]. Grando
A further example of the use of argumentation schemes and knowledge bases in support of clinical decision support is provided by Atkinson
Another decision support system that leverages argumentation is in organ transplant allocation mechanisms. There is a known shortage of viable organs, therefore the allocation process needs to be as efficient as possible. In their papers [27,28] Tolchinsky
A related challenge in the clinical domain is reasoning with all the available evidence on treatment outcomes. In [14] Hunter
Automation of model selection has recently been central to the “Automatic Statistician” project [16] where a different approach to automation of model selection is proposed, compared to the work articulated herein. The approach and type of analysis tackled by the “Automatic Statistician” is different from the one this paper is focusing on. Lloyd
The evaluation of methodologies and prototypes leveraging argumentation for clinical decision support has been through case studies and user studies. ArguEIRA [12] was evaluated by clinicians assessing the tool’s output, CARREL [28] was similarly evaluated on a set of examples as well as DRAMA [2,3] where examples were also used to ensure the proposed argument scheme and knowledge base were comprehensive enough. The initial evaluation approach we took is similar in nature as it is initially case study based.
Conclusion and future work
This paper reports on an application of argumentation theory to the analysis of clinical data. Our contributions presented in this paper are a formalisation of the argument scheme and its associated critical questions for this domain, and an extended knowledge base containing preference orders for the models that enable the instantiation of an Extended Argumentation Framework (EAF). We also present an implementation of these formalisations in Z notation. These elements offer a novel approach to supporting the automation of the process of statistical model selection. The application of the method we have proposed herein supports an end user by suggesting the recommended statistical model to use given their specific research question and the data available.
Our application of EAFs as well as the formalisation of the argument schemes in Z notation sets this work apart from all the work cited. Our approach provides an example of an application of argumentation and preferences with a prototype application in the clinical domain. The use of Z notation to bridge the gap between the definitions and the implementation has enabled all of the specifications required of the system to be articulated, furthermore the strength of Z notation as a step between definitions and implementation has enabled the introduction of variables to account for potential errors. The advantage of using EAFs is their support for reasoning that leverages different sets of preferences through their representation as meta level arguments. This enables the reasoning to argue at the preference as well as the argument level. In future work we will investigate the benefit of exploring a meta-level argumentation representation [21].
The initial steps in evaluation of the method proposed herein were achieved through the use of case studies from the clinical domain, one of which is articulated in this paper. There is a requirement for further evaluation through user studies. The first of which will assess whether the outlined method will provide the same recommended model and justification when compared to what a statistician would recommend based on the same data, research question and available models. This will then be followed up by a user study where the prototype will be used by clinicians. The latter would enable us to ascertain whether this is usable and acceptable to the end user. A further aspect we will be researching is how to best present the results of the EAF to the end user. These evaluation steps are ongoing work to be reported in future publications, more detailed plans are outlined in [22].
