Skip to main content

Hybrid architecture based intelligent diagnosis assistant for GP

Abstract

As the first point of contact for patients, General Practitioners (GPs) play a crucial role in the National Health Service (NHS). An accurate primary diagnosis from the GP can alleviate the burden on specialists and reduce the time needed to re-confirm the patient’s condition, allowing for more efficient further examinations. However, GPs have broad but less specialized knowledge, which limits the accuracy of their diagnosis. Therefore, it is imperative to introduce an intelligent system to assist GPs in making decisions. This paper introduces two data augmentation methods, the Complaint Symptoms Integration Method and Symptom Dot Separating Method, to integrate essential information into the Integration dataset. Additionally, it proposes a hybrid architecture that fuses the features of words from different representation spaces. Experiments demonstrate that, compared to commonly used pre-trained attention-based models, our hybrid architecture delivers the best classification performance for four common neurological diseases on the enhanced Integration dataset. For example, the classification accuracy of the BERT+CNN hybrid architecture is 0.897, which is a 5.1% improvement over both BERT and CNN with 0.846. Finally, this paper develops an AI diagnosis assistant web application that leverages the superior performance of this architecture to help GPs complete primary diagnosis efficiently and accurately.

Peer Review reports

Background

In many healthcare systems, including the United Kingdom and Ireland, General Practitioners (GPs) are the first point of contact for patients and most general problems are addressed in primary care [1]. GP plays a central role in ensuring that the patient receives a timely and preliminary diagnosis. However, if GP is unable to make decisions, the patient will be referred to a specialist.

For a patient eligible for referral, GP will write a referral letter to a specialist, outlining the reason why this patient is being referred, as well as the patient’s past medical history, including medications, allergies and any other relevant information. After 2 weeks (urgent situations) or 18 weeks (non-urgent) wait, an appointment with a hospital specialist will be offered to the patient.

With the outbreak of COVID-19, the NHS is under extreme pressure. Hospital Episode Statistics (HES) data shows that the number of unsuccessful GP referrals has jumped from 238,859 in February 2020 to a staggering 401,115 in November 2021 (an 87% increase), because there is no capacity in secondary care, and those referrals are rejected. During the specialist consultation process, almost two-thirds of the time is spent on confirming the contents of the referral letter and doing basic examinations, then they upgrade to the professional tests and make a decision. If this initial confirmation and basic examination process are completed before the specialist consultation, this process will be accelerated which can save a lot of time and more patients will have a chance to get health care. GP is the most suitable role to take over this work. As the first point of contact for patients, GP has the best understanding of the patient’s situation. Because GP has a broader, less specialized knowledge base, the accuracy of their primary diagnosis is limited. If there is an assistant for the GP to improve the accuracy of their primary diagnosis, it will be helpful to relieve the pressure on the specialist. A lot of research has been devoted to developing AI-assisted diagnosis methods [2,3,4,5] or tools (i.e., Symptomate, Patient, HealthTap, Ada and etc.). However, they are mainly facing the patient, and the connection between GP and specialists is seldom considered.

Therefore, the purpose of the paper is to develop an intelligent GP assistant which can improve the accuracy of the GP’s primary diagnosis. As the referral letter includes abundant information about the patient’s past medical history including medications, allergies and any other relevant information. The proposed system takes the complaints in the referral letter and the final diagnosis from specialists to train an AI model. We collaborate with neurologists at the University Hospital Dorset (UHD) to collect the source datasets. The proposed architecture supports the diagnosis of 4 common neurological diseases including epilepsy-recurrent seizures, headache, dorsalgia and cerebral infarction.

Diagnosing on the basis of a referral letter can be seen as the text classification task. With the development of the research in natural language processing, there are mainly two types of models that can be used to complete this task: the traditional deep learning-based text classification models including Convolutional Neural Network (CNN) [6, 7], Long Short-Term Memory networks (LSTM) [8] and their variants; the pre-trained attention-based models including Transfomer [9], Bidirectional Encoder Representations from Transformers (BERT) [10], DistilBERT [11], XLNet [12] and Robustly Optimized BERT approach (Roberta) [13].

Distinct from the typical text classification task, the symptoms present in the content for classification will play a pivotal role in determining the final outcome. Our dataset not only contains the long complaint text but also includes short symptom phrases (extracted symptoms). For the traditional deep learning-based model, the feature of each word uses static (fixed) word vectors that come from the pre-trained word embedding models, they have a better performance in short text classification. As embedding models themselves, the pre-trained attention-based structures have the advantage of extracting the long-distance dependencies in the text by using a multi-head attention approach to calculate the similarities between words. Therefore, fusing them together can not only enrich the representation of our neurology dataset but also can benefit from both short-text symptoms and long-text complaints.

A hybrid architecture is proposed in this paper, which fuses the features of words in different representation spaces to make the primary diagnosis from a referral letter. This structure jointly optimizes the pre-trained attention-based model and traditional deep learning model, which makes full use of the two different representation spaces. Besides, two data augmentation methods (Complaint-Symptoms Integration Method and Symptom Dot Separating Method) have also been proposed for this architecture. Subsequent experiments demonstrate that this hybrid architecture has a better performance on the accuracy of the text classification task. Finally, we develop an AI diagnosis assistant web application which leverages the superior performance of this architecture to help GPs complete referral tasks more efficiently and accurately. Our code is available at https://github.com/ruibin-wang/referral-letter-classification.

The contribution of this paper can be summarized as follows:

  • This paper proposed a Hybrid Architecture which fuses the features of words in different representation spaces.

  • This paper proposed a Symptom Dot Separating Method to avoid semantic confusion of the extracted symptoms.

  • A Complaint-Symptoms Integration Method is proposed which is proven to have a positive effect on the accuracy of the disease classification tasks.

  • An AI diagnosis assistant web application is developed to help GPs complete primary diagnosis more efficiently and accurately.

Related work

The referral letters, encompassing detailed records of patients’ historical illnesses and current symptomatic presentations, offer a valuable resource for assisting GPs in decision-making. This study introduces a novel approach, utilizing referral letters as a predictive tool for disease diagnosis. To our knowledge, this is the first initiative to employ referral letters for disease diagnosis. Currently, the primary methodologies for aiding diagnosis include task-oriented dialogue systems and text classification methods. This proposed use of referral letters aims to complement these existing techniques, potentially enhancing the diagnostic process.

The task-oriented dialogue systems

In the realm of task-oriented methodologies, as referenced in studies [2,3,4,5], the primary focus lies on developing a coherent dialogue policy. This policy aims to elicit detailed information from patients through conversation to facilitate accurate diagnoses. The foundational data in these methods is derived from dialogues between patients and doctors. It is important to note that the nature of this source data differs from the focus of our current research problem.

Text classification methods

The complaint in the referral letter presents the patient’s medical information with a long text. Text classification methods can be used for mining useful information in referral letters and classifying it into the correct disease categories. There are three main text classification methods: machine learning-based methods, traditional deep learning-based methods and pre-trained attention-based methods.

Machine learning-based methods: Machine learning approaches in this domain primarily relied on predefined features, such as associated symptoms and disease status, as inputs. These methods employed classifiers like Support Vector Machine (SVM) [14], tree-based methods [15,16,17,18,19,20], Logistic Regression [21], and Naive Bayes (NB) [22] for classification purposes. However, these approaches, deeply rooted in statistical theory, often necessitated extensive feature engineering. Such reliance on manually crafted features could overlook scenarios not encompassed in the predefined sets. This limitation renders these traditional methods less effective for the classification of referral letters, where the range of features might not be counted.

Traditional deep learning-based methods: The advent of word embedding techniques, as referenced in [14, 22,23,24,25], has significantly enhanced the efficacy of traditional deep learning-based methods in text classification tasks. These methods, including Convolutional Neural Networks (CNN) [6, 7], Long Short-Term Memory networks (LSTM) [8], and their various adaptations [7, 26, 27], leverage word embeddings to encode textual semantics into a word representation space. In these approaches, each word in the text is represented by static word vectors derived from pre-trained word embedding models such as fastText [16], word2vec [17,18,19], and GloVe [20]. The unique filter design of CNN and the attention mechanisms within LSTM make these models particularly adept at identifying local and positionally invariant features in texts, such as key phrases or terms crucial for classification. Despite this, the architectural constraints of CNN and LSTM, coupled with the use of static word embeddings, limit their effectiveness in classifying longer texts. This limitation arises from the inherent inability of these models to dynamically adapt word representations based on the surrounding textual context, which is often essential for understanding more extended text passages.

Fine-tuning on the pre-trained attention-based methods: Word representation varies across different semantic spaces, a concept exemplified by pre-trained attention-based models like Transfomer [9], BERT [10], DistilBERT [11], XLNet [12], Roberta [13]. These embedding models, trained on extensive datasets, offer rich contextual representations of words. Additionally, the incorporation of multi-head self-attention mechanisms within their architecture enables them to effectively handle long texts with complex dependencies. This capability is crucial for accurate text classification, as it allows for a more nuanced understanding of the text’s semantic structure.

In this research, we focus on a dataset of referral letters, where the key symptoms detailed within the text are crucial for classification outcomes. Additionally, the dependency attributes of these symptoms play a significant role in determining the final classification results. Therefore, it is essential to propose a method that can simultaneously consider both the overall context and the local key information within the text. In our discussion on text classification methodologies, CNN and LSTM techniques excel in recognizing local and positionally invariant features in text. However, their efficacy diminishes when dealing with extended text sequences. In contrast, pre-trained attention-based models show superior performance with longer texts but lack the precision to consistently highlight all key symptoms in the context. Therefore, we propose a hybrid architecture that involves an integration of these two model types. This hybrid architecture leverages the strengths of each model type, enhancing the system’s capacity for accurate text interpretation and classification. Complemented by appropriate data augmentation methods designed to emphasize key symptoms within the referral letters, the resultant architecture benefits from the synergistic combination, leading to a more robust and nuanced performance in text analysis. The subsequent section will detail the proposed data augmentation methods and the hybrid architecture.

Methods

This paper proposed two data augmentation methods and a hybrid architecture which will be introduced in this section.

Data pre-processing

This paper collected a dataset of referral letters from neurologists at UHD. Our dataset includes records of four common neurological diseases: epilepsy-recurrent seizures (G40), headache (R51), dorsalgia (M54), and cerebral infarction (I63), along with their corresponding complaint texts. An example of this dataset can be found in Table 1, and the dataset’s statistics are presented in Table 2. In Tables 1 and  2, the referral letter presents the patient’s clinical condition and medical history information in the complaint text. The codes G40, R51, I63, and M54 correspond to the International Classification of Disease (ICD) codes. As shown in Fig. 1, before training the hybrid architecture, the complaint text in the referral letter is processed with the following four steps.

Table 1 Referral letter examples for four diseases
Table 2 Statistic of the neurology dataset
Fig. 1
figure 1

Flow chart of the data processing

Step1: Data cleaning

The data source utilized in this study comprises referral letters, typically formatted as emails, written by GPs addressing specialists. Upon examining the referral letters, it was observed that a minor portion of the data contained special characters, such as “$$", “**", “Å", “??" and “XXX", in addition to standard greeting phrases. These elements, unrelated to clinical content, have the potential to obscure critical information within the text. The removal of such characters and words is crucial for diminishing dataset noise and enhancing the focus on pertinent information, which is anticipated to improve the accuracy of disease diagnosis. This paper employs character-matching techniques to systematically eliminate these irrelevant elements.

Step2: Extract symptoms

Symptoms are the key information for a GP to make decisions. A clear and abundant description of symptoms will improve the accuracy of diagnosis. Therefore, this step is to extract symptoms in the given text. Although MIE [28], BERN [29] and Stanza [30] can recognize some symptoms, the accuracy is not satisfactory, and negated symptoms cannot be recognized. This paper adopts the Watson Annotator for Clinical Data of IBMFootnote 1, which can provide a high recognition accuracy for symptoms. Negated symptoms are symptoms that the patient does not have, but which have been mentioned in the complaint text, they will be removed in this step, for example, the blue text in Table 3.

Step3: Symptom Dot Separating Method

Most symptoms consist of multi-words such as falling down, pass out and tunnel vision in Table 3. If these symptoms are connected directly in a sequence as the input for models, the relation between connection words (for example the word ‘down’ and ‘dizziness’ in the processed symptoms part of Table 3) will be calculated, which will extract confusing information from the sequence and has a negative effect on the accuracy of disease classification. Therefore, this paper proposed a Symptom Dot Separating Method, using an unmeaningful symbol dot (‘.’) to integrate the extracted symptoms into training, validating and testing data. Lateral experiments will prove that under the same network structure, the classification performance of training data processed with the Symptom Dot Separating Method is better than that without using this method. The processed symptoms can be seen in Table 3.

Step4: Complaint-Symptoms Integration Method

As described in the Related work section, to effectively harness the potential of referral letters in assisting GPs with decision-making, it is essential to employ text classification methods. For machine learning approaches, predefined features (such as related symptoms or fitness states) are essential for disease diagnosis. Task-oriented dialogue systems [2,3,4,5], aim to develop effective dialogue policies to gather symptom information from patients, employing predefined features as slots to be filled based on conversation content. This indicates the pivotal role of symptoms or fitness states in diagnosis processes. However, defining these features beforehand is labour-intensive and has difficulty addressing complex situations. Employing Transformer-based models leverages their capability to handle long dependencies, thereby obviating the need for labour-intensive feature definition. This allows for the direct use of extensive descriptive text in disease diagnosis, as referenced in [31]. However, a limitation of these Transformer-based approaches is their equal treatment of each word, which may lead to a diminished focus on disease-specific information within the text.

This paper proposes a Complaint- Symptoms Integration Method, which combines the complaint text and processed symptoms with a simple symbol “[SEP]". This symbol is often used in pre-trained attention-based models as a special connection between two different sentences. Our integration method not only keeps the original semantics of the patient’s historical medical information but also highlights key information in the complaint text. Subsequent experiments have shown that the classification accuracy of this method is higher than that of training complaint text and extracting symptoms alone.

Table 3 Example of complaint text processing

Table 3 shows an example of complaint text processing. It can be clearly seen that the related symptoms are correctly extracted (marked in red in the complaint text), while negated symptoms are also obtained (marked in blue) and not included in the processed symptoms. Besides, three datasets Symptoms, Complaints, and Integrations are generated to validate the performance of our proposed methods and architectures. The Symptoms dataset only contains the symptoms processed by Step3 (for example processed symptoms in Table 3). The Complaints dataset only contains the cleaned complaint text processed by Step1(for example complaint text in Table 3). The Integration dataset consists of data generated through Step4 (for example network input in Table 3).

The hybrid architecture

Our architecture integrates a pre-trained attention-based model (such as BERT, DistilBERT, XLNet, or Roberta) with a traditional deep learning model (like CNN, LSTM, or their variants). This hybrid approach leverages the strengths of both model types. Pre-trained models, having been extensively trained on large datasets, are adept at comprehending complex language and semantic nuances. They provide a context-rich word representation, which is vital for precise text classification. Meanwhile, the CNN or LSTM models are particularly effective in identifying local and positionally invariant features in text, like key phrases or terms that are critical for classification. By merging these two model types, our architecture benefits from their complementary capabilities. This synergy enhances the system’s ability to interpret and classify text, yielding a more robust and nuanced performance.

The input of this architecture is the three datasets processed in “Data pre-processing” section and the output is the prediction of the 4 common neurological diseases. We take BERT+CNN as an example to explain the training process. As depicted in Fig. 2, the left part is the BERT branch, and the right part is the CNN branch (the Bert branch can be changed to other pre-trained attention-based models).

Fig. 2
figure 2

The hybrid architecture

Given the neural network input data as \(X=(x_1,...x_i,...,x_m)\), four diseases as \(Y=(y_1,...,y_4)\) in which \(x_i\) is the ith word in the text and m is the length of the text.

BERT Branch

The first hidden state H in Fig. 2 of the BERT model presents the features of the whole input text, it is used as the input of the lateral linear layer. This paper utilizes the sigmoid function in the softmax layer as

$$\begin{aligned} Bert_{logits} = Softmax(Linear(H)). \end{aligned}$$
(1)

The Cross-Entropy Loss is used as our loss function, the loss of the BERT branch is described as

$$\begin{aligned} Loss_{bert} = CrossEntropyLoss(Bert_{logits},Y). \end{aligned}$$
(2)

CNN Branch

Word embedding is a learned representation for text where words that have the same meaning would have a similar representation. As mentioned in the introduction part, there are mainly four pre-trained word embedding models, because of the wide coverage of different semantic lexical, the ‘crawl-300d-2M.vec’ [32] (2-million-word vectors trained on Common Crawl (600B tokens)) of the Google’s fastText are utilized as the embedding layer of CNN as shown in Eq. (3).

$$\begin{aligned} X_{vector} =Embedding(X). \end{aligned}$$
(3)

The 1-dimension CNN is chosen as the convolutional layer, which is shown below:

$$\begin{aligned} CNN_{output} = conv1d(X_{vector}). \end{aligned}$$
(4)

The max pooling layer and linear classification layer are shown in Eq. (5).

$$\begin{aligned} CNN_{linear}=Linear(Maxpooling(CNN_{output})). \end{aligned}$$
(5)

The softmax layer is described as following

$$\begin{aligned} CNN_{logist} = Softmax(CNN_{linear}). \end{aligned}$$
(6)

The loss function of the CNN model is the same as BERT, as shown

$$\begin{aligned} Loss_{CNN}= CrossEntropyLoss(CNN_{logist},Y). \end{aligned}$$
(7)

A parameter \(\alpha\) for the hybrid architecture is used to fuse the two branches, presented below

$$\begin{aligned} Loss = \alpha *Loss_{CNN}+ (1-\alpha )*Loss_{Bert}. \end{aligned}$$
(8)

The \((m+1)th\) fusion weight \(\alpha _{m+1}\) is updated by the following equation

$$\begin{aligned} \alpha _{m+1}=\alpha _m-(Loss_{CNN}-Loss_{Bert})*lr*\alpha _m. \end{aligned}$$
(9)

In which the parameter lr is the learning rate.

Experiment

Experiment settings

This paper prepares three training datasets, namely the Symptoms dataset, Complaints dataset and Integration dataset to validate the performance of our proposed architecture and methods. The Symptoms dataset and Complaints dataset consists of processed symptoms and complaints, respectively, and the Integration dataset consists of texts processed by the Complaint-Symptoms Integration Method. To validate the hybrid architecture and the proposed training strategies, the training, validation and testing samples are divided in a ratio of 7:2:1, which have 179/51/27 samples, respectively. Meanwhile, four tasks are arranged using classification accuracy as the metric.

Bert branch baseline

Comparing the performance of different pre-trained attention-based models on our proposed hybrid architecture, four commonly used sentence classification models are adopted. The details of the four models are as below:

  • BERT is a pre-trained language model that has achieved state-of-the-art performance in various natural language processing (NLP) tasks. It uses a bidirectional transformer architecture, which allows it to consider the entire context of a sentence or text when generating its representations.

  • DistilBERT is a smaller and faster version of BERT, designed for resource-constrained environments, while still maintaining high accuracy in various NLP tasks. It uses a distillation technique to transfer the knowledge from the larger BERT model to a smaller one.

  • XLNet is a transformer-based language model that aims to address some of the limitations of previous models like BERT. It uses a permutation-based approach for generating representations, which allows it to capture more diverse and complex relationships between words in a sentence.

  • RoBERTa (Robustly Optimized BERT approach) is another variant of BERT, which was pre-trained on a larger dataset and with improved training techniques. It has achieved state-of-the-art performance on several NLP tasks, outperforming BERT and other previous models.

Parameters setting: The batch size is set to 8, dropout \(p=0.5\), the optimization method is Adam [33] and the learning rate is set as 1e-6. For the CNN branch, this paper chooses the 1-dimension CNN which has three kernel sizes (2,3,4) and each kernel has two filters. Besides, the initial value of \(\alpha\) for the hybrid architecture is set as 0.3 with lr=1e-5. The hidden unit of CNN uses the fixed embedding representation vector (dimension: 300) and the epoch is set as 200.

The configuration of the computer is Windows 10, Nvidia RTX 2080super, Intel Core (TM) i7-9700 and RAM 32G.

Experimental design

In order to validate the diagnostic accuracy of our proposed methods, three experiments are designed. Besides, three training methods are adopted for each architecture, including training on the symptoms processed with the Symptom Dot Separating Method, training on the complaint text only and training on the data processed with the Complaint-Symptoms Integration Method.

Fine-tuning the pre-trained attention-based model: This step will further fine-tune the pre-trained attention-based models such as BERT, DistilBERT, XLNet and Roberta on the neurology referral letter dataset. The architecture of this step can be found in the left BERT branch of Fig. 2.

The hybrid architecture: This part is designed to compare the accuracy of different combinations between BERT+CNN, DistilBERT+CNN, Roberta+CNN and XLNet+CNN to validate the performance of the proposed hybrid architecture.

Comparison with the sequential model: As shown in Fig. 3, the sequential model consists of the pre-trained attention-based network and CNN network, and the pre-trained attention-based model can be seen as the embedding of CNN. A single embedding representation space is utilized in this architecture, and it is used to compare the effectiveness of the hybrid architecture which fuses two different embedding representation spaces. Four models are designed for this task, including the BERT-CNN, DistilBERT-CNN, Roberta-CNN and XLNet-CNN. The hidden unit of CNN in this part keeps the same dimension as the pre-trained attention-based embedding model (768).

Fig. 3
figure 3

Bert-CNN sequential model

Validation of adding dot: By comparing the performance of adding dots and without dots for the extracted symptoms, this task is designed to validate the effect of the Symptom Dot Separating Method on our proposed hybrid architecture.

Main results

Results of the design tasks on our neurology dataset are presented in Table 4. As described in “Experimental design” section, the first vertical column of Table 4 shows three designed architectures, and the second column shows different models in each architecture. Symptoms, Complaints and Integration are three training datasets described in “Data pre-processing” section.

Table 4 Results of the four designed experiments

As shown in the black bold data of pre-trained attention-based models in Table 4, the classification accuracy of training on the Integration dataset for each model is significantly higher than the results of training on Complaints and extracted Symptoms alone. With an average of 3% and 3.9% improvement on the Symptoms and Complaints datasets respectively. Besides, CNN also has an obvious improvement. It indicates that training on the data processed with our data augmentation methods will produce better classification accuracy. Comparing the result in the Complaints column of pre-trained models with CNN, it can be found that fine-tuning on pre-trained attention-based models perform better than CNN (as 0.734 is lower than the other results in the complaints column of Table 4 and highlighted in blue). But our proposed Complaint-Symptoms Integrating training method ‘re-energizing’ CNN, makes it outperform Roberta and XLNet and achieve comparable performance to BERT and DistilBERT. This also can be evidence of the positive effect of our proposed data integration method.

Two conclusions can be drawn from the hybrid architecture of Table 4:

1. By comparing the performance of hybrid architectures and their associated pre-trained attention-based models under the same training dataset (for example comparing BERT+CNN and BERT on the Symptoms, Complaints or Integration dataset), it can be found that on both Symptoms and Integration datasets, all hybrid architectures outperform pre-trained attention-based models. Besides, BERT+CNN architecture obtains the best classification accuracy on our integration dataset with a 5.1% improvement.

To demonstrate the efficacy of the hybrid architecture, Fig. 4 provides a visual representation of the training and validation processes in terms of both loss and accuracy. Specifically, Fig. 4 showcases the performance metrics of the BERT+CNN architecture, serving as a representative example. This illustration aids in comprehending the architecture’s effectiveness during the training phase and its subsequent validation. For Fig. 4, the left vertical axis represents the loss, with the red line indicating training loss and the green line denoting validation loss. The right vertical axis is designated for accuracy, where the blue line represents validation accuracy and the orange line shows training accuracy. The horizontal axis corresponds to the number of training epochs. From the data in Fig. 4, the BERT+CNN model demonstrates robust performance, as indicated by the converging trends in loss and accuracy, without evidence of overfitting. Notably, around 75 epochs, the model appears to reach a plateau in its learning capacity. This is reflected by the stabilization of training accuracy, despite a continuing decrease in training loss. Consequently, the validation values exhibit fluctuations around this convergence point.

Fig. 4
figure 4

Training and validation loss and accuracy for the BERT+CNN architecture

2. As for the hybrid architectures in Table 4, by comparing the classification accuracy of different training datasets under the same architecture, the blue bold data in Table 4 shows that all non-sequential architectures benefit from our Integration dataset: models trained on it consistently achieve better performance than either Symptoms or Complaints on their own.

These two conclusions demonstrate the positive effect of our hybrid architecture and two data augmentation methods on neurology disease classification tasks. Our method brings a maximum of 11% improvement for disease classification. The reasons for this improved performance are twofold: the two data augmentation methods enrich the content of training data, and they combined the original complaint text with the extracted symptoms. This highlights the key information in the complaint text. Another is that the hybrid architecture fuses two embedding spaces, which enriches the representation of our neurology dataset. Therefore, training on them results in better classification accuracy.

Comparing the classification accuracy of the sequential architectures and their related hybrid architectures (such as BERT+CNN and BERT-CNN) in Table 4, our proposed pre-trained attention-based models+CNN outperforms pre-trained attention-based models-CNN on all three training datasets. This is because our architecture fuses two different embedding representation spaces, which can improve the accuracy of neurological disease classification tasks significantly.

Due to the superior performance of the hybrid architecture BERT+CNN, it is chosen to validate the performance of the Symptom Dot Separating Method, the result can be found in Table 5. This table shows that adding dots between each symptom improves the classification accuracy of our proposed hybrid architecture. The main reason is that each symptom is separated, without semantic meaning, adding dots can avoid semantic confusion.

Table 5 Validating the performance of the Symptom Dot Separating Method

This paper proposed three methods, including the hybrid architecture, the Symptom Dot Separating Method and the Complaint-Symptoms Integration Method which are proven to have positive effects on the accuracy of neurological disease classification tasks. This means that it can be a good assistant tool for GPs to make the primary diagnoses.

Application

In order to help GPs complete diagnosis more efficiently and accurately, this paper applied the hybrid architecture to develop an AI diagnosis assistant web application, which can help GPs make a primary diagnosis by talking/texting with them. Flowchart (left part in the figure) and demo screenshot of the conversation between a GP and AI diagnosis assistant (right part in the figure) can be found at Fig. 5. The natural language understanding part in Fig. 5 follows the same steps as the data processing in “Data pre-processing”. Besides, both Speech-to-text (STT) and Text-to-speech (TTS) functions are implemented using state-of-art methods, which can better improve the experience of this AI diagnosis assistant web application. A demonstration video which shows the process of both writing and speaking assistant process can be found at https://www.youtube.com/watch?v=Coj1xGYOCBw.

Fig. 5
figure 5

The flowchart and demo of the assistance diagnosis application which is developed on the hybrid architecture

Conclusion

This paper introduces a hybrid architecture to help GP to improve the accuracy of primary diagnosis. Comparative experiments on the neurology dataset demonstrate that our proposed model has satisfactory classification accuracy. We also proposed two data augmentation methods (the Symptom Dot Separating Method and the Complaint-Symptoms Integration Method) which were proved effective in improving the predicting accuracy of the neurology disease. This paper also applied the hybrid architecture to develop an AI diagnosis assistant web application, which can efficiently help GPs make a primary diagnosis by talking/texting with them. The accuracy of this architecture needs to be improved, and in the future, the relation between each symptom would be presented via the knowledge graph. The dataset and the source code can be found at the GitHub link which is shared in the Introduction section.

Availability of data and materials

The dataset utilized in the study is currently inaccessible as the collaborating doctor has no plan to share the data, citing complicated policies that must be followed for releasing a dataset. The sample data and source code can be found at the following link: https://github.com/ruibin-wang/referral-letter-classification.

Notes

  1. https://acd-try-it-out.mybluemix.net/

References

  1. O’Donnell CA. Variation in GP referral rates: what can we learn from the literature? Fam Pract. 2000;17(6):462–71.

    Article  PubMed  Google Scholar 

  2. Lewenberg Y, Bachrach Y, Paquet U, Rosenschein J. Knowing what to ask: A bayesian active learning approach to the surveying problem. In: AAAI. 2017;31(1). https://0-doi-org.brum.beds.ac.uk/10.1609/aaai.v31i1.10730.

  3. Shim H, Hwang SJ, Yang E. Joint active feature acquisition and classification with variable-size set encoding. NeurIPS; 2018. p. 31.

  4. Wei Z, Liu Q, Peng B, Tou H, Chen T, Huang XJ, Wong KF, Dai X. Task-oriented dialogue system for automatic diagnosis. In: ACL. 2018;2:201–7. https://aclanthology.org/P18-2033.

  5. Xu L, Zhou Q, Gong K, Liang X, Tang J, Lin L. End-to-End Knowledge-Routed Relational Dialogue System for Automatic Diagnosis. AAAI. 2019;33:7346–53. https://0-doi-org.brum.beds.ac.uk/10.1609/aaai.v33i01.33017346.

    Article  Google Scholar 

  6. Kim Y. Convolutional Neural Networks for Sentence Classification. EMNLP. 2014. https://0-doi-org.brum.beds.ac.uk/10.3115/v1/D14-1181.

  7. Conneau A, Schwenk H, Barrault L, Lecun Y. Very Deep Convolutional Networks for Text Classification. In: EACL. 2017;1:1107–16. https://aclanthology.org/E17-1104.

  8. Mousa A, Schuller B. Contextual bidirectional long short-term memory recurrent neural network language models: a generative approach to sentiment analysis. In: EACL. Valencia; 2017. p. 1023–1032. https://aclanthology.org/E17-1096.

  9. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I. Attention is all you need. NeurIPS; 2017. p. 30.

  10. Alaparthi S, Mishra M. BERT: a sentiment analysis odyssey. Journal of Marketing Analytics. 2021;9. https://0-doi-org.brum.beds.ac.uk/10.1057/s41270-021-00109-8.

  11. Sanh V, Debut L, Chaumond J, Wolf T. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108 . 2019.

  12. Yang Z, Dai Z, Yang Y, Carbonell J, Salakhutdinov RR. Le QV. Xlnet: generalized autoregressive pretraining for language understanding. NeurIPS; 2019. p. 32.

  13. Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, Levy O, Lewis M, Zettlemoyer L, Stoyanov V. Roberta: a robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 . 2019.

  14. Joachims T, Ls I, Str B. Text categorization with support vector machines: learning with many relevant features. In: ECML. 1998. p. 137–42. https://0-doi-org.brum.beds.ac.uk/10.1007/BFb0026683.

  15. Kononenko I. Inductive and Bayesian learning in medical diagnosis. Appl Artif Intell Int J. 1993;7(4):317–37.

    Article  Google Scholar 

  16. Kononenko I. Machine learning for medical diagnosis: history, state of the art and perspective. Artif Intell Med. 2001;23(1):89–109.

    Article  CAS  PubMed  Google Scholar 

  17. Xu Z, Kusner M, Weinberger K, Chen M. Cost-sensitive tree of classifiers. In: International conference on machine learning. PMLR; 2013. p. 133–141.

  18. Kohavi R. Scaling up the accuracy of naive-bayes classifiers: a decision-tree hybrid. In: KDD. 1996;96:202–7. https://0-dl-acm-org.brum.beds.ac.uk/doi/10.5555/3001460.3001502.

  19. Nan F, Wang J, Saligrama V. Feature-budgeted random forest. In: International conference on machine learning. PMLR; 2015. p. 1983–1991.

  20. Hayashi Y. A neural expert system with automated extraction of fuzzy if-then rules and its application to medical diagnosis. NeurIPS; 1990. p. 3.

  21. Genkin A, Lewis DD, Madigan D. Large-scale Bayesian logistic regression for text categorization. Technometrics. 2007;49(3):291–304.

    Article  Google Scholar 

  22. McCallum A, Nigam K. A comparison of event models for naive bayes text classification. In: AAAI. 1998;752(1):41–8. https://aaai.org/papers/041-ws98-05-007.

  23. Lin CY, Hovy E. Automatic evaluation of summaries using n-gram co-occurrence statistics. In: NAACL. 2003. p. 150–7. https://aclanthology.org/N03-1020.

  24. Zhang W, Yoshida T, Tang X. TFIDF, LSI and multi-word in information retrieval and text categorization. In: IEEE SMC. IEEE; 2008. p. 108–113.

  25. Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J. Distributed representations of words and phrases and their compositionality. NeurIPS; 2013. p. 26.

  26. Zhang X, Zhao J, LeCun Y. Character-level convolutional networks for text classification. NeurIPS; 2015. p. 28.

  27. Kim Y, Lee H, Jung K. Attnconvnet at semeval-2018 task 1: Attention-based convolutional neural networks for multi-label emotion classification. arXiv preprint arXiv:1804.00831 . 2018.

  28. Zhang Y, Jiang Z, Zhang T, Liu S, Cao J, Liu K, Liu S, Zhao J. MIE: A Medical Information Extractor towards Medical Dialogues. In: ACL. 2020. p. 6460–6469. https://aclanthology.org/2020.acl-main.576.

  29. Lee J, Yoon W, Kim S, Kim D, Kim S, So CH, Kang J. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics. 2020;36(4):1234–40.

    Article  CAS  PubMed  Google Scholar 

  30. Zhang Y, Zhang Y, Qi P, Manning CD, Langlotz CP. Biomedical and clinical English model packages for the Stanza Python NLP library. JAMIA. 2021;28(9):1892–9.

    PubMed  PubMed Central  Google Scholar 

  31. Zhou M, Li Z, Tan B, Zeng G, Yang W, He X, Ju Z, Chakravorty S, Chen S, Yang X, Zhang Y. On the Generation of Medical Dialogs for COVID-19. In: ACL-IJCNLP). Online; 2021. p. 886–896. https://0-doi-org.brum.beds.ac.uk/10.18653/v1/2021.acl-short.112.

  32. Mikolov T, Grave E, Bojanowski P, Puhrsch C, Joulin A. Advances in pre-training distributed word representations. arXiv preprint arXiv:1712.09405 . 2017.

  33. Kingma DP, Ba J. Adam: A Method for Stochastic Optimization. In: 3rd ICLR; 2015. arXiv:1412.6980 .

Download references

Acknowledgements

During the research process and paper preparation, I received selfless assistance from my supervisor and numerous colleagues. I am immensely grateful for their invaluable contributions.

Funding

This research has been supported by the Neuravatar Project funded by High Education Innovation Fund (UK-HEIF), match funded PhD scholarship from Bournemouth University and Shandong Hengdao Ruyi Digital Communication Ltd, the Cloud based Virtual Production Project funded by the Arts and Humanities Research Council (UK-AHRC Ref: AH/W009323/1). We state that the funding bodies played no role in the design of the study and collection, analysis, and interpretation of data and in writing the manuscript.

Author information

Authors and Affiliations

Authors

Contributions

Project administration, supervision and conceptualization: RP, XY, HL and JZ; Methodology, experiment, visualization and writing-original draft preparation: RW; Writing—review and editing: XY, RW, HL and KJ; Web application development: KJ; Funding acquisition HL; Validation, formal analysis and data curation: RW and RP. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Ruibin Wang.

Ethics declarations

Ethics approval and consent to participate

We would like to clarify that the referral letters used in our study have been fully anonymized. Any personal or identifiable information has been removed, and therefore, ethical approval is not required.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, R., Jayathunge, K., Page, R. et al. Hybrid architecture based intelligent diagnosis assistant for GP. BMC Med Inform Decis Mak 24, 15 (2024). https://0-doi-org.brum.beds.ac.uk/10.1186/s12911-023-02398-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://0-doi-org.brum.beds.ac.uk/10.1186/s12911-023-02398-8

Keywords