The emergency plans for water diversion projects suffer from weak knowledge correlation, inadequate timeliness, and insufficient support for intelligent decision-making. This study incorporates knowledge graph technology to enable intelligent recommendations for emergency plans in water diversion projects. By employing pre-trained language models (PTMs) with entity masking, the model's ability to recognize domain-specific entities is enhanced. By leveraging matrix-based two-dimensional transformations and feature recombination, an interactive convolutional neural network (ICNN) is constructed to enhance the processing capability of complex relationships. By integrating PTM with ICNN, a PTM–ICNN method for joint extraction of emergency entity relationships is constructed. By utilizing the Neo4j graph database to store emergency entity relationships, an emergency knowledge graph is constructed. By employing the mutual information criterion, intelligent retrieval and recommendation of emergency plans are achieved. The results demonstrate that the proposed approach achieves high extraction accuracy (F1 score of 91.33%) and provides reliable recommendations for emergency plans. This study can significantly enhance the level of intelligent emergency management in water diversion projects, thereby mitigating the impact of unforeseen events on engineering safety.

  • Pre-trained language models with entity masking to improve the ability of the models to recognize domain entities.

  • Matrix-based two-dimensional transformation and feature reorganization to enhance the processing of complex relationships.

  • Intelligent retrieval and recommendation of emergency plans based on mutual information criterion.

Abbreviations

Expansion

PTMs

pre-trained language models

PTM–ICNN

fusion of PTM and ICNN

BILSTM

bidirectional long short-term memory

CNN

convolutional neural network

GPTs

generative pre-trained transformers

BERT–CRF

fusion of BERT and CRF

ICNN

interactive convolutional neural network

RNNs

recurrent neural networks

CRF

conditional random field

BERT

bidirectional encoder representations from transformers

BERT–BILSTM–CRF

fusion of BERT, BILSTM, and CRF

BERT–CNN

fusion of BERT and CNN

Variables

Meaning

, K, and

word vector matrices

number of attention heads

relation embedding vector

one-dimensional convolutional kernel

feature map obtained after the one-dimensional convolutional calculation

feature alternating matrix

feature map obtained after the two-dimensional convolution calculation

masked keyword sequence

Vector concatenation

trainable weight matrix

count of entity relationships

entity to be retrieved

number of correctly predicted entities

scaling factor

entity embedding vector

convolution operator

vector alternating matrix

two-dimensional convolution kernel

text sequence

extracted text features

embedding matrix for entity relationships

Softmax function

input entity

joint probability distribution of the two random variables

number of predicted entities that are not actual entities

number of entities that were not predicted

Emergency plans serve as critical foundations for engineering emergency response and precise decision-making. Constructing emergency plans as knowledge graphs and enabling intelligent recommendations are the keys to advancing the level of intelligence in emergency management (Kreibich et al. 2022). Currently, emergency plans are predominantly in document form, making it difficult to achieve effective knowledge integration and updates. They suffer from weak knowledge association, limited timeliness, and insufficient support for intelligent decision-making (Liu & Chen 2019; Tang et al. 2022). Hence, investigating the methods for constructing an emergency knowledge graph in water diversion projects and achieving intelligent recommendations of emergency plans is a crucial issue that urgently needs to be addressed in the field of water diversion emergency management.

Traditional methods for constructing knowledge graphs primarily rely on rule-based matching techniques, where textual information is extracted and knowledge is built using logical rules devised by experts (Popovski et al. 2019; Zhu & Zheng 2020). Mykowiecka et al. (2009) combined semantic rules and dictionaries to build a relational knowledge graph between patients and drugs, providing support for medical decision-making and knowledge discovery. Chiticariu et al. (2013) utilized regular expressions and rule templates to construct a biomedical knowledge graph, facilitating automatic recommendation of biomedical knowledge. Milosevic et al. (2019) integrated rule matching and template matching techniques to extract event information from biomedical literature, enabling support for biomedical research. Rule-based methods necessitate frequent rule definition and adjustment, which limits the scalability and expansiveness of knowledge graphs (Bouraoui & Schockaert 2019; Lan et al. 2020).

The knowledge graph construction approach based on deep learning utilizes deep learning models to automatically extract entities and relationships from textual data, offering advantages such as strong scalability and high automation (Ma et al. 2021a, 2021b; Li et al. 2022). Deep learning models can be categorized into individual models and hybrid models (Banan et al. 2020; Fan et al. 2020; Chen et al. 2022). Individual models, such as the one presented by Zeng & Wang (2022), employ recurrent neural networks (RNNs) to establish an intelligent recommendation knowledge graph, thereby augmenting the precision and dependability of recommendation systems. Zhou et al. (2020) have put forth a knowledge graph construction methodology, focusing on the industrial domain, which leverages the power of bidirectional long short-term memory (BILSTM) neural networks. This approach has resulted in notable enhancements in both the efficiency and quality of industrial manufacturing. However, individual models often suffer from poor generalization and low precision in knowledge extraction (Afan et al. 2021; Xiao & Zhang 2021; Fan & Wang 2022).

Hybrid models can effectively address the shortcomings and limitations of individual models. Chang et al. (2022) enhanced the accuracy and reliability of knowledge graph construction by combining conditional random fields (CRFs) with BILSTM. Zhou et al. (2020) addressed the issue of BILSTM's inability to capture spatial dependencies in text by proposing a convolutional neural network (CNN)-based BILSTM model for constructing an industrial process knowledge graph. With the emergence of advanced pre-trained models such as a bidirectional encoder representation from transformers (BERTs) and generative pre-trained transformers (GPTs), scholars have sought to leverage these models to enhance the limitations of deep learning models (Lee & Toutanova 2018; Radford et al. 2018). Meng et al. (2022) developed a knowledge graph for power equipment failures based on BERT–BILSTM–CRF, offering more accurate and reliable support for fault information in the field of electrical power. Zhang et al. (2021) employed BERT–CRF to construct a knowledge graph for the management of chemical hazardous materials. Currently, hybrid models based on pre-training still face issues such as fixed pre-training patterns and low accuracy in domain knowledge extraction.

Building upon the aforementioned discussions, this research proposes an intelligent recommendation approach for emergency plans in water diversion projects, based on knowledge graph, to address issues such as weak knowledge associations, poor timeliness, and insufficient intelligent decision support. The main contributions are as follows:

  • (1)

    By employing a pre-trained language model (PTM) with entity masking, the model's ability to identify domain-specific entities is enhanced.

  • (2)

    By leveraging matrix-based two-dimensional transformations and feature recombination, the interactive CNN (ICNN) is constructed to enhance the processing capability of complex relationships.

  • (3)

    By combining the PTM with ICNN, we construct a PTM–ICNN-based joint extraction method for emergency entity relationships.

  • (4)

    Utilizing the mutual information criterion, we achieve intelligent retrieval and recommendation of emergency plans.

The remaining sections of this paper are organized as follows: Section 2 introduces the relevant work of PTM–ICNN. Sections 3 and 4 provide detailed explanations of the construction process of PTM–ICNN and the process of intelligent recommendation of emergency plans. Section 5 presents the experimental design, results, and discussions. Finally, in Section 6, the paper concludes and highlights future research directions.

Pre-trained language models

PTMs aim to acquire extensive linguistic knowledge through self-supervised learning on text corpora (Zhang et al. 2021). Early pre-training language models primarily relied on n-gram and rule-based methods. Cichosz (2018) introduced the Global Vectors model, which accomplishes pre-training by leveraging the statistical co-occurrence information of words and global semantic information. Ma et al. (2021a, 2021b) proposed the Word Vector model, which achieves pre-training by representing words as continuous vectors through self-supervised learning.

With the advancement of deep learning, Han et al. (2021) introduced the transformer model, which employs self-attention mechanism for sequence modeling. The transformer model effectively captures long-range dependencies in text through its self-attention mechanism and has demonstrated significant performance improvements in machine translation tasks. Subsequently, Lee & Toutanova (2018) proposed the BERT model, building upon the foundation of the transformer. The BERT model employs a strategy of pre-training and fine-tuning, engaging in large-scale unsupervised training on vast amounts of data, followed by fine-tuning on downstream tasks. The groundbreaking innovations of the BERT model have yielded outstanding results across multiple natural language processing tasks, sparking widespread scholarly interest in PTMs.

Convolutional neural network

A CNN is a feedforward neural network renowned for its formidable feature extraction capabilities (Yamashita et al. 2018; Tan & Le 2019). The origins of CNNs can be traced back to the 1980s. LeCun et al. (1989) introduced the LeNet model, which was developed for the task of handwritten digit recognition. This model incorporates the structural components of convolution and pooling and incorporates the use of the backpropagation algorithm for training. With the emergence and development of big data, researchers have embarked on exploring deep CNN models. Krizhevsky et al. (2017) introduced the AlexNet model, which achieved a significant breakthrough in the ImageNet image classification competition by employing multiple convolutional layers and fully connected layers.

The initial application of the CNN in natural language processing tasks was primarily focused on text classification. Chen (2015) proposed a text classification model based on the CNN, which utilizes convolution operations on sentences to extract local features. The model has achieved favorable performance on multiple text classification tasks. Subsequently, the CNN has been employed in sequence modeling tasks, such as named entity recognition and part-of-speech tagging. Gehring et al. (2017) proposed a sequence labeling model based on the CNN, which performs convolution and pooling operations on the input sequence to make label predictions.

PTM construction

The emergency plans for water diversion projects encounter challenges involving a plethora of technical terminologies and complexities in entity naming rules. Leveraging PTMs allows the transfer of linguistic knowledge acquired during the pre-training phase to the task of extracting emergency knowledge, thereby significantly enhancing entity recognition accuracy. Conventional PTMs, such as BERT, typically employ a character-level masking strategy during pre-training, where a solitary [MASK] character is utilized to conceal individual characters within the text. The character-level masking strategy employed by BERT falls short in fully concealing the entire entity, leading to biases and erroneous predictions when the model attempts to forecast the masked entity. To facilitate the model's effective extraction of contextual features from the input sequence, this study employs the pre-training strategy of entity masking. By randomly masking certain entities within the input sequence, the model can leverage context to infer the original meaning of the masked positions, thus enhancing its understanding of domain-specific entities and consequently improving the accuracy of entity recognition tasks. The PTM proposed in this study is a self-supervised deep language model that undergoes pre-training on multi-layer transformer structures through a random masking mechanism. It comprises input, feature extraction, and output layers, with the fundamental implementation as follows.

Input layer: Employing an entity lexicon, the text undergoes tokenization and entity analysis. Multiple consecutive [MASK] markers are utilized to mask the entire entity. The masked input sequence is composed of character embedding vectors, word embedding vectors, and positional embedding vectors. By employing vector addition, we integrate the character embedding vectors, word embedding vectors, and positional embedding vectors together as inputs for the PTM model. As depicted in Figure 1, the strategy for random entity masking is as follows: (1) 80% of the entities in the input sequence are replaced with alternative entities. (2) 10% of the entities in the input sequence are masked with [MASK]. (3) The remaining 10% of the input sequence remains unchanged.
Figure 1

The basic process of stochastic [MASK] in the pre-training phase.

Figure 1

The basic process of stochastic [MASK] in the pre-training phase.

Close modal
Feature extraction layer: The feature extraction layer of PTM consists of a 24-layer transformer network, with each transformer network comprising six encoder layers. The transformer network employs a self-attention mechanism, which calculates contextual features of words or characters and allocates weights to each word or character accordingly. In such cases, word vectors or character vectors encapsulate semantic information related to their context. Taking word embeddings as an example, the self-attention mechanism of transformer is defined as follows.
(1)
Among them, Q, K, and V represent word vector matrices, where the dot product matrix of Q and represents the degree of correlation for each word, and denotes the scaling factor. Building upon this foundation, multiple self-attention layers are concatenated using a multi-head structure, resulting in a more interpretable multi-head attention mechanism, as defined in the following formula.
(2)
(3)

Among them, W represents the weight matrix, signifies the number of attention heads, and , , , respectively, denote the weight parameter matrices for Q, K, and V. To expedite the convergence of the model, this study incorporates residual networks and applies normalization techniques.

Output layer: The output layer of PTM comprises the results of feature extraction performed by 24 layers of transformer encoders and decoders, presented in the form of word vectors. These word vectors encapsulate the rich semantic information of the input text.

ICNN construction

CNN-based relation classification is a prevalent natural language processing task, aiming to discern and categorize the relationships between entities within a given text. CNN exhibits remarkable generalization capabilities and high computational efficiency (Zhou et al. 2022). However, the relationships among emergency entities in water diversion projects possess intricacies and uncertainties. For the extraction of features in complex relationships, conventional methods often resort to one-dimensional convolutions for feature extraction. For instance, the utilization of multi-channel CNN and recurrent CNN treats word vectors as one-dimensional time series data, employing one-dimensional convolutions to extract local features. This one-dimensional convolution-based feature extraction method typically captures only local information, exhibiting limited comprehension of global context. This study employs two-dimensional transformations and feature recombination to construct vector alternating matrices and feature alternating matrices, enhancing the feature interactions of entities. This approach effectively improves the model's capability to handle complex relationships, thereby enhancing the accuracy of relation classification. ICNN comprises two main modules: feature recombination and convolution computation. Feature recombination involves reordering the embedding matrix of entity relations for input triplets, enhancing the interactivity of entity relations. Figure 2 presents the three embedding arrangement methods utilized in this study: original arrangement, vector alternation, and feature alternation. The convolution computation module is designed to extract and integrate features from the rearranged embedding matrix, and its fundamental implementation is as follows.
Figure 2

Three ways of interacting with embedded vectors: (a) original arrangement; (b) vector alternation; and (c) feature alternation.

Figure 2

Three ways of interacting with embedded vectors: (a) original arrangement; (b) vector alternation; and (c) feature alternation.

Close modal
Feature reorganization: Let and represent the embeddings of emergency entity relations, where denotes the entity embedding vector, and represents the relation embedding vector. Firstly, we perform a dimensional transformation on the embedding vectors. The transformation equation for entity embedding vectors is defined as follows.
(4)
Among them, , and A represents the entity embedding matrix after the dimensional transformation. Next, the matrix A is subjected to feature rearrangement, which primarily involves two methods: vector alternation and feature alternation, defined as follows.
(5)
(6)

Among them, and , respectively, denote the reorganized matrices after vector alternation and feature alternation.

Convolution computation: Utilizing 3 × 3 and 3 × 1 two-dimensional convolution kernels, we perform feature extraction on matrices and with the extraction process defined as follows.
(7)

Among them, represents the dimensionality of the two-dimensional convolutional filter, and denotes the convolution operator. represents the feature map obtained after the two-dimensional convolution computation. Then, utilizing matrix addition, the feature map is fused, and subsequently, through inner product operations, the model achieves its final output. The ICNN model's output consists of the relational features between entities.

PTM–ICNN model construction

The PTM–ICNN model developed in this study comprises two primary modules: the pre-training module and the convolution computation module. By leveraging the pre-training module, we can uncover the semantic features of the text and enhance the accuracy of named entity recognition. By conducting convolutional computations, we can extract interactive features of entity relations, thus enhancing the ability to extract complex relationships. Using PTM–ICNN, we achieve the joint extraction of entity relationships in water diversion projects' emergency plans. Figure 3 illustrates the fundamental process of the PTM–ICNN model, and its basic implementation is as follows.
Figure 3

Data processing flow for the PTM–ICNN joint entity relationship extraction model.

Figure 3

Data processing flow for the PTM–ICNN joint entity relationship extraction model.

Close modal

Pre-training module: Initialize the text sequence, referred to as , and apply entity masking to the text sequence using the entity masking mechanism. The masked input sequence is constructed using character embedding vectors, word embedding vectors, and positional embedding vectors. By utilizing vector addition, we integrate the character embedding vectors, word embedding vectors, and positional embedding vectors together as inputs to the model. After multiple layers of bidirectional language encoding, the output of the pre-training phase consists of extracted textual features, denoted as .

Convolutional calculation module: The extracted text features are encoded into embedding vectors. Based on the spatial transformation and feature reorganization of vectors, the vector alternation matrix and the feature alternation matrix are constructed. The embedding vectors undergo feature extraction using one-dimensional convolutional kernels of size 3*1 and two-dimensional convolutional kernels of size 3*3, resulting in the generation of convolutional feature maps. The feature maps are fused using matrix addition, and the fused features are then used as inputs to the fully connected layer, ultimately enabling the joint extraction of entity relationships. The objective function of the model is defined as follows.
(8)
Among them, represents vector concatenation, represents the embedding matrix for entity relationships, W represents the trainable weight matrix, and g represents the Softmax function. For model training, the adoption of cross-entropy as the loss function is defined as follows.
(9)

Among them, N represents the count of entity relationships.

The research focuses on the comprehensive emergency plans at the provincial and municipal levels, as well as 17 specific emergency plans released from 2014 to 2021 by the South-to-North Water Diversion Project in China. The South-to-North Water Diversion Project in China is a large-scale inter-basin water diversion project implemented to alleviate water scarcity issues in northern China. Since its comprehensive implementation, the project has conveyed a total volume of 58.6 billion cubic meters of water, effectively alleviating water scarcity issues in northern China. The extraction of emergency entity relationships in the water diversion project, based on PTM–ICNN, primarily consists of five stages: text pre-processing, ontology construction, entity relationship extraction, knowledge graph triple storage, and knowledge retrieval and recommendation. Figure 4 presents the fundamental process flow of the method, which unfolds as follows.
  • (1)

    Text pre-processing: For emergency plan text data, firstly, eliminate frequently occurring but semantically insignificant words to reduce the dimension of the feature space and minimize noise interference on the model. Next, convert the textual data into numerical vector representations to provide inputs for model learning.

  • (2)

    Ontology construction: From the perspective of emergency management, the emergency response process primarily comprises four stages: forecasting and early warning, graded response, emergency handling, and post-event support. The forecasting and early warning stage involves factors such as rainfall, runoff, water levels, and reservoir dams. The emergency handling stage encompasses institutions, responsibilities, and more. The post-event support involves resources, equipment, personnel, and other considerations. The emergency knowledge graph ontology represents emergency concepts and their related relationships, comprising concept nodes and concept relation edges. The emergency knowledge graph ontology constructed in this study primarily encompasses five aspects of conceptual relationships. (1) The emergency response process encompasses four distinct stages: forecast and warning, graded response, emergency disposal, and post-event support. (2) Natural entities such as meteorological conditions, hydrology, and rivers. (3) Institutions, departments, and other social entities. (4) Reservoirs, dams, and other hydraulic engineering entities. (5) Various specialized terminologies in the field of hydraulics.

  • (3)

    Joint extraction of entity relationships: The emergency plan document for the South-to-North Water Diversion Project in China primarily comprises two types of data: semi-structured tabular data and unstructured textual data. Furthermore, there are structured real-time monitoring data available. Regarding structured real-time data, it is directly stored in a relational database. As for semi-structured tabular data, it is extracted using regular expressions for matching. For unstructured textual data, extraction is performed using the PTM–ICNN model. Initially, train the PTM–ICNN model using well-annotated training datasets. Subsequently, employ the trained PTM–ICNN model to extract entity relationships from unannotated emergency plan texts. Finally, integrate the extracted entity relationships to obtain comprehensive and precise entity relationship triplets.

  • (4)

    Storage of knowledge graph triplets: Following the processes of named entity recognition and relationship extraction, the emergency data of the water diversion project has been transformed from unstructured textual data into structured knowledge triplets. This study utilizes the Neo4j graph database to store the knowledge triplets of entity relationships. When storing entity nodes, each node is identified by a unique identifier such as an ID or URI. When storing relationship edges, the semantic relationships between entities are represented using the relationship type and direction between nodes. The emergency knowledge graph adopts the knowledge representation method of triples, namely ‘entity–relation–entity’ and ‘entity–property–property value’. The graph traversal method combines conditional traversal and depth-first traversal techniques. Depth-first traversal involves starting from a single node and systematically exploring connected nodes by following specified relationship types and depth constraints until reaching the desired depth of nodes. Conditional traversal involves incorporating conditional statements during graph traversal to filter and restrict nodes based on specific criteria. Regarding the query mechanism of Neo4j, it relies on the Cypher query language, which offers statements such as MATCH, CREATE, and DELETE, among others.

  • (5)
    Knowledge retrieval and recommendation: Knowledge retrieval and recommendation often employ text similarity algorithms to compute the similarity between the target text and the candidate text. Prominent text similarity algorithms, such as jaccard and word2vec, are primarily built upon word embedding techniques and the cosine theorem. However, this approach often only considers shared vocabulary among texts, neglecting the overall semantic information and contextual correlations. The knowledge retrieval method based on mutual information criterion considers the co-occurrence frequency of vocabulary, enabling a more comprehensive reflection of the semantic similarity between texts. Furthermore, the approach based on mutual information criterion exhibits higher sensitivity toward the co-occurrence of rare vocabulary, enabling more effective handling of domain-specific knowledge retrieval and matching, thereby enhancing the reliability of knowledge recommendations. The retrieval and recommendation of emergency plans for water diversion projects require a predetermined similarity threshold. When the input content meets the similarity threshold, the recommended solution can be directly outputted. Otherwise, utilizing the mutual information criterion for entity similarity matching, the returned set of recommendations comprises the top 5 solutions with the highest similarity. The formula for calculating similarity can be defined as follows.
    (10)
Figure 4

PTM–ICNN-based emergency knowledge graph construction and retrieval.

Figure 4

PTM–ICNN-based emergency knowledge graph construction and retrieval.

Close modal

Among them, represents the input entity, represents the entity to be retrieved, and represents the joint probability distribution of the two random variables.

This section validates the effectiveness of the proposed method through joint extraction of emergency entity relationships, hyperparametric sensitivity analysis, ablation experiments and case study experiments. Firstly, an explanation is provided regarding the datasets, baseline models, and evaluation indexes utilized in the experiment. Furthermore, the experiment is divided into four groups to achieve various research objectives.

  • (1)

    In the experiment of entity relationship joint extraction on emergency plan data for water diversion projects, the PTM–ICNN model was utilized, and the experimental results were compared and analyzed with previous state-of-the-art models to evaluate the effectiveness of the PTM–ICNN model;

  • (2)

    Conducting a sensitivity analysis of the hyperparameters aims to validate the model's performance sensitivity to different parameter settings;

  • (3)

    By conducting ablation experiments, the impact of entity masking-based pre-training tasks and interactive convolution on the performance of PTM–ICNN can be analyzed;

  • (4)

    Conducting a case study experiment on intelligent recommendations for emergency plans, the experiment aims to intricately demonstrate how our proposed method ensures the reliability of plan recommendations.

Datasets

This study utilizes the comprehensive emergency plans released by provincial and municipal levels, as well as 17 specialized emergency plans from the South-to-North Water Diversion Project in China between 2014 and 2021, as experimental data. The comprehensive emergency plans at the provincial and municipal levels, as well as the specific contingency plans, all originate from the 47 subordinate management offices of the South-to-North Water Diversion Project in China. The provincial and municipal-level emergency plans delineate the unified emergency response procedures for significant risk events within the respective province or city. The process comprises four stages: forecasting and early warning, graded response, emergency handling, and post-event support. There are 17 specialized emergency plans, encompassing major risk events such as floods, fires, traffic accidents, earthquakes, water pollution, and others. These plans outline the emergency response procedures for each respective event. After data pre-processing, a total of 733,041 records were obtained, which included provincial and municipal emergency plans as well as specialized emergency plans. Employing a 5-fold cross-validation method, the emergency plan data was divided into five sets, distributed according to predetermined proportions. Among these, four sets comprising a total of 586,432 records were utilized as the training set, while the remaining set of 146,608 records served as the validation set. The aforementioned steps were repeated five times, with each iteration using a distinct validation set. Ultimately, the average of the five validation results was taken as the model's performance level, thereby assessing the overall effectiveness of the model.

Baseline models

This study selects two widely used models, the individual model, and the hybrid model, as the baseline models. By comparing the two types of models, a more comprehensive understanding of the applicability and reliability of different models in entity relation extraction tasks can be achieved. The individual models comprise FastText (Oral et al. 2020), TextCNN (Bu et al. 2022), and BILSTM (Zhou et al. 2020; Lin et al. 2022). The hybrid models consist of BERT–CNN (Qi et al. 2021) and BERT–BILSTM–CRF (Wang et al. 2021; Meng et al. 2022). Regarding the model's parameter configuration, excessively high or low learning rates can lead to an increase in the model's loss or a decrease in convergence speed. An excessively high or low dropout rate can lead to model overfitting or underfitting. In this study, we ensure the fundamental performance of the model while comprehensively considering the fairness and reliability of the experiments, thereby eliminating the impact of parameter variations on the experimental results. Overall, the parameter configurations of the PTM–ICNN model and the baseline model are presented in Table 1.

Table 1

Values of the model parameters

HyperparametersDescriptionValues
Hidden layer dimension Number of hidden layer neurons in the network. 32 
Embedding dimension Vector dimensionality when converting text data to vector representation. 200 
Dropout rate The proportion or probability of randomly discarding neurons. 0.2 
Epochs Number of model training times. 500 
Max length Enter the maximum length of the sentence 512 
Batch size Number of input sentences for one round of model iteration 32 
Learning rate Learning rate of the model 0.0001 
Optimizer Model optimizer Adam 
HyperparametersDescriptionValues
Hidden layer dimension Number of hidden layer neurons in the network. 32 
Embedding dimension Vector dimensionality when converting text data to vector representation. 200 
Dropout rate The proportion or probability of randomly discarding neurons. 0.2 
Epochs Number of model training times. 500 
Max length Enter the maximum length of the sentence 512 
Batch size Number of input sentences for one round of model iteration 32 
Learning rate Learning rate of the model 0.0001 
Optimizer Model optimizer Adam 

Evaluation indexes

For model evaluation, this study employs accuracy, recall, and F1 score as the evaluation indexes. Accuracy measures the proportion of correctly identified entity relationships by the model. Recall measures the model's ability to correctly identify the number of entity relationships. The F1 score is the harmonic mean of accuracy and recall, used to evaluate the overall performance of a model. Using evaluation indexes such as accuracy, recall, and F1 score for assessing the entity relation extraction task offers the advantages of comprehensive evaluation, addressing sample imbalances, and ensuring strong interpretability. The higher the values of accuracy, recall, and F1 score, the better the performance of the model. Accuracy (P), recall (R), and F1 score are defined as follows.
(11)
(12)
(13)

Among them, denotes the number of correctly predicted entities, denotes the number of predicted entities that are not actual entities, and denotes the number of entities that were not predicted.

Analysis of the results of the joint extraction of entity relationships

Figure 5 and Table 2 present the experimental results of different models. In terms of entity extraction, FastText exhibited the lowest performance with accuracy, recall, and F1 scores of 0.828, 0.873, and 0.836, respectively. FastText utilizes character-level n-gram features to represent text, which exhibits limited capability in comprehending complex semantics. The BERT–BILSTM–CRF model achieves excellent performance with accuracy, recall, and F1 scores of 0.925, 0.938, and 0.923, respectively. Based on pre-training and a bidirectional language model, BERT–BILSTM–CRF possesses the ability to deeply comprehend the intricate relationships between vocabulary, syntax, and semantics, thereby offering a more comprehensive contextual representation. The PTM–ICNN model proposed in this study achieved an accuracy, recall, and F1 score of 0.946, 0.973, and 0.959, respectively. Compared to BERT–BILSTM–CRF, PTM–ICNN exhibited improvements of 2.27% in accuracy, 3.73% in recall, and 3.90% in F1 score. In the domain of emergency knowledge extraction for water diversion projects, emergency plan texts exhibit numerous specialized terminologies and complex entity naming conventions. The BERT–BILSTM–CRF pre-training strategy based on character masking fails to fully mask the entire entity, leading to biases in the model's predictions for masked entities and consequently resulting in erroneous forecasts. Hence, in the domain of emergency entity extraction for water diversion projects, BERT–BILSTM–CRF exhibits inferior performance compared to PTM–ICNN. In relation extraction, FastText achieves an accuracy of 0.337, a recall of 0.246, and an F1 score of 0.115. BERT–BILSTM–CRF achieves an accuracy of 0.830, a recall of 0.817, and an F1 score of 0.817. PTM–ICNN achieves an accuracy of 0.872, a recall of 0.87, and an F1 score of 0.868. In comparison to BERT–BILSTM–CRF, PTM–ICNN demonstrates a notable enhancement in accuracy, recall, and F1 score, with improvements of 5.06, 6.61, and 6.24%, respectively. The emergency entity relationships in the water diversion project exhibit characteristics of uncertainty and complexity. Contextual information plays a vital role in accurately identifying entities and determining their relationships. The n-gram-based FastText model lacks the ability to capture such contextual information, resulting in the lowest performance in entity relationship extraction. Compared to the BERT–BILSTM–CRF relation extraction model, the relationship extraction model proposed in this paper, based on interactive convolution, exhibits stronger capability in handling complex relationship data and demonstrates superior performance. Overall, the PTM–ICNN proposed in this paper outperforms the current state-of-the-art models in terms of entity relation extraction, indicating its effectiveness in performing the task of extracting emergency entity relations in water diversion projects.
Table 2

Evaluation index values for each model

ModelsEntity extraction
Relationship extraction
PRF1PRF1
PTM–ICNN 0.946 0.973 0.959 0.872 0.871 0.868 
FastText 0.828 0.873 0.836 0.337 0.246 0.115 
BERT–CNN 0.891 0.919 0.892 0.609 0.597 0.547 
TextCNN 0.805 0.858 0.827 0.525 0.315 0.241 
BERT–BILSTM–CRF 0.925 0.938 0.923 0.830 0.817 0.817 
BILSTM 0.796 0.822 0.808 0.815 0.804 0.798 
ModelsEntity extraction
Relationship extraction
PRF1PRF1
PTM–ICNN 0.946 0.973 0.959 0.872 0.871 0.868 
FastText 0.828 0.873 0.836 0.337 0.246 0.115 
BERT–CNN 0.891 0.919 0.892 0.609 0.597 0.547 
TextCNN 0.805 0.858 0.827 0.525 0.315 0.241 
BERT–BILSTM–CRF 0.925 0.938 0.923 0.830 0.817 0.817 
BILSTM 0.796 0.822 0.808 0.815 0.804 0.798 

Values in bold indicate the maximum value of the current metric.

Figure 5

Comparison of performance levels by model: (a) entity extraction assessment results and (b) relationship extraction assessment results.

Figure 5

Comparison of performance levels by model: (a) entity extraction assessment results and (b) relationship extraction assessment results.

Close modal
To thoroughly validate the effectiveness of the proposed model, PTM–ICNN was subjected to 200 statistical experiments, along with the baseline model. The model's performance was evaluated and tested using four statistical variables: mean, median, 10th quantile, and 90th quantile. Among these, the mean is used to describe the overall average performance level of the model. The median describes the central tendency of the model's performance. Unlike the mean, the median is less sensitive to extreme values in the data, making it a better indicator of the central trend of the model's performance. The 10th quantile and 90th quantile, respectively, denote the values at the 10 and 90% positions in the data, providing information about the lower and upper bounds of the data distribution. Figure 6(a)–6(c) illustrates the accuracy, recall, and F1 scores of different models and statistical variables in the context of entity extraction. It is evident that the quantiles of BERT–BILSTM–CRF remain at approximately 93% for both the median and mean, while the 10th quantile and 90th quantile are maintained at 89 and 95%, respectively, resulting in a quantile range of 0.6. The mean and median of PTM–ICNN consistently hover around 95%, while the 10th quantile and 90th quantile remain steady at 95 and 96%, respectively, resulting in a quantile difference of 0.1. The mean and median of PTM–ICNN surpass those of the BERT–BILSTM–CRF model, while the difference between the 10th and 90th quantiles is smaller than that of BERT–BILSTM–CRF. This indicates that PTM–ICNN exhibits superior overall performance compared to BERT–BILSTM–CRF and is relatively stable. Figures 6(d)–6(f) illustrate the accuracy, recall, and F1 scores of different models and statistical variables in the context of relation extraction. The mean and median of BERT–BILSTM–CRF remain around 80%, with the 10th and 90th quantiles maintaining values of 77 and 82%, respectively. The difference between the quantiles is 0.5. The mean and median of PTM–ICNN remain around 87%, with the 10th and 90th quantiles maintaining values of 85 and 87%, respectively. The difference between the quantiles is 0.2 indicating that PTM–ICNN outperforms BERT–BILSTM–CRF in relation extraction and exhibits relative stability.
Figure 6

Performance levels of different models with different statistical variables: (a) level of accuracy of entity extraction with different statistical variables for different models; (b) level of recall of entity extraction with different statistical variables for different models; (c) F1 level of entity extraction with different statistical variables for different models; and (d) accuracy level of relationship extraction with different statistical variables for different models; (e) recall level of relationship extraction with different statistical variables for different models; and (f) F1 level of relationship extraction with different statistical variables for different models.

Figure 6

Performance levels of different models with different statistical variables: (a) level of accuracy of entity extraction with different statistical variables for different models; (b) level of recall of entity extraction with different statistical variables for different models; (c) F1 level of entity extraction with different statistical variables for different models; and (d) accuracy level of relationship extraction with different statistical variables for different models; (e) recall level of relationship extraction with different statistical variables for different models; and (f) F1 level of relationship extraction with different statistical variables for different models.

Close modal

Hyperparametric sensitivity analysis

To investigate the impact of hyperparameters on the performance of the PTM–ICNN model, this study conducted a detailed analysis and comparison of several key hyperparameters, including the entity relationship embedding dimension, dropout rate, and number of iterations. The entity relationship embedding dimension is denoted as L ∈ {40, 80, 120, 160, 200, 240, 280, 320, 360, 400, 440, 480, 520, 560, 600}. The dropout rate is denoted as R ∈ {0.04, 0.05, 0.06, 0.07, 0.08, 0.09, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9}. The number of iterations is denoted as E ∈ {10, 20, 30, 40, 50, 60, 70, 80, 90, 100, 110, 120, 130, 140, 150}. To ensure experimental fairness, all hyperparameters except those under current investigation were set to the same values as discussed in Section 5.2. The experimental results are presented in Figure 7.
Figure 7

Performance levels of the PTM–ICNN model with different hyperparameters: (a) accuracy index level for the number of iterative rounds and entity relationship embedding dimension; (b) recall index level for the number of iterative rounds and entity relationship embedding dimension; (c) F1 index level for the number of iterative rounds and entity relationship embedding dimension; (d) accuracy index level for the number of iterative rounds and dropout rates; (e) recall index level for number of iterative rounds and dropout rates; and (f) F1 index level for number of iterative rounds and dropout rates.

Figure 7

Performance levels of the PTM–ICNN model with different hyperparameters: (a) accuracy index level for the number of iterative rounds and entity relationship embedding dimension; (b) recall index level for the number of iterative rounds and entity relationship embedding dimension; (c) F1 index level for the number of iterative rounds and entity relationship embedding dimension; (d) accuracy index level for the number of iterative rounds and dropout rates; (e) recall index level for number of iterative rounds and dropout rates; and (f) F1 index level for number of iterative rounds and dropout rates.

Close modal

Figure 7(a)–7(c) illustrates the relationship between dropout rate, number of iteration rounds, and evaluation indexes. When the dropout rate is less than 0.2, the model exhibits excellent performance and remains relatively stable. When the dropout rate exceeds 0.2, the model's performance deteriorates and its stability decreases. At a dropout rate of 0.9, the model performs the worst. The experimental results indicate that the dropout rate is a sensitive parameter. Setting a dropout rate too high implies that more neurons are randomly discarded during the training process, leading to a weakening of the network's expressive capability. This could potentially lead to the model encountering the issue of underfitting, where the model fails to adequately fit the data, resulting in a lower accuracy level. Figure 7(d)–7(f) illustrates the relationship between entity relation embedding dimension, number of iteration rounds, and evaluation indexes. The model performs well when the embedding dimension is in the range of 80–200. When the embedding dimension of the model exceeds 200, overfitting occurs, leading to a decline in overall performance. The experimental results also demonstrate that the embedding dimension for entity relations is a sensitive hyperparameter. When the model's embedding dimension is set too large, it implies that each entity feature is represented as a lengthy vector. Embedding with such high dimensionality will escalate the quantity of model parameters, resulting in a more intricate model structure, thereby inducing the issue of overfitting.

Ablation experiments

To analyze the impact of entity-masked training tasks and interactive convolution on model performance. In this study, two variant models, namely PTM–ICNN–CUT and PTM–ICNN–UNI, were designed based on the PTM–ICNN model. PTM–ICNN–CUT denotes the exclusion of the entity masking pre-training task from the PTM–ICNN model. PTM–ICNN–UNI represents the exclusion of interactive convolution from the model. Table 3 displays the average values of accuracy, recall, and F1 score for the PTM–ICNN model and its variant models in the entity relation extraction task.

Table 3

Results of the index comparison between PTM–ICNN and variant models

ModelsPRF1
PTM–ICNN–CUT 0.749 0.679 0.712 
PTM–ICNN–UNI 0.896 0.898 0.896 
PTM–ICNN 0.909 0.922 0.913 
ModelsPRF1
PTM–ICNN–CUT 0.749 0.679 0.712 
PTM–ICNN–UNI 0.896 0.898 0.896 
PTM–ICNN 0.909 0.922 0.913 

Values in bold indicate the maximum value of the current metric.

According to Table 3, the PTM–ICNN–CUT model demonstrates accuracy, recall, and F1 scores of 0.749, 0.679, and 0.712, respectively. The PTM–ICNN–UNI model achieved accuracy, recall, and F1 scores of 0.896, 0.898, and 0.896, respectively. The PTM–ICNN model achieved accuracy, recall, and F1 scores of 0.909, 0.922, and 0.913, respectively. Compared to the PTM–ICNN–CUT model, the PTM–ICNN model exhibited an improvement of 21.36% in accuracy, 35.79% in recall, and 28.23% in F1 score. The removal of the entity masking pre-training task in the PTM–ICNN–CUT model has resulted in a diminished ability to extract domain entities from emergency plans, leading to a decline in model performance. Compared to PTM–ICNN–UNI, PTM–ICNN demonstrated improvements of 1.50% in accuracy, 2.63% in recall, and 1.93% in F1 score. By removing the interactive convolution from PTM–ICNN–UNI, the model's capability to extract complex relationships has been weakened, resulting in a decline in overall performance.

Analysis of case experiment results

This study evaluates the reliability of the proposed method by measuring whether the recommended solution set includes the correct emergency plans. Table 4 presents the accuracy, recall, and F1 values for the top 1, 3, and 5 recommended solutions as correct solutions. The table reveals that the accuracy of the correct solution being ranked first in the recommended solution set is 72.32%, while it is 74.30% for the top three and 75.36% for the top five. Considering the current recommendation outcomes in conjunction with the actual conditions of the South-to-North Water Diversion Project, they largely meet the demand for accurate recommendations of emergency plans. For the minority of correct solutions that are not included in the recommended set, the primary causes leading to the failure of recommendations are as follows: (1) Due to potential annotation errors during manual supervision of data labeling, the entity recognition algorithm may inaccurately identify certain entities. (2) The insufficient amount of training data for the entity recognition model resulted in inaccurate identification of relevant entities. (3) The excessive presence of similar entities within certain emergency knowledge results in the correct solution being located beyond the top five recommended options.

Table 4

Evaluation index values for the top 1, 3, and 5 scenarios in the recommendation set as the correct scenario

PRF1
Top 1 72.32 72.55 71.96 
Top 3 74.30 74.51 74.13 
Top 5 75.36 75.49 75.05 
PRF1
Top 1 72.32 72.55 71.96 
Top 3 74.30 74.51 74.13 
Top 5 75.36 75.49 75.05 

In response to the issues of weak knowledge association, poor timeliness, and insufficient intelligent decision support in the emergency plans for water diversion projects. Based on the knowledge graph technology, this paper proposes a knowledge-driven intelligent recommendation method for emergency plans in water diversion projects, yielding the following main conclusions:

  • (1)

    Taking into consideration the abundant domain knowledge present in the text of emergency plans for water diversion projects, we propose a pre-training task based on entity masking. Compared to BERT–BILSTM–CRF, PTM–ICNN showed improvements of 2.27% in accuracy, 3.73% in recall, and 3.90% in F1 score for entity extraction, effectively meeting the extraction needs of emergency entities in water diversion projects.

  • (2)

    To address the issue of complex relationships in emergency plan texts, an ICNN has been proposed. The experimental results demonstrate that PTM–ICNN achieved accuracy, recall, and F1 score of 0.872, 0.871, and 0.868, respectively, making it the top-performing model. This model exhibits stronger capabilities in handling complex relationships.

  • (3)

    The entity relationship triplets are stored and the emergency knowledge graph of the water diversion project is constructed using the Neo4j graph database.

  • (4)

    The proposed work is an emergency plan retrieval and recommendation method based on the mutual information criterion, with a recommendation accuracy of 75.36% for the top-5. Considering the actual operational conditions of the South-to-North Water Diversion Project, it can effectively meet the requirements for accurately recommending emergency plans.

However, the approach presented in this paper has not taken into account the multimodal information in emergency plan data, leading to a gap between knowledge graph-based emergency plan recommendations and practical engineering applications. Future work will focus on addressing the challenge of extracting multimodal knowledge and investigating the construction of a multimodal knowledge graph to enhance the effectiveness of our approach.

L.W., X.L., and Y.L. conceptualized the study; L.W. and H.L. did data curation; L.W. and J.L. did formal analysis; Y.L. acquired funds; L.W. and X.L. performed the methodology; L.W., X.L. and Y.L. wrote the original draft; and L.W. and X.L. wrote, reviewed, and edited the article.

This work was supported in part by grants from the National Natural Science Foundation of China (72271091), Projects of Open Cooperation of Henan Academy of Sciences (220901008), and Major Science and Technology Projects of the Ministry of Water Resources (SKS-2022029).

This data was provided by the Ministry of Water Resources of China and the South-to-North Water Diversion Central Line Administration and is available on the official website of the Ministry of Water Resources of China: http://www.mwr.gov.cn/.

The authors declare there is no conflict.

Afan
H. A.
,
Ibrahem Ahmed Osman
A.
,
Essam
Y.
,
Ahmed
A. N.
,
Huang
Y. F.
,
Kisi
O.
,
Sherif
M.
,
Sefelnasr
A.
,
Chau
K. W.
&
El-Shafie
A.
2021
Modeling the fluctuations of groundwater level by employing ensemble deep learning techniques
.
Engineering Applications of Computational Fluid Mechanics
15
(
1
),
1420
1439
.
Banan
A.
,
Nasiri
A.
&
Taheri-Garavand
A.
2020
Deep learning-based appearance features extraction for automated carp species identification
.
Aquacultural Engineering
89
,
102053
.
Bouraoui
Z.
&
Schockaert
S.
2019
Automated rule base completion as bayesian concept induction
. In:
Proceedings of the AAAI Conference on Artificial Intelligence
, Vol.
33
(
01
).
Bu
W.
,
Yang
W.
,
Chen
D.
&
Ding
T.
2022
Chinese relation extraction based on characters and words fusion
. In:
Third International Conference on Artificial Intelligence and Electromechanical Automation (AIEA 2022)
, Vol.
12329
.
SPIE
.
Chang
C.
,
Tang
Y.
,
Long
Y.
,
Hu
K.
,
Li
Y.
,
Li
J.
&
Wang
C. D.
2022
Multi-Information Preprocessing Event Extraction With BiLSTM
–CRF
Attention for Academic Knowledge Graph Construction
. In:
IEEE Transactions on Computational Social Systems
.
Chen
Y.
2015
Convolutional Neural Network for Sentence Classification
.
MS thesis
,
University of Waterloo
.
Chen
C.
,
Zhang
Q.
,
Kashani
M. H.
,
Jun
C.
,
Bateni
S. M.
,
Band
S. S.
,
Dash
S. S.
&
Chau
K. W.
2022
Forecast of rainfall distribution based on fixed sliding window long short-term memory
.
Engineering Applications of Computational Fluid Mechanics
16
(
1
),
248
261
.
Chiticariu
L.
,
Li
Y.
&
Reiss
F.
2013
Rule-based information extraction is dead! long live rule-based information extraction systems!
. In:
Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing
.
Cichosz
P.
2018
A case study in text mining of discussion forum posts: classification with bag of words and global vectors
.
International Journal of Applied Mathematics and Computer Science
28
(
4
),
787
801
.
Gehring
J.
,
Auli
M.
,
Grangier
D.
,
Yarats
D.
&
Dauphin
Y. N.
2017
Convolutional sequence to sequence learning
.
International Conference on Machine Learning. PMLR
70
,
1243
1252
.
Han
K.
,
Xiao
A.
,
Wu
E.
,
Guo
J.
,
Xu
C.
&
Wang
Y.
2021
Transformer in transformer
.
Advances in Neural Information Processing Systems
34
,
15908
15919
.
Kreibich
H.
,
Van Loon
A. F.
,
Kai Schröter
K.
,
Ward
P. J.
,
Mazzoleni
M.
,
Sairam
N.
,
Abeshu
G. W.
,
Agafonova
S.
,
Aghakouchak
A.
,
Aksoy
H.
,
Alvarez-Garreton
C.
,
Aznar
B.
,
Balkhi
L.
,
Barendrecht
M. H.
,
Biancamaria
S.
,
Bos-Burgering
L.
,
Bradley
C.
,
Budiyono
Y.
,
Buytaert
W.
,
Capewell
L.
,
Carlson
H.
,
Cavus
Y.
,
Couasnon
A.
,
Coxon
G.
,
Daliakopoulos
I.
,
de Ruiter
M. C.
,
Delus
C.
,
Erfurt
M.
,
Esposito
G.
,
François
D.
,
Frappart
F.
,
Freer
J.
,
Frolova
N.
,
Gain
A. K.
,
Grillakis
M.
,
Grima
J. O.
,
Guzmán
D. A.
,
Huning
L. S.
,
Ionita
M.
,
Kharlamov
M.
,
Khoi
D. N.
,
Kieboom
N.
,
Kireeva
M.
,
Koutroulis
A.
,
Lavado-Casimiro
W
,
Li
H.-Y.
,
LLasat
M. C.
,
Macdonald
D.
,
Mård
J.
,
Mathew-Richards
H.
,
McKenzie
A.
,
Mejia
A.
,
Mendiondo
E. M.
,
Mens
M.
,
Mobini
S.
,
Mohor
G. S.
,
Nagavciuc
V.
,
Ngo-Duc
T.
,
Huynh
T. T. N.
,
Nhi
P. T. T.
,
Petrucci
O.
,
Nguyen
H. Q.
,
Quintana-Seguí
P.
,
Razavi
S.
,
Ridolfi
E.
,
Riegel
J.
,
Sadik
Md. S.
,
Savelli
E.
,
Sazonov
A.
,
Sharma
S.
,
Johanna Sörensen
J.
,
Souza
F. A. A.
,
Stahl
K.
,
Steinhausen
M.
,
Stoelzle
M.
,
Szalińska
W.
,
Tang
Q.
,
Tian
F.
,
Tokarczyk
T.
,
Tovar
C.
,
Tran
T. V. T.
,
Van Huijgevoort
M. H. J.
,
van Vliet
M. T. H.
,
Vorogushyn
S.
,
Wagener
T.
,
Wang
Y.
,
Wendt
D. E.
,
Wickham
E.
,
Yang
L.
,
Zambrano-Bigiarini
M.
,
Blöschl
G.
&
Baldassarre
G. D.
2022
The challenge of unprecedented floods and droughts in risk management
.
Nature
608
(
7921
),
80
86
.
Krizhevsky
A.
,
Sutskever
I.
&
Hinton
G. E.
2017
Imagenet classification with deep convolutional neural networks
.
Communications of the ACM
60
(
6
),
84
90
.
Lan
L. T. H.
,
Tuan
T. M.
,
Ngan
T. T.
,
Giang
N. L.
,
Ngoc
V. T. N.
&
Van Hai
P.
2020
A new complex fuzzy inference system with fuzzy knowledge graph and extensions in decision making
.
Ieee Access
8
,
164899
164921
.
LeCun
Y.
,
Boser
B.
,
Denker
J. S.
,
Henderson
D.
,
Howard
R. E.
,
Hubbard
W.
&
Jackel
L. D.
1989
Backpropagation applied to handwritten zip code recognition
.
Neural Computation
1
(
4
),
541
551
.
Lee
J. D. M. C. K.
&
Toutanova
K.
2018
‘Pre-training of deep bidirectional transformers for language understanding.’ arXiv preprint arXiv:1810.04805
.
Lin
H.
,
Gharehbaghi
A.
,
Zhang
Q.
,
Band
S. S.
,
Pai
H. T.
,
Chau
K. W.
&
Mosavi
A.
2022
Time series-based groundwater level forecasting using gated recurrent unit deep neural networks
.
Engineering Applications of Computational Fluid Mechanics
16
(
1
),
1655
1672
.
Ma
S.
,
Li
A.
,
Zhao
X.
&
Song
Y.
2021
Learning BiLSTM-based embeddings for relation prediction in temporal knowledge graph
.
In Journal of Physics: Conference Series
(Vol.
1871
, No.
1
, p.
012050
).
IOP Publishing
.
2021a
Learning BiLSTM-based embeddings for relation prediction in temporal knowledge graph
.
Journal of Physics: Conference Series
1871
(
1
).
IOP Publishing
.
Ma
Z.
,
Zheng
W.
,
Chen
X.
&
Yin
L.
2021b
Joint embedding VQA model based on dynamic word vector
.
PeerJ Computer Science
7
,
e353
.
Meng
F.
,
Yang
S.
,
Wang
J.
,
Xia
L.
&
Liu
H.
2022
Creating knowledge graph of electric power equipment faults based on BERT–BiLSTM–CRF model
.
Journal of Electrical Engineering & Technology
17
(
4
),
2507
2516
.
Milosevic
N.
,
Gregson
C.
,
Hernandez
R.
&
Nenadic
G.
2019
A framework for information extraction from tables in biomedical literature
.
International Journal on Document Analysis and Recognition (IJDAR)
22
,
55
78
.
Mykowiecka
A.
,
Marciniak
M.
&
Kupść
A.
2009
Rule-based information extraction from patients’ clinical data
.
Journal of Biomedical Informatics
42
(
5
),
923
936
.
Oral
B.
,
Emekligil
E.
,
Arslan
S.
&
Eryigit
G.
2020
Information extraction from text intensive and visually rich banking documents
.
Information Processing & Management
57
(
6
),
102361
.
Popovski
G.
,
Kochev
S.
,
Korousic-Seljak
B.
&
Eftimov
T.
2019
FoodIE: A Rule-based Named-entity Recognition Method for Food Information Extraction
. In
ICPRAM 12
, p.
915
.
Qi
T.
,
Qiu
S.
,
Shen
X.
,
Chen
H.
,
Yang
S.
,
Wen
H.
,
Zhang
Y.
,
Wu
Y.
&
Huang
Y.
2021
KeMRE: knowledge-enhanced medical relation extraction for Chinese medicine instructions
.
Journal of Biomedical Informatics
120
,
103834
.
Radford
A.
,
Narasimhan
K.
,
Salimans
T.
&
Sutskever
I.
2018
Improving Language Understanding by Generative pre-training
.
Tan
M.
&
Le
Q.
2019
Efficientnet: rethinking model scaling for convolutional neural networks
.
International Conference on Machine Learning. PMLR
6105
6114
.
Tang
W.
,
Zhang
X.
,
Feng
D.
,
Wang
Y.
,
Ye
P.
&
Qu
H.
2022
Knowledge graph of alpine skiing events: a focus on meteorological conditions
.
PLos one
17
(
9
),
e0274164
.
Tan
M.
&
Le
Q.
2019
Efficientnet: rethinking model scaling for convolutional neural networks
.
International Conference on Machine Learning.
PMLR
97
,
6105
6114
.
Yamashita
R.
,
Nishio
M.
,
Do
R. K. G.
&
Togashi
K.
2018
Convolutional neural networks: an overview and application in radiology
.
Insights Into Imaging
9
,
611
629
.
Zeng
F.
&
Wang
Q.
2022
Intelligent recommendation algorithm combining RNN and knowledge graph
.
Journal of Applied Mathematics
2022
,
1
11
.
Zhang
Z.
,
Han
X.
,
Zhou
H.
,
Ke
P.
,
Gu
Y.
,
Ye
D.
,
Qin
Y.
,
Su
Y.
,
Ji
H.
,
Guan
J.
,
Qi
F.
,
Wang
X.
,
Zheng
Y.
,
Zeng
G.
,
Cao
H.
,
Chen
S.
,
Li
D.
,
Sun
Z.
,
Liu
Z.
,
Huang
M.
,
Han
W.
,
Tang
J.
,
Li
J.
,
Zhu
X.
&
Sun
M.
2021
CPM: A large-scale generative Chinese pre-trained language model
.
AI Open
2
,
93
99
.
Zhou
B.
,
Bao
J.
,
Liu
Y.
&
Song
D.
2020
BA-IKG: BiLSTM Embedded ALBERT for Industrial Knowledge Graph Generation and Reuse
. In:
2020 IEEE 18th International Conference on Industrial Informatics (INDIN)
, Vol.
1
.
IEEE
.
Zhou
Z.
,
Wang
C.
,
Feng
Y.
&
Chen
D.
2022
JointE: Jointly utilizing 1D and 2D convolution for knowledge graph embedding
.
Knowledge-Based Systems
240
(
2022
),
108100
.
This is an Open Access article distributed under the terms of the Creative Commons Attribution Licence (CC BY 4.0), which permits copying, adaptation and redistribution, provided the original work is properly cited (http://creativecommons.org/licenses/by/4.0/).