COVID-19 evolves rapidly and an enormous number of people worldwide desire instant access to COVID-19 information such as the overview, clinic knowledge, vaccine, prevention measures, and COVID-19 mutation. Question answering (QA) has become the mainstream interaction way for users to consume the ever-growing information by posing natural language questions. Therefore, it is urgent and necessary to develop a QA system to offer consulting services all the time to relieve the stress of health services. In particular, people increasingly pay more attention to complex multi-hop questions rather than simple ones during the lasting pandemic, but the existing COVID-19 QA systems fail to meet their complex information needs. In this paper, we introduce a novel multi-hop QA system called COKG-QA, which reasons over multiple relations over large-scale COVID-19 Knowledge Graphs to return answers given a question. In the field of question answering over knowledge graph, current methods usually represent entities and schemas based on some knowledge embedding models and represent questions using pre-trained models. While it is convenient to represent different knowledge (i.e., entities and questions) based on specified embeddings, an issue raises that these separate representations come from heterogeneous vector spaces. We align question embeddings with knowledge embeddings in a common semantic space by a simple but effective embedding projection mechanism. Furthermore, we propose combining entity embeddings with their corresponding schema embeddings which served as important prior knowledge, to help search for the correct answer entity of specified types. In addition, we derive a large multi-hop Chinese COVID-19 dataset (called COKG-DATA for remembering) for COKG-QA based on the linked knowledge graph OpenKG-COVID19 launched by OpenKG①, including comprehensive and representative information about COVID-19. COKG-QA achieves quite competitive performance in the 1-hop and 2-hop data while obtaining the best result with significant improvements in the 3-hop. And it is more efficient to be used in the QA system for users. Moreover, the user study shows that the system not only provides accurate and interpretable answers but also is easy to use and comes with smart tips and suggestions.
The serious situation of COVID-19 is ongoing. By January 16, 2022, more than 5.54 million people had died from the plague2, raising increasing anxiety about health problems in individuals. The pandemic has severely affected people's lives, and people dramatically demand accurate, efficient, and instant access to epidemic information. However, large information about COVID-19 on various Web sites is not well organized and not specialized for the general public. Question Answering systems based on COVID-19 knowledge as a convenient interaction way are popular among more and more people. There are two existing paradigms for COVID-19 QA: Information Retrieval Question answering (IRQA) and Question Answering over Knowledge Graph (KGQA). The IRQA systems of COVID-19 are based on textual question-answer pairs [1, 2, 3, 4], getting answers by computing similarity between the asked question and questions/answers in the dataset. IRQA systems can naturally answer simple questions that people frequently ask. In contrast, KGQA methods over COVID-19 dataset [5, 6, 7, 8] give answers to complex questions covering multiple relations over structural KGs. Besides, KQGA techniques can reason for new knowledge in QA tasks.
On the other hand, the pandemic has been spreading for a long time until now, and people have some basic understanding of COVID-19. So people are no longer satisfied with asking simple questions, like “what are the clinical symptoms of patients with COVID-19?”. They are more inclined to express complex multi-hop questions, such as the 2-hop question that “What are the related diseases having similar symptoms to COVID-19?” and the 3-hop question that “how to check the related diseases having similar symptoms to COVID-19?”. So we choose to use multi-hop KGQA techniques to build COVID-19 QA system.
However, there are some limitations of existing KGQA techniques and current COVID-19 KGQA datasets. Existing methods [9, 10] often represent knowledge graph and questions by using separate models, raising issues that heterogeneous embeddings from different spaces should be fitted to a common space. Additionally, a schema that defines a useful, high-level structure of a KG has been neglected in the current multi-hop KGQA tasks . Schema information as important prior knowledge can be helpful to search for correct entities of specified types. What's more, public COVID-19 KGs [12, 13, 14] suffer from knowledge sparsity especially the knowledge people would like to ask for daily, which will further affect the quality of downstream QA tasks. In this paper, we improve KGQA performance by proposing COKG-QA: multi-hop Question Answering over COVID-19 Knowledge Graphs. COKG-QA proposes some improvements in terms of these constraints mentioned above. The architecture of our system is illustrated in Figure 1, and the main contributions of our paper are as follows:
We introduce COKG-QA to demonstrate the importance of embedding projection mechanism and schema information in multi-hop KGQA task②. More precisely, embeddings of entities, schema, and questions from different spaces are transferred into one common one by a projection method to align important features. Furthermore, entity embeddings are incorporated with its type embeddings to predict answers of specified types.
There rarely exist comprehensive KGQA datasets③ managed for COVID-19 especially lacking multi-hop questions. Benefiting from OpenKG-COVID19 , we derive a large multi-hop Chinese COVID-19 KGQA dataset, COKG-DATA. It consists of abundant knowledge, which provides an important foundation for building a superior question answering system.
Experiments in the paper prove that COKG-QA is of high quality and also robust to further generalize to new knowledge. In order to facilitate people's demand for COVID-19 consulting services, we develop a user-friendly interactive application④ based on COKG-QA. The application not only provides accurate and interpretable answers but also is easy to use and has functions of smart tips and recommendations.
Architecture of the COKG-QA system.
2. RELATED WORK
At the moment of the epidemic, researchers have released some relevant datasets and built question answering systems based on natural language processing techniques to help people conveniently obtain information about COVID-19. We introduce these significant efforts and KGQA techniques in this section.
COVID-19 datasets and QA systems: Some useful KGs have been launched to advance COVID-19 research during the ongoing pandemic. However, the published COVID-19 KGs have limited data size and are more academically medical, which are not applicable for users' daily consulting needs. For example, the coronavirus Knowledge Graph  has 27 relations and limited entity types. Publications, case statistics, and molecular data are structured  to explore biomedical knowledge, such as specific genes, proteins, etc. KG-COVID-19  also focuses on SARS-CoV-2 and COVID-19 related heterogeneous biomedical data to construct KGs. Based on the public KGs, some KGQA systems are developed. Template matching method  using Naive Bayes algorithm over a KG is adopted to establish a QA system of COVID-19. QA system like  employs a rule-based classifier for recognizing users' intentions and also adopts templates to parse natural questions of users. To make the framework not limited to predefined rules, some work like  introduces a relatively general framework based on the knowledge embedding method tranE . Although these QA systems are developed for COVID-19, they fail to provide optimal performance for users' diverse questions.
KGQA: There are many state-of-the-art KGQA methods, and we briefly review these three types : (1) logic-based methods; (2) path-based methods; (3) embedding-based Methods. Logic-based methods are widely discussed due to the advantages of high accuracy and strong interpretability. GQE (Graph Query Embedding) , Query2Box , BetaE  represent the query as a directed acyclic computational graph to generate logic form query embedding. Pathbased methods take the topic entity in the question to search along multiple triples of KG to find the answer entity or relation. To alleviate the issue that the search space of Path-Ranking Algorithm  is large, DeepPath  allows the path attributes to be controllable. Teacher-student network is adopted in NSM  to learn intermediate supervision signals. Some other works like [26, 27] regard KG reasoning as a sequential path decision process. Embedding based methods [11, 28] measure the similarity between question embeddings and candidate answer embeddings to get the right answer. For example, the state-of-the-art method EmbedKGQA represents questions by pre-trained model and represent knowledge graph embeddings by ComplEx , and select answer through the score function of ComplEx. Relational Graph Convolutional Networks (R-GCN) method  aggregates embeddings of specific multiple relations in KG to predict answers. Research that KGs incorporate text corpus based on embedding methods [9, 10, 31] also attract much attention.
3. TASK DEFINITION
To alleviate people's anxiety about health problems caused by the COVID-19 pandemic, we are determined to develop an effective KGQA system focusing on complex multi-hop questions. In addition, the functions of smart tips and recommendations make the QA system consumer friendly. Considering the questions tend to be asked daily, data derived from OpenKG-COVID19 will be curated elaborately and finally formed into the multi-hop KGQA dataset, COKG-DATA. Moreover, we propose COKG-QA extending the state-of-the-art EmbedKGQA model with simple and efficient modifications, so that it can achieve superior and practical performance for the QA system. In the following sections, we first describe the extension in COKG-QA and the details of COKG-DATA. Based on the two modules, the performance of the KGQA system will be demonstrated finally.
As mentioned in the related work, EmbedKGQA  is a good work considering multi-hop reasoning. We extend it in the following several aspects to achieve better performance in terms of accuracy and coverage in the context of COVID-19 question answering. We first give a brief introduction of EmbedKGQA and then describe our improvements in detail in the following subsections.
An instance triple in a KG can be represented as <h, r, t>, where h represents the head entity and t represents the tail entity linked by relation r. Given a set of entities E and relations R, a Knowledge Graph G is a set of triples K such that K ⊆ E × R χ E. KGQA task searches answer entity for a natural language question q including muti-hop relations over a KG. Inspired by EmbedKGQA, we also employ KG Embedding Module, Question Embedding Module, and Answer Selection Module in our method. In this paper, we extend EmbedKGQA over COKGDATA by adding Embedding Projection and Schema-Aware Module. In addition, we also add a Topic-Entity-Aware Filter at inference to predict answer entity only related to the topic entity in question. The architecture can be seen in Figure 2. Details are described as follows.
Overview of COKG-QA framework.
4.2 Embedding Projection
We regard embeddings generated by different models as heterogeneous. Like triples in instance level, <s h, r, s t> is a triple in schema level, where s h represents the head type and s t stands for the tail type linked by relation r. Schema embeddings of s h, s t ∊ E' are also trained by ComplEx  method to enhance searching answer, but schema model and instance model are trained separately. What's more, question embedding is produced by pre-trained model RoBERTa  which leverages quite another technique paradigm. Therefore, these three embeddings are heterogeneous. Even though it helps to maintain their characteristics of schema, instance, and question by separate models, it is hard to model representations in the final KGQA model. Fully Connected (FC) linear layers like “firewalls” can maintain and project important features in transfer learning , especially when the source domain and target domain are quite different. Therefore, it is reasonable to project these embeddings before being transferred into one common space. We respectively define question embedding, entity embedding, schema embedding by
where eq is question embedding. And en is entity embedding trained by instance triples, while e's_n is entity type embedding produced by triples in schema level.
4.3 Schema-Aware Module
Existing KGQA methods only focus on instance facts in the KG, which ignores the well-constructed prior knowledge in the schema. The schema contains valuable structure information of a KG, which defines concepts and properties of these concepts. Entities in KG are linked to their corresponding concepts by entity types [34, 35]. We add Schema-Aware Module by combining entity embedding with corresponding entity type embedding which will be helpful to filter answer entities of specified types. This is good enough for the model to understand which type of the topic entity is and which type of the answer entity will be. Specifically, the topic entity representation in the question and the tail entity representation as the answer is constructed by adding the corresponding entity type embedding. Question representation embedded by using RoBERTa can't encode relation embedding in the schema level because there is no relation type label for question in a real application. But we concatenate entity type with the given question to imply that the question is relevant with a certain entity type like the input shown in Figure 2.
And the specific function is
where φ is the ComplEx scoring function described in section 4.1 and Eh is the topic entity embedding and E's_h is its corresponding type embedding. Ea stands for the right answer entity and Eâ means negative entity. σ ∊ E is the set of answer entities. All these embeddings are all transferred by Embedding Projection.
4.4 Topic-Entity-Aware Filter
Because COKG-DATA we collect is very large, it's necessary to add a filter to get the topic entity⑤ related entities, including 1-hop, 2-hop, and 3-hop entities at inference like EmbedKGQA to predict more relevant answer entity. We first make a map between topic entities and its multi-hop entities with 3-hop number, and then we predict answers among the multi-hop entities based on the best-trained model.
5. COKG-DATA CONSTRUCTION
Existing COVID-19 QA systems fail to perform complex reasoning with a non-KG dataset. We organize COKG-DATA based on seven sub-KGs (i.e., encyclopedia, prevention, goods, medical, epidemiology, character) of OpenKG-COVID19 launched by OpenKG, which people are more prone to ask daily. COKG-DATA is a new challenging question-answer benchmark that contains single-hop questions and multi-hop questions concerning diseases, symptoms, drugs, etc. The overview of the selected graphs by COKG-DATA is depicted in Appendix A.1. With the large and diverse COKG-DATA, multi-hop KGQA is an appealing and useful task to satisfy people's complex query needs during the pandemic. We spend much time cleaning data based on OpenKG-COVID19 and collecting multi-hop questions. Details are shown in A.2.
5.1 Human Check
To make sure that the questions in COKG-DATA are natural and meaningful, we recruited four volunteers whose research fields are all Knowledge Graph and Question Answering to check the quality of the dataset. We got random samples in proportion to the number of the questions sorted by each relation defined in the cleaned OpenKG-COVID19. These four volunteers were asked to rate the sampled questions with three choices: 1 for Weird; 2 for Natural; 3 for Meaningful. We have optimized COKG-DATA four times by removing or modifying the weird question-answer pairs through the scoring process. The sampled number for the last turn is 4, 000, and the average score by volunteers is 2.8 demonstrating the high quality of COKG-DATA. The final statistics for each hop questions of COKG-DATA are shown in Table 1. COKG-DATA will keep up with OpenKG-COVID19 to update for more sufficient knowledge for users.
|Dataset .||Train .||Dev .||Test .|
|Dataset .||Train .||Dev .||Test .|
In this section, we first present the experimental setup, the COKG-QA results on COKGDATA, and then analyze answer errors.
6.1 Experimental Settings
We follow the same split proportion (i.e., 3:1:1) of train/validation/test for all datasets of 1-hop, 2-hop, and 3-hop questions. The number of each hop questions are summarized into Table 1. We choose batch size of 90, 64, 32 and corresponding learning rate 5e-5, 2e-5, 1e-6 for training model across 2 NVIDIA RTX2080ti GPUs. Additionally, we set the patience number as 10 meaning that it will stop training when the accuracy score has decreased ten times and the maximum limitation epoch is 100. ComplEx embedding was obtained based on OpenKE⑥ and the dimension of ComplEx embedding and question embedding in COKG-QA are all 400. Weight decay as a popular and necessary regularization technique was set as 1e-1.
We compare our model with two state-of-the-art models, including EmbedKGQA  and TransferNet . Since EmbedKGQA reasons answer through link prediction which can alleviate the KG incompleteness problem and avoid the problem of uneven distribution of data, we take extensions over it in our implementations. TransferNet is an Effective method and competitive enough as a baseline which achieves best performance on public multi-hop datasets, such as MetaQA , WebQSP , and CompWebQ .
EmbedKGQA  regard multi-hop KGQA task as link prediction and search for answer entity based on question embedding and knowledge embeddings, which mitigates the problem of KG incompleteness and can predict answer in unlimited neighbors.
TransferNet  proposes a unified architecture for label and text data. In this framework, TransferNet calculates the relations corresponding to different positions of the question under attention mechanism at each step and further gets the answer entity.
6.3 Main Results
In Table 2, we compare EmbedKGQA and TransferNet with COKG-QA on our COKGDATA datasets. COKG-QA performs better than EmbedKGQA in all hop data, while TransferNet outperforms COKG-QA in 1-hop and 2-hop questions. But TransferNet obtains the lowest accuracy in the 3-hop questions. TransferNET attends to different parts of the question to search for the corresponding relation at each step, which makes it sensitive to both the quality and quantity of each-hop relations in the graph. Therefore, we assume that the small amount of 3-hop data of COKG-DATA causes the bad performance for TransferNET. However, EmbedKGQA and COKG-QA both regard the multi-hop KGQA task as link prediction which takes a multi-hop relation as a single relation in KG Embedding Module. For example, each relation of “complication||commonly used medicine||usage and dosage”, “medication||medication ingredient” and “precaution” is equally seen as a single relation to put in one triple. So COKG-QA avoids the problem of data imbalance which is very common in the real world and poses challenge to neural models. What's more, TransferNET has a high complexity of computation and large memory storage problems because it computes the probability of an entity being activated as the answer entity for multi-times, which would also affect the inference speed.
|Model .||1-hop .||2-hop .||3-hop .|
|Model .||1-hop .||2-hop .||3-hop .|
|Model .||1-hop .||2-hop .||3-hop .|
|Model .||1-hop .||2-hop .||3-hop .|
Note: The results reported in this table are hits@1. The subscript of COKG-QAsam is named by the first letter of each word of Schema-Aware Module.
COKG-QAep for adding Embedding Projection, and COKG-QAteaf for Topic-Entity-Aware Filter.
6.4 Ablation studies
Table 4 shows ablation studies of the effects of adding Schema-Aware Module, adding Embedding Projection and Topic-Entity-Aware Filter. We demonstrate the importance of each improvement by leveraging the same train set, validation set, test set, and hyperparameters. We briefly analyze the effect of each component in this section.
|Dataset .||1-hop .||2-hop .||3-hop .|
|Dataset .||1-hop .||2-hop .||3-hop .|
6.4.1 Effect of Embedding Projection
Since all the entities embedding are frozen during COKG-QA training as EmbedKGQA does, the features of entities embedding are quite different from question embedding. Besides, entity embeddings and type embeddings are also learned from different trained models. So it is necessary to bridge a projection to transform these important features in different vector spaces into a common vector space. The comparison results of projection (in COKG-QAep row) and without projection can be seen in Table 3. Although Embedding Projection does not provide as much improvement as Schema-Aware Module, the 2.61% absolute improvement in 1-hop questions and soft better performance in other questions demonstrates that Embedding Projection advances the capability of COKG-QA compared to EmbedKGQA.
6.4.2 Effect of Schema-Aware Module
We concatenate entity embedding to the corresponding entity type embedding to build a contextual KG embedding for COKG-QA. Furthermore, an ablation test was performed to evaluate the effect of the only Schema-Aware Module. The results listed in Table 3 marked by COKG-QAsam show Schema-Aware Module leads to a better performance of an average increase by 1.82%, which indicates the effectiveness of enriching entity embedding by adding schema information.
Effect of Topic-Entity-Aware Filter
To select an answer entity in the range of the 3-hop neighborhoods of the topic entity, the filter could competitively deliver better inference results with more than a 10% increase, which further ensures to provide a robust QA system on COKG-DATA.
6.5 Answer analysis
To fully analyze the results of the experiments, we collect all wrong answers of the test set to try to find some useful reasons. Through observation, we find that the wrong samples containing digital numbers (in their digit or word form) account for 33.92%. And there are about 11.94% percentage entities including numbers in the selected sub-graphs, which is not a negligible data size. Numerical reasoning or discrete reasoning is a more challenging task  with only question-answer pairs supervision. Therefore, we experimented with two types of data, i.e, numerical question-answer pairs (inserted with numbers) and non-numerical question-answer pairs, to probe the impact of data types. We also tested their corresponding 2-hop, 3-hop questions. Table 4 shows the results for different types of datasets using our model.
Numerical Data Analysis. COKG-QA over only numerical data reaches 49.19% hits@1 in the 1-hop data with a 30.87% absolute decrease compared to the model with all data (without Topic-Entity-Aware Filter condition). It highlights the fact that it is harder to model text with numbers. Besides, entities and relations distributions in the numerical dataset are also observed and show that the uneven distributions may be another key factor for the worse performance. The right histogram in Figure 3 gives entities and relations distributions of Numerical Data.
Distribution of COKG-DATA by data types.
Non-numerical Data Analysis. As expected, non-numerical data with large samples is still hard to optimize, because non-numerical data accounts for the majority and the distribution of the non-numerical dataset is similar to all data. However, without the numerical problems, the experimental results of non-numerical data are better than of all data. The left histogram in Figure 3 presents the visualization of COKG-DATA distribution according to the first 30 multi-hop relations sorted by entity number. We can see that both numerical and non-numerical data have long-tail data problems, for which data augmentation to compensate  or enhancing the recall of long-tail entities  are directions that can be considered.
7. COKG-QA PERFORMANCE
The superior performance of COKG-QA illustrated by the extensive experiments above will promise an effective QA system. Therefore, we devise an interactive Web QA application based on COKG-QA for people. A friendly design of QA system can improve user experience [39, 40]. We discuss the considerations designed in the QA application in this section.
7.1 Interpretability of Answers
Unlike most KGQA systems giving direct answers, our system will explain the intermediate context for the multi-hop questions to make the answer for multi-hop questions interpretable. An answer will be inferred based on the best-trained model by computing ComplEx score. But the answer based on EmbedKGQA model is not understandable. For example, the answer to the 2-hop question “What are the types of drugs recommended for pediatric intracranial tumors” is “Chemical drugs, prescription drugs and medical insurance drug for work-related injury”, which would pose users a question like “what are the respective recommended drugs corresponding the drug types mentioned in the answer above?”. In other words, people not only want to achieve the final answer but also want to figure out what the intermediate results are. So we offer an interpretable answer “ The recommended drug for pediatric intracranial tumors glycerol fructose injection is a chemical drug; the recommended drug for pediatric intracranial tumors piracetam glucose injection is a medical insurance work injury drug…”. The process for the interpretable response is as follows: (1) When the QA system gets a multi-hop question, the topic entity will be recognized first. (2) Subsequently, the not direct tail answer is obtained by ranking scores based on the question and the recognized head. (3) To get an interpretable final answer, we need to search out the intermediate relations and get intermediate entities. Questions and corresponding multi-hop relations having the same head and answer labeled in the dataset are filtered out. Furthermore, we select the interpretable answer corresponding to the question in the dataset that has the same multi-hop relations or is most similar to the user's question to be the final response.
7.2 Sources of Answers
We give the sources of answers with the corresponding URL to help users to trace the context, which also increases the credibility of the system. The answer sources of our system give evidence by offering graph names in selected sub-graphs. Multiple graph names are shown if the user's question covers multiple linked graphs. The example can be seen in Figure 4.
User-friendly functions of our QA system.
7.3 Use Feedback
We design thumbs-up and thumbs-down buttons to encourage users to provide feedback, which will be used to improve the COKG-QA model. When users give positive feedback, the system will randomly generate a thank you sentence. When users thumb down, a bubble will pop up and three options are displayed for users: Incorrect answer, incomplete answer and customized opinions. The custom options provide space for users to flexibly come out with suggestions and further benefit to improve the effectiveness of the QA system.
7.4 Ease of Use
Many medical terms are uncommon or difficult to remember for users, such as disease names and treatments. The automatic input prompt function is significant and practical to improve the usability of the system. Our system supports autocompletion in many scenarios. For example, users can just use a single word, pinyin, first letters of multiple words, or even fuzzy search. Tips in the input box can expand the focus of users' queries to help complete questions that users want to ask as shown in Figure 5. Besides, our system can also recommend questions relevant to the topic entity, which allows users to explore more about the original question.
Usability of our QA system.
In this paper, we introduce a multi-hop KGQA method named COKG-QA to develop a QA system for COVID-19 consulting services and meet people's tailored medical information needs. Multi-hop KGQA techniques have attracted increasing attention of researchers for the ability to handle complex multi-hop questions and reasoning. We extend the state-of-the-art method EmbedKGQA by adding Embedding Projection and Schema-Aware Module in this paper. EmbedKGQA represents knowledge graph embedding based on ComplEx and represents questions using RoBERTa. Although it is reasonable and convenient to represent different specified embeddings, these representations come from heterogeneous vector spaces which will influence the optimal performance. We adapt the important features of questions and knowledge embeddings from different spaces into a common semantic one by adopting an embedding projection mechanism. What's more, current KGQA methods ignore the schema implication for entity representation. COKG-QA learns entity embeddings by summing their corresponding type information to help search for the right answer entity of specified types. And to ensure superior performance, we also add a Topic-Entity-Aware Filter to select the answer from the topic entity's neighbor entities in the 3-hop relation range. Furthermore, we publish a large multi-hop Chinese COVID-19 KGQA dataset COKG-DATA based on the open license of CC BY SA to provide a comprehensive knowledge foundation for COKG-QA. Extensive experiment results showed that COKG-QA is robust as a QA engine and can further generalize to new fields. Based on COKG-QA, we also develop a user-friendly interactive application. The application can generate interpretable answers and is easy to use with functions of smart tips and recommendations.
DATA AVAILABILITY STATEMENT
The data that support the findings of this study are available in the ScienceDB repository: https://doi.org/10.57760/sciencedb.02062. To reuse the data, please cite the data as: Du, H.F., et al.: COKG-QA: Multi-hop question answering over COVID-19 knowledge graphs. Data Intelligence 4(3), 2022. doi: https://doi.org/10.57760/sciencedb.02062.
This work was supported by the Fundamental Research Funds for the Central Universities with grant Nos. 22120220069 and the National Nature Science Foundation of China with Grant No. 62176185 and was supported in part by the Shanghai Artificial Intelligence Innovation and Development Fund grant 2020-RGZN-02026.
Du H.F. (firstname.lastname@example.org), Wang H.F. (email@example.com) designed the model architecture. Le Z.W. (firstname.lastname@example.org) participated in the discussion about the task definition of the project and lead the implementation of the collection of COKG-DATA and the training of COKG-QA with Du H.F.. Du H.F.also developed the Web application. Chen Y.W. (email@example.com) and Yu J. (firstname.lastname@example.org) added the contrast experiments and made result analysis. All the authors have made valuable contributions in writing and revising the manuscript.
APPENDICES A. Details of COKG-DATA
A.1 Overview of COKG-DATA
We elaborately select seven sub-graphs that contain topics people are more concerned about during the COVID-19 epidemic. The specific graphs selected by COKG-DATA are demonstrated as follows.
The encyclopedia KG gives us a general understanding of SARS-CoV-2 and COVID-19, and relevant viruses and diseases information.
The prevention KG provides prevention guidance published by the government for individuals, organizations in different places.
The goods KG is expanded around materials supply status during the epidemic, covering daily protective equipment, medical devices, and drugs.
The medical KG and the health KG are complementary to exploit COVID-19 related knowledge about various diseases, drugs, symptoms, examination methods, and hospitals.
The epidemiology KG employs the general techniques of epidemiology to study the distribution of diseases and influencing factors, exploring the causes of disease, clarifying the laws of epidemics for controlling and eradicating diseases effectively.
The character KG records concepts such as characters, battles, achievements for the pandemic, articles, resumes of heroes, etc.
A.2 Data Curation
Data cleaning. To ensure the quality of the QA dataset, we have cleaned some bad cases in OpenKG-COVID19 and removed triples that are not practical for QA: (1) some triples contain empty string, punctuation entities, or useless numbers; (2) some triples are weird to compose natural questions, e.g., (Doctors of Xinhua hospital, work in, Xinhua hospital) (3) the head entity is same with the tail entity in some triples, such as triples with “alias” relation. We filter out these bad triples described above and remove them. In addition, relation patterns including symmetry and inversion exist in OpenKG-COVID19. We extend triples for these relation patterns of OpenKG-COVID19. After data cleaning and relation extension, the knowledge graph dataset contains 112,246 entities, 209 relations, and 787,056 triples.
Multi-hop Questions Collection. We leverage fact triples in the selected sub-graphs of OpenKG-COVID19 as single-hop data. Further, we manually design 47 relations for 2-hop questions and 23 relations for 3-hop questions, in which the combined relations must be reasonable and natural. Specifically, the range of the front relation must be the same with the domain of the back relation in a 2-hop relation. For example, the range of “selected drug” relation is “drug” which must be consistent with the domain of “usage and dosage” in the 2-hop relation “Selected drug Usage and dosage”. The same rule applies to the 3-hop relations collection process. Similar to multi-hop dataset MetaQA , we employ neural translation models in Helsinki-NLP Opus-MT project⑦ to introduce more diverse and natural statements with the same meaning. Opus-mt-zh-en model is leveraged to translate sentences from Chinese to English, and then opus-mt-zh-en is used to translate back to Chinese. Furthermore, to create a large-scale unified knowledge base from the top level, entity alignment and relation alignment have been completed to eliminate inconsistency problems.
http://openkg.cn, the largest Chinese open knowledge graph community pushing the development of public KGs, open-source tools, and best practices in vertical sectors in China.
You can access the system at http://cokg-qa.openkg.cn/qa/
We are only concerned about questions with single topic entity in this paper.
http://openke.thunlp.org/, an Open-source Framework for Knowledge Embedding.
https://github.com/Helsinki-NLP/Opus-MT, a project offers tools and resources for open translation services