Coreference resolution in Vietnamese electronic medical records

Conclusion In this paper, we propose a system to resolve coreference in Vietnamese electronic medical records. Our contributions are threefold. First, to the best of our knowledge, our work is the first to explore this NLP problem on Vietnamese EMRs. Second, we discover and define rules to annotate verbs in a Vietnamese clinical corpus as their use is preferred to describe symptoms. Finally, our work shows that lexical similarity plays an important role in determining coreferential relationship among mentions in Vietnamese EMRs. By using bag-of-words vectors to encode the matching tokens, our system achieves an F1 score of 91.4%. These could provide a basis for further NLP research on Vietnamese EMRs when clinical texts from hospitals in Vietnam are more available. Despite having a high performance, there remains some unsolved cases. These include but not limited to detecting synonyms, hypernyms, and extracting contextual clues to distinguish non-corefential mentions when their lexical strings are the same. We suggest them for future works.

pdf11 trang | Chia sẻ: hachi492 | Ngày: 28/01/2022 | Lượt xem: 241 | Lượt tải: 0download
Bạn đang xem nội dung tài liệu Coreference resolution in Vietnamese electronic medical records, để tải tài liệu về máy bạn click vào nút DOWNLOAD ở trên
VNU Journal of Science: Comp. Science & Com. Eng., Vol. 34, No. 2 (2018) 33–43 Coreference Resolution in Vietnamese Electronic Medical Records Hung D. Nguyen1,∗, Tru H. Cao2 1Faculty of Information Technology, Monash University, Victoria, Australia 2Faculty of Computer Science and Engineering, Ho Chi Minh University of Technology, Ho Chi Minh City, Vietnam Abstract Electronic medical records (EMR) have emerged as an important source of data for research in medicine and information technology, as they contain much of valuable human medical knowledge in healthcare and patient treatment. This paper tackles the problem of coreference resolution in Vietnamese EMRs. Unlike in English ones, in Vietnamese clinical texts, verbs are often used to describe disease symptoms. So we first define rules to annotate verbs as mentions and consider coreference between verbs and other noun or adjective mentions possible. Then we propose a support vector machine classifier on bag-of-words vector representation of mentions that takes into account the special characteristics of Vietnamese language to resolve their coreference. The achieved F1 score on our dataset of real Vietnamese EMRs provided by a hospital in Ho Chi Minh city is 91.4%. To the best of our knowledge, this is the first research work in coreference resolution on Vietnamese clinical texts. Received 15 August 2018, Revised 16 November 2018, Accepted 25 December 2018 Keywords: Clinical text, support vector machine, bag-of-words vector, lexical similarity, unrestricted coreference. 1. Introduction Coreference resolution is the task of determining whether two mentions in a document refer to the same real-world entity, i.e. there exists an “identity” relation between them. This is a basic natural language processing (NLP) task that plays an important role in many applications such as question answering, text summarization, and machine translation. The problem of resolving coreference in texts has received a lot of attention among the ∗ Corresponding author. Email: dngu0042@student.monash.edu https://doi.org/10.25073/2588-1086/vnucsce.210 NLP community for the last 20 years. In the early days, the focus was primarily put on the general domain of mostly newswire corpora. Firstly approached with hand-crafted methods using discourse theories such as focusing or centering [1, 2], coreference resolution received the first learning-based treatment by Connolly et al. in 1994 [3] that casted it as a classification problem. Since then, several supervised models have been proposed to resolve coreference in the general domain, namely, the mention-pair model [4], the entity-mention model [5], and the ranking model [6]. Through achievements in the newswire 33 34 H.D. Nguyen, T.H. Cao / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 34, No. 2 (2018) 33–43 domain, recently this task has been investigated in other domains as well. One of them is the clinical domain, which proved to have critical applications but had been left with little attention [7]. To address this, i2b2 – one of the seven NIH-funded national centers for biomedical computing in USA – organized a shared task in 2011 where various teams joined to resolve coreference in English discharge summaries – a type of electronic medical records (EMR). This challenge was part of a series of effort to automatically extract knowledge from clinical documents and release annotated datasets to the NLP community. Containing vast and valuable medical knowledge, EMRs have significant potential in assisting medical practitioners with treatment and healthcare, such as predicting the possibility of diseases [8, 9] as well as facilitating the study of patients’ health. However, in Vietnam, EMRs are still at an early development stage as Vietnamese hospitals have just started to digitize them recently. Therefore to contribute to this development, we propose a method in this paper to resolve coreference among mentions in Vietnamese EMRs. To the best of our knowledge, our work is the first to explore this NLP problem in Vietnamese clinical documents. By doing this, we aim to provide a groundwork for future solutions and applications, especially when Vietnamese datasets are more mature and accessible. Similar to the general domain, the goal of a coreference resolution system in the clinical domain is to produce all coreferential chains for a given document, where each chain contains mentions referring to the same entity. For example, in the sentence “Bé ho từ hôm qua, ở nhà bé có uống thuốc nhưng không bớt ho”, both underlined mentions “ho” refer to the same symptom “cough”, hence they are put in the same chain. Mentions that do not corefer with any others are called singletons. These singletons can be viewed as single-mention chains to evaluate a coreference resolution system. According to our observation, verbs are often used in Vietnamese EMRs to describe abnormal behaviors or actions indicating tests/treatments. As can be seen in the example above, two mentions of the problem “ho” (cough) are used as verbs. Although the original coreference problem and the 2011 i2b2 shared task did not take coreference between verbs, and between a verb and a noun into account, the 2011 CoLNN challenge on unrestricted coreference considered such cases possible [10]. Motivated by this work, we also annotate verbs alongside nouns for coreference resolution in Vietnamese EMRs. I2b2 defined five different semantic classes to categorize mentions in the clinical domain, namely, Person, Problem, Test, Treatment, and Pronoun. The Person class is used for mentions referring to hospital’s staffs or patients and their relatives, whereas the three medical classes Problem, Test and Treatment represent those particular to the clinical domain. A coreferential chain can only belong to one of the first four classes because a pronoun refers to an entity of these classes. In our experiments, we use the same guidelines provided by i2b2 to annotate our dataset but extend it to include verbs as well. Our method in this paper takes both EMR’s texts and all labeled mentions as input, then produces coreferential chains as mentioned above. One thing to note, however, is that as we observed in our dataset, there lacks of Person and Pronoun mentions. For Person, only a small number of mentions are used and they mostly refer to the same patient. Similarly, it is not pronouns but rather hypernyms that are preferably used to refer to previously mentioned entities. Therefore, in this work we only consider coreference resolution among Problem/Test/Treatment mentions. 2. Related work In the general domain, some early methods for resolving coreference were heuristic or rule-based. H.D. Nguyen, T.H. Cao / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 34, No. 2 (2018) 33–43 35 They required sophisticated knowledge source or relied on computational theories of discourse such as centering or focusing. Since the 1990s, research in coreference has shifted its attention to machine learning approaches with the advents of three important classes of supervised methods, namely, the mention-pair model [4], the entity-mention model [5], and the ranking model [6]. The main idea of the mention-pair model is composed of two distinct steps. The first step is a pairwise classification process, where each pair of mentions is taken to determine its coreferential relation. In this step, simply generating all C2n pairs of mentions in the text often leads to too many negative pairs being present, which might introduce bias into the trained classifier. To tackle this issue, some works proposed heuristic methods for reducing the number of negative pairs [11, 12]. The second step of the mention-pair model involves constructing coreferential chains from the pairwise classification results. There are several methods for this task, including closest-first clustering [11], best-first clustering [13], correlation clustering [14], and graph partitioning algorithm [15]. Although many clustering algorithms have been proposed, only a few works attempted to compare their effectiveness. For example, best-first clustering was reported to have better performance than closest-first clustering in [13]. The entity-mention model treats coreference resolution as a supervised clustering problem by determining whether a mention belongs to a preceding cluster or not. This involves cluster-level features such as all relevance, most relevance or any relevance between the given mention and a cluster based on a certain aspect. For example, the relevance in terms of gender indicates whether the mention has the same gender as all, most, or any other mentions in the cluster. On the other hand, the ranking approach tries to rank mentions and chooses the best candidate to be an anaphora for an antecedent. To further improve the performance of these models, especially the mention-pair model, some works explored the topic of features design. The work in [16] stated that lexical features such as string matching, name alias, and apposition contribute the most to the effectiveness of these models. They also proposed some variations of the string matching feature to deal with cases where simple string matching is not sufficient. In that work, the authors treated two mentions as two bags of words and computed their similarity using a metric such as the dot product. In our system, we leverage this bag-of-words model as a way to provide more information about the matching tokens to improve the classifier’s performance. In the clinical domain, i2b2 introduced the 2011 shared task in which various teams competed to resolve coreference in clinical texts. Three classes of methods were used, namely, the rule-based, supervised, and hybrid ones [17, 18]. The system achieving the best result [19] is a supervised one that uses the mention-pair model and a wide range of features, including those from the general domain as well as the different characteristics of mentions in the clinical domain. To prevent class imbalance, the authors simply filtered out the obvious negative pairs where the two mentions belong to two different semantic classes. 3. Proposed method Taking the work in [19] as the basic idea, we also apply the mention-pair model to our Vietnamese corpus with the same instance filtering process. The input to our system includes both raw EMR’s content and all labeled mentions presented in the text. The overall process consists of the following steps (see Figure 1): preprocessing, generating pairs of mentions as classification instances, extracting features from these pairs and feed them to the SVM model to determine whether each pair is coreferential along 36 H.D. Nguyen, T.H. Cao / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 34, No. 2 (2018) 33–43 Medical records List of mentions Preprocessing Generating mention pairs Filtering pairs of different classes Extracting features SVM classification Best-first clustering Coreferential chains Fig. 1. The overall coreference resolution process with its confidence score, and finally producing coreferential chains using the best-first clustering algorithm in [13] that utilizes confidence scores from the previous step. The details of these steps are described in the following sections. 3.1. Preprocessing One of the main differences between English and Vietnamese language lies in the way words are constructed. In English, each lexical token represents a single word in most cases, while in Vietnamese a word can consist of one or multiple tokens. This is because each token in Vietnamese represents a single syllable rather than a word. Therefore in many situations, we need to distinguish between two or more single-syllable words and the multi-syllable one constructed from them [20]. Take two tokens “buồn” and “nôn” for example; when standing alone, these two represent two single-syllable words that have their own meanings (“sad” and “vomit” respectively). However, when combined together, they form a very different word “buồn nôn”, which means “nausea”. This characteristic of Vietnamese can affect important features such as string matching, which is the most influential feature in determining coreferential pairs. For example, while two mentions “buồn nôn” and “nôn” have their lexical strings partially matched, they represent two different health problems, which are “nausea” and “vomiting” respectively. For our system to be able to know which tokens should go together and which should stand alone depending on the context, we use the tool named vnTokenizer from [20] to segment words in the input text as well as to separate its sentences. The outcome of this step is that tokens which should be combined to form a multi-syllable word are grouped together using underscores (such as “buồn_nôn”), and each sentence is put on its own line. 3.2. Resolving coreference Generating mention pairs From n mentions in the input text, our system considers all C2n possible pairs and determines their coreferential relation. For the obviously negative cases where the two mentions belong to two different semantic classes, our system filters them out beforehand without the need to use the classifier. This step is necessary to avoid class imbalance, which heavily affects the classifier’s performance. H.D. Nguyen, T.H. Cao / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 34, No. 2 (2018) 33–43 37 Table 1. Examples of cases where partially matching tokens do not indicate coreferential relation Mention 1 Mention 2 Description nôn vomit buồn nôn nausea overlapping at syllable token “nôn”; some of these cases are solved by the preprocessing step where mention 2 becomes “buồn_nôn” đau bụng abdominal pain đầy bụng dyspepsia overlapping at modifier “bụng”; these cases state different symptoms occurring in the same body part ho nhiều khi thay đổi tư thế cough when changing position cảm giác khó thở khi nằm dyspnea when lying overlapping at preposition “khi” ho nhiều serious cough sổ mũi nhiều serious rhinorrhea overlapping at quantifier “nhiều”, which describes the seriousness of two different medical problems Extracting features for coreferential relations Each pair of mentions is represented by a feature vector containing useful information for our SVM classifier to determine their coreferential relation. As mentioned in the first section, because our dataset lacks mentions in Person and Pronoun classes, our system only resolves coreference among those from the three medical classes: Problem, Test, and Treatment. One observation we have in our dataset is that there tends to be simple medical terms and sentence constructions. The majority of cases where two mentions are coreferential are when their lexical strings are fully identical or have some matching tokens. On one hand, when two mentions are written exactly the same, they are very likely to be coreferential, and thus a simple boolean value is sufficient enough to inform our classifier. For this feature (called Full-String-Matching), we compare the lexical strings of two mentions, and set the value of the feature to 1 or 0 depending on whether they are equal to each other or not. For example, the value of Full-String-Matching will be 1 if the pair is (“ho”, “ho”), or 0 if the pair is (“nôn”, “sốt”). On the other hand, there are many cases such as (“sốt”, “sốt cao”) where the two mentions are not exactly identical, but they share some keywords indicating their coreferential relation. For this, we need to also extract a feature that compares the two mentions’ substrings (called Partial-String-Matching). However, a boolean value is not very usefull in this case because two mentions’ lexical strings can overlap at modifiers, prepositions, or syllable tokens, but not the actual words describing the medical problem, test or treatment. In cases of overlapping at syllable tokens, only some of them are solved by the preprocessing step but not all due to the low accuracy of the tool since its primary target is the general domain. Examples of some of these cases are shown in Table 1. To tackle the problem of partially matching tokens mentioned above, instead of using boolean value, we adapt the bag-of-words model used in [16] to encode our Partial-String-Matching feature. In [16], the authors actually measured the similarity between two bag-of-words vectors representing two mentions using a metric such as cosine-similarity. However in our method, we directly use the bag-of-words vector to represent the matching tokens and append it to the mention pair’s feature vector. This way, we can provide our classifier the exact tokens the two mentions overlap at. The bag-of-words vector is created using the binary scheme [16], which assigns weight 1 to a token if it occurs in the matching set, and 0 otherwise. To demonstrate it more clearly, suppose s1 and s2 are two sets of tokens taken from mentions m1 and m2 respectively. The 38 H.D. Nguyen, T.H. Cao / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 34, No. 2 (2018) 33–43 Table 2. Features used in our coreference resolution system Feature Description Lexical Full-String-Matching A boolean value indicating whether the two mentions have their string fully identical Partial-String-Matching A bag-of-words vector representing the matching tokens between the two mentions Distance Mention-Distance The number of mentions occurring between the two mentions Sentence-Distance The number of sentences occurring between the two mentions matching set sm is the intersection of s1 and s2, that is sm = s1 ∩ s2. The bag-of-words vector representing sm, denoted by vm, has its dimension equal to the vocabulary size of the training set. For each token, its corresponding vm’s element is assigned 1 if the token occurs in sm, and 0 otherwise. The value of Partial-String-Match feature is the vector vm. Along with Full-String-Matching and Partial-String-Matching features, we also use two other common features in the general domain to compute the distance between two mentions of a pair. One is the number of sentences in between (Sentence-Distance), and the other is the number of mentions in between (Mention-Distance). These distance features give our classifier useful hints based on this observation: the further the two mentions are from each other, the less likely they are coreferential. As stated in [19], besides lexical features, some other semantic clues in the text can also affect the coreferential relationship between two mentions even when their lexical strings are fully identical. For instance, different locations where the same medical problem appears, different times when the same test is conducted, or different ways of consuming the same drug. In Vietnamese EMRs, however, there seems to have little of such contextual information since most of the text is preferably organized by listing rather than narration. Therefore, we do not extract those semantic features. Still, our system achieves high performance by using only string matching and distance features as shown later in the Evaluation section. Table 2 summarizes all four features used in our system. Constructing coreferential chains In this step, our system takes the coreferential confidence scores of all mention pairs generated from the SVM classifier to make decision on how to form coreferential chains. We use the best-first clustering algorithm [13] for this step, in which for each mention, our system finds the best candidate such that this pair is coreferential and achieves the highest confidence score. Finally, the output coreferential chains are the results of chaining those pairs that have one mention in common. 4. Evaluation 4.1. Annotation guidelines As part of the raw clinical corpora, i2b2 also released guidelines assisting their annotators in marking the ground truths of interest in the corresponding tasks, such as mentions or coreferential chains. Since most of the works in English coreference define the problem for noun phrases only, i2b2’s guidelines also comply with this rule but extend it to include adjective phrases describing medical problems as well. As we observe in Vietnamese EMRs, verbs are often used to describe patient’s medical problems (especially symptoms) or actions taken to treat patients. However, the i2b2’s guidelines do not cover such cases in details but only state some specific examples involving verbs that should not H.D. Nguyen, T.H. Cao / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 34, No. 2 (2018) 33–43 39 Table 3. Examples of verbs that should and should not be annotated Should Should not “Cháu bị bệnh hai ngày nay. Ở nhà cháu ho, sốt”. Verbs that describe abnormal behaviors, such as “ho” (to cough) and “sốt” (to have a fever). “Kích cỡ của khối u :::: tăng ::: lên”. Verbs that indicate the outcome of an event. In this case, the verb “tăng lên” indicates that a tumor (“khối u” in the example) has grown in size. “Bệnh nhân được mổ ruột thừa”. Verbs that indicate actions performed to treat a patient. In this case “mổ ruột thừa” means to operate a surgery that removes the patient’s appendicitis. “Bệnh nhân được cho :::: uống thuốc hạ sốt”. Verbs that indicate the application of a treatment and that treatment is present in the sentence. In this case, the verb “uống” indicates the oral use of antipyretic (“thuốc hạ sốt”). “Bệnh nhân được :: đo huyết áp”. Verbs that indicate the application of a test and that test is present in the sentence. In this case, “đo” means to measure a patient’s blood pressure (“huyết áp”). be annotated. In 2011, the CoNLL challenge was organized to resolve unrestricted coreference that takes verbs into account and considers coreference related to verbs possible [10]. Take the text “Sales of passenger cars grew 22%. The strong growth followed year-to-year increases” from [10] for example; both underlined mentions refer to the same event and should be included in the system’s output. Motivated by this work, we have extended the current i2b2’s guidelines to include verbs where they, by themselves, describe abnormal behaviors related to medical problems or actions performed to treat patients. In cases where the name of a treatment or test is present and the verb is only used to describe their applications, it is not annotated as we adhere to the i2b2’s guidelines. Examples of our extended rules for verbs are shown in Table 3. 4.2. Dataset and experimental settings Our dataset is provided by a hospital in Ho Chi Minh city, whose name is confidential for data privacy, and consists of 687 raw text documents. To provide our system true labels for training and testing, we manually annotate mentions and coreferential chains from the dataset using the extended rules discussed above. Table 4 shows the statistics after we annotated the dataset. We evaluate our system using 5-fold cross validation on the entire dataset. We use LibSVM [21] to train and test our SVM models, which are configured with the Radial Basic Function (RBF) kernel. As recommended by LibSVM’s developers, the trade-off parameter C and the kernel parameter γ are chosen by performing a grid search on C ∈ {2−5, 2−3, . . . , 215} and γ ∈ {2−15, 2−13, . . . , 23}. 4.3. Evaluation metrics Similar to those in the 2012 i2b2 Challenge, our system is evaluated using three evaluation metrics, namely, MUC, B-CUBED, and CEAF. Each metric computes the precision and recall for each document. The unweighted average on a set of n documents is then computed to get the overall performance. Every F1 score is computed from the corresponding average precision and recall. Before we go into details, the followings are some terminologies used through out all of the three metrics: key refers to the set of manually annotated coreferential chains (the ground truth), denoted by G. response refers to the set of coreferential 40 H.D. Nguyen, T.H. Cao / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 34, No. 2 (2018) 33–43 Table 4. Statistics of the dataset Class No. mentions No. coreferential chains Problem 4122 747 Test 224 34 Treatment 1887 155 Total 6233 936 chains produced by a system, denoted by S . MUC metric This metric considers each coreferential chain as a list of links between pairs of mentions and evaluates a system based on the least number of incorrect links needed to be removed and missing links needed to be added to create a correct chain. These incorrect links and missing links can be considered as precision errors and recall errors respectively. From [22], we use the following formulas to compute the precision and recall for each document d: PMUC = ∑ s∈S (|s| − m(s,G))∑ s∈S (|s| − 1) RMUC = ∑ g∈G (|g| − m(g, S ))∑ g∈G (|g| − 1) where m(s,G) is calculated as the number of chains in G intersecting s plus the number of mentions in s not contained in any chain in G. B-CUBED metric The B-CUBED metric (or B3) evaluates a system by giving a score to each mention in a document rather than relying on the links in coreferential chains [23]. According to the authors, this metric addresses the two following weaknesses in the MUC scorer: 1. It does not take into account singletons, because there are no links in such mentions. 2. All kinds of errors take the same level of punishment, although some cause more performance loss than the others. From [23], we use the following formulas to compute the precision and recall for each mention: Pm = |sm ∩ gm| |sm| , Rm = |sm ∩ gm| |gm| where sm and gm respectively are the response chain and key chain that contain mention m. The precision and recall for the whole document are then computed as follows: PB 3 = 1 |M| ∑ m∈M Pm, RB 3 = 1 |M| ∑ m∈M Rm CEAF metric This metric is proposed as another method to overcome the above shortcomings of the MUC metric, where the precision and recall are derived from the optimal alignment between the response chains S and the key chains G. According to [24], an alignment between S and G (|S | ≤ |G|) is defined as H = {(s, h(s)) | s ∈ S }, where h : S → G is injective (when |S | > |G|, the roles of S and G are reversed), which means: 1. ∀s ∈ S ,∀s′ ∈ S : s , s′ ⇔ h(s) , h(s′) 2. |H| = |S | The similarity score of H, denoted by Φ(H), is the sum of all the similarity scores between s and h(s) in H, denoted by φ(s, h(s)): Φ(H) = ∑ s∈S φ(s, h(s)) The goal of this metric is to calculate the optimal alignment H∗ in which Φ(H∗) is maximized. The result is then used to compute H.D. Nguyen, T.H. Cao / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 34, No. 2 (2018) 33–43 41 Table 5. Results of system using bag-of-words for Partial-String-Matching feature (Part-BOW) MUC B3 CEAF Average P R F P R F P R F P R F All 84.4 81.3 82.8 97.2 96.7 96.9 94.2 94.7 94.5 91.9 90.9 91.4 Problem 86.0 90.2 88.0 96.8 97.9 97.3 95.5 94.5 95.0 92.8 94.2 93.5 Test 71.3 88.9 79.1 99.2 99.7 99.5 99.2 98.8 99.0 89.9 95.8 92.7 Treatment 73.7 47.9 58.1 98.9 95.7 97.3 93.7 96.3 95.0 88.8 80.0 84.2 Table 6. Results of system using boolean for Partial-String-Matching feature (Part-Bool) MUC B3 CEAF Average P R F P R F P R F P R F All 63.5 47.5 54.4 95.1 90.5 92.8 86.0 90.2 88.0 81.5 76.1 78.7 Problem 81.9 48.9 61.3 96.6 89.5 92.9 84.3 90.8 87.4 87.6 76.4 81.6 Test 74.9 98.6 85.1 99.3 100 99.6 99.5 99.0 99.3 91.2 99.2 95.0 Treatment 34.9 41.2 37.8 94.1 94.6 94.3 91.0 90.5 90.7 73.3 75.4 74.4 the precision and recall: PCEAF = Φ(H∗)∑ s∈S φ(s, s) , RCEAF = Φ(H∗)∑ g∈G φ(g, g) There are four ways to compute the similarity score between two coreferential chains proposed by [24]. We use φ4 as recommended by i2b2: φ4(s, g) = 2|s ∩ g| |s| + |g| 4.4. Results and discussion In this section, we show the experimental results of our system and compare the two variants of the Partial-String-Matching features, where one is implemented using boolean values and the other using bag-of-words vectors (named Part-Bool and Part-BOW respectively). The Part-BOW system achieves 91.9% in precision, 90.9% in recall and 91.4% in F1 (see Table 5). Compared to Part-Bool (Table 6), the F1 score is improved by an amount of 12.7%, which shows the effectiveness of the bag-of-words model. These results prove that coreference in Vietnamese EMRs largely depends on lexical characteristics. Due to the syllabic nature of how Vietnamese words are constructed, a simple boolean value indicating whether two mentions have any similarity in their lexical strings is not sufficient. Knowing the exact tokens two mentions overlap at by the use of bag-of-words vectors, a classifier can be trained to distinguish most of the cases where these matching tokens do not suggest coreferential relationship. Regarding the results of each class, in our best system (Part-BOW), the Problem class has the highest F1 score of 93.5%, Test achieves 92.7%, and Treatments 84.2%, which is the lowest. This shows that bag-of-words highly improves coreference performance among Problem mentions (an increase of 11.9% from the boolean variant), where there are usually long phrases consisting of multiple words and syllables. As for the lowest F1 of the Treatment class, there are cases where hypernyms are used 42 H.D. Nguyen, T.H. Cao / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 34, No. 2 (2018) 33–43 to refer to the previously mentioned treatments. In English, when a hypernym is used for such purpose, it often comes after a definite article “the”, giving a hint that it actually refers to a previous mention. While there is no definite article in Vietnamese, there are words such as “này”, “đó” used for such purpose but they are not strictly enforced. For example, consider the text “Sau điều trị bệnh nhân khỏi, cho xuất viện.” from our dataset; the underlined mention “điều trị” means “general treatment”. When used in such a context, it implies one or many specific treatments previously mentioned in the document. In the case where it refers to two or more treatments, the coreference is of the type Set/Subset and is excluded from i2b2’s definition. In the other case where it refers to only one treatment, the coreference is of the type Identity and should be resolved. As can be seen in the example, there are no words such as “này” or “đó” used. This poses a problem to be solved in future works. 5. Conclusion In this paper, we propose a system to resolve coreference in Vietnamese electronic medical records. Our contributions are threefold. First, to the best of our knowledge, our work is the first to explore this NLP problem on Vietnamese EMRs. Second, we discover and define rules to annotate verbs in a Vietnamese clinical corpus as their use is preferred to describe symptoms. Finally, our work shows that lexical similarity plays an important role in determining coreferential relationship among mentions in Vietnamese EMRs. By using bag-of-words vectors to encode the matching tokens, our system achieves an F1 score of 91.4%. These could provide a basis for further NLP research on Vietnamese EMRs when clinical texts from hospitals in Vietnam are more available. Despite having a high performance, there remains some unsolved cases. These include but not limited to detecting synonyms, hypernyms, and extracting contextual clues to distinguish non-corefential mentions when their lexical strings are the same. We suggest them for future works. Acknowledgements This work is funded by Vietnam National University at Ho Chi Minh City under the grant of the research program on Electronic Medical Records (2015-2020). References [1] J. Hobbs, Readings in natural language processing, Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, 1986, Ch. Resolving Pronoun References, pp. 339–352. [2] S. Lappin, H. J. Leass, An algorithm for pronominal anaphora resolution, Comput. Linguist. 20 (4) (1994) 535–561. [3] D. Connolly, J. D. Burger, D. S. Day, A machine learning approach to anaphoric reference, in: Proceedings of International Conference on New Methods in Language Processing, 1994, pp. 255–261. [4] J. F. McCarthy, W. G. Lehnert, Using decision trees for conference resolution, in: Proceedings of the 14th International Joint Conference on Artificial Intelligence - Volume 2, IJCAI’95, Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, 1995, pp. 1050–1055. [5] X. Yangy, J. Su, G. Zhou, C. L. Tan, An np-cluster based approach to coreference resolution, in: Proceedings of the 20th International Conference on Computational Linguistics, COLING ’04, Association for Computational Linguistics, Stroudsburg, PA, USA, 2004. doi:10.3115/1220355.1220388. [6] X. Yang, G. Zhou, J. Su, C. L. Tan, Coreference resolution using competition learning approach, in: Proceedings of the 41st Annual Meeting on Association for Computational Linguistics - Volume 1, ACL ’03, Association for Computational Linguistics, Stroudsburg, PA, USA, 2003, pp. 176–183. doi:10.3115/1075096.1075119. [7] O. Uzuner, A. Bodnari, S. Shen, T. Forbush, J. Pestian, B. R. South, Evaluating the state of the art in coreference resolution for electronic H.D. Nguyen, T.H. Cao / VNU Journal of Science: Comp. Science & Com. Eng., Vol. 34, No. 2 (2018) 33–43 43 medical records, Journal of the American Medical Informatics Association 19 (5) (2012) 786–791. doi:10.1136/amiajnl-2011-000784. [8] O. Uzuner, Recognizing obesity and comorbidities in sparse data, Journal of the American Medical Informatics Association 16 (4) (2009) 561–570. doi:10.1197/jamia.m3115. [9] A. Stubbs, C. Kotfila, H. Xu, Özlem Uzuner, Identifying risk factors for heart disease over time: Overview of 2014 i2b2/UTHealth shared task track 2, Journal of Biomedical Informatics 58 (2015) S67–S77. doi:10.1016/j.jbi.2015.07.001. [10] S. Pradhan, L. Ramshaw, M. Marcus, M. Palmer, R. Weischedel, N. Xue, Conll-2011 shared task: Modeling unrestricted coreference in ontonotes, in: Proceedings of the Fifteenth Conference on Computational Natural Language Learning: Shared Task, CONLL Shared Task ’11, Association for Computational Linguistics, Stroudsburg, PA, USA, 2011, pp. 1–27. [11] W. M. Soon, H. T. Ng, D. C. Y. Lim, A machine learning approach to coreference resolution of noun phrases, Computational Linguistics 27 (4) (2001) 521–544. doi:10.1162/089120101753342653. [12] V. Ng, C. Cardie, Identifying anaphoric and non-anaphoric noun phrases to improve coreference resolution, in: Proceedings of the 19th international conference on Computational linguistics -, Association for Computational Linguistics, 2002. doi:10.3115/1072228.1072367. [13] V. Ng, C. Cardie, Improving machine learning approaches to coreference resolution, in: Proceedings of the 40th Annual Meeting on Association for Computational Linguistics, ACL ’02, Association for Computational Linguistics, 2002, pp. 104–111. doi:10.3115/1073083.1073102. [14] N. Bansal, A. Blum, S. Chawla, Correlation clustering, Machine Learning 56 (1-3) (2004) 89–113. doi:10.1023/b:mach.0000033116.57574.95. [15] C. Nicolae, G. Nicolae, Bestcut: A graph algorithm for coreference resolution, in: Proceedings of the 2006 Conference on Empirical Methods in Natural Language Processing, EMNLP ’06, Association for Computational Linguistics, Stroudsburg, PA, USA, 2006, pp. 275–283. [16] X. Yang, G. Zhou, J. Su, C. L. Tan, Improving noun phrase coreference resolution by matching strings, in: Natural Language Processing, IJCNLP ’04, Springer Berlin Heidelberg, 2005, pp. 22–31. doi:10.1007/978-3-540-30211-7_3. [17] B. Rink, K. Roberts, S. M. Harabagiu, A supervised framework for resolving coreference in clinical records, Journal of the American Medical Informatics Association 19 (5) (2012) 875–882. doi:10.1136/amiajnl-2012-000810. [18] H. Yang, A. Willis, A. de Roeck, B. Nuseibeh, A system for coreference resolution in clinical documents, in: Proceedings of the 2011 i2b2/VA/Cincinnati Workshop on Challenges in Natural Language Processing for Clinical Data, i2b2, 2011. [19] Y. Xu, J. Liu, J. Wu, Y. Wang, Z. Tu, J.-T. Sun, J. Tsujii, E. I.-C. Chang, A classification approach to coreference in discharge summaries: 2011 i2b2 challenge, Journal of the American Medical Informatics Association 19 (5) (2012) 897–905. doi:10.1136/amiajnl-2011-000734. [20] L. H. Phuong, N. T. M. Huyen, A. Roussanaly, H. T. Vinh, A hybrid approach to word segmentation of vietnamese texts, in: Language and Automata Theory and Applications, Springer Berlin Heidelberg, pp. 240–249. doi:10.1007/978-3-540-88282-4_23. [21] C.-C. Chang, C.-J. Lin, LIBSVM, ACM Transactions on Intelligent Systems and Technology 2 (3) (2011) 1–27. doi:10.1145/1961189.1961199. [22] M. Vilain, J. Burger, J. Aberdeen, D. Connolly, L. Hirschman, A model-theoretic coreference scoring scheme, in: Proceedings of the 6th conference on Message understanding, MUC6 ’95, Association for Computational Linguistics, 1995, pp. 45–52. doi:10.3115/1072399.1072405. [23] A. Bagga, B. Baldwin, Algorithms for scoring coreference chains, in: In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference, 1998, pp. 563–566. [24] X. Luo, On coreference resolution performance metrics, in: Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing, HLT ’05, Association for Computational Linguistics, 2005, pp. 25–32. doi:10.3115/1220575.1220579.

Các file đính kèm theo tài liệu này:

  • pdfcoreference_resolution_in_vietnamese_electronic_medical_reco.pdf
Tài liệu liên quan