• Aucun résultat trouvé

Similarity Clustering

Dans le document Data Mining in Biomedicine Using Ontologies (Page 194-200)

Text Summarization Using Ontologies

8.4 Data Summarization Through Background Knowledge

8.4.2 Similarity Clustering

While we may consider connectivity clustering as ontology-based in a genuine sense it is not the only possible direction. Alternatively, cluster applying, given similar-ity measures, over the set of concepts should also be considered. Obvi ously, if the measure is derived from an ontology, and thereby does reflect this, then so will the clustering. Below we will assume an ontology-based similarity mea sure sim, as briefly touched upon in Section 8.2.3 but, make no further assumptions of the type and characteristics of this measure.

We may expect a pattern similar to connectivity clustering in the derivation of summaries in an approach based on similarity, when the similarity measure closely reflects connectivity in the ontology, as the simple shortest path measure does.

Beside the example ontology from Figure 8.3 we will use the following set of clusters:

{C1= {size, number}

C2 = {government, state, committee}

C3= {defender, man, servant, woman}

C4 = {cost, bribe, price, fee}

C5= {fortress, fortifi cation, stockade}}

where C1,...,C4 are from SemCor and C5 is from the example ontology in Figure 8.2. The former is created by use of a categorical agglomerative hierarchial cluster-ing, with the shortest path similarity measure [12] over all documents in SemCor containing the word jury. We then picked from one of the levels in the clustering some meaningful clusters, and furthermore added a structure cluster (fortress, for-tifi cation, stockade) from Figure 8.2 to capture another aspect.

8.4.2.1 A Hierarchical Similarity-Based Approach

With a given path-length depen dent similarity measure derived from the ontology, a lub-centered, agglomerative, hierarchical clustering can be performed as follows.

Initially each cluster corresponds to an individual element of the set to be sum-marized. At each particular stage, the two clusters that are most similar are joined together. This is the principle of conventional hierarchical clustering; however rath-er than replacing the two joined clustrath-ers with their union, as in the conventional approach they are replaced by their lub. Thus, given a set of concepts C = {c1, ..., cn}, summarizers can be derived as follows.

ALGORITHM—Hierarchical clustering summary.

INPUT: Set of concepts C = {c1, ..., cn}.

OUTPUT: Generalizing description δ(C) for C.

Let the instantiated ontology for

1. C be OC= (LC , ≤C , R)

As was also the case with the connectivity clustering, to obtain an appropriate description of C we might have to apply δ several times and at some point m we have that δm(C) = Top.

Figure 8.5 illustrates the application of δ for a total of 15 times to the set of concepts from the previous example. At each step is shown the lub with which the two clusters is replaced. Thus, the first step is to replace cost and price with cost, the second to replace government and state with government, and so on.

8.4.2.2 Simple Least Upper Bound-Based Approach

A straightforward similarity based approach is simply to apply a crisp clustering to the set of concepts C = {c1, ..., cn} leading to {C1,...,Ck} and then provide the set of { ,cˆ1 ,cˆk} = {lub(C1), ..., lub(Ck)} for the division of C as summary5 to also take into account the importance of clusters in terms of their sizes the summary can be modified by the support of the generalizing concepts, support(x, C), that for a given concept specifies the fraction of elements from the set C covered:

( )

,

{

y y C y, x

}

support x C

C

∈ ≤

= (8.9)

leading to a fuzzyfied (weighted) summary, based on the division (crisp cluster ing) of C into {C1,...,Ck}:

( ( )

i ,

) ( )

i

isupport lub C C lub C

(8.10)

5. Observe that we put no restrictions on the clustering here, but, of course, the general idea is that the clustering applies a similarity measure that is ontology-based and that the ontology reflected is the instantiated ontology over the set of concepts C.

Figure 8.5 An illustration of the hierarchical clustering summary. The merging of two clusters is shown with their lub.

8.4 Data Summarization Through Background Knowledge 179

To illustrate this lub-based approach, consider the set of clusters and their least upper bounds in Table 8.1.

From these clusters the fuzzyfied summary {.13/magnitude + .19/organi zation + .25/person + .25/cost + .19/defensive structure} can be generated. Obviously, this approach to summarization will not be very tolerant as regards noise in the clusters given. Consider the following example, in which bribe is replaced by politics and stockade by radiator. The least upper bounds of the respective clusters becomes more general as illustrated in Table 8.2.

As a result, the summary becomes {.13/magnitude + .19/organization + .25/per-son + .25/relation + .19/artifact}. To get around this problem, we can introduce a soft definition of lub and combine this again with crisp clusters to get more specific cluster-based summaries. Obviously, this approach to summariza tion will not be very tolerant as regards noise in the clusters given. The lub for a cluster including an outlier may well be the top element or something similarly useless.

8.4.2.3 A Soft Least Upper Bound Approach

To get around this problem we can introduce a soft definition of lub and combine this again with crisp clusters to get more specific cluster-based summaries.

A soft definition of lub for a (sub)set of concepts C′ should comprise upper boundness as well as leastness (or least upperness), expressing, respectively the portion of concepts in C′ that are generalized and the degree to which a concept is least upper with regard to one or more of the concepts in C′.

Upper boundness can be expressed for a set of concepts C′ by μub(C′) simply as the support as regards C′:

Table 8.1 A Set of Crisp Clusters and Their Least Upper Bounds from WordNet

Cluster Lub

{number, size} Magnitude

{committee, government, state} Organization {defender, man, servant, woman} Person {bribe, cost, fee, price} Cost

{fortifi cation, fortress, stockade} Defensive structure

Table 8.2 A Set of Crisp Clusters with Noise and Their Least Upper Bounds from WordNet

Cluster Lub

{number, size} Magnitude

{committee, government, state} Organization {defender, man, servant, woman} Person {cost, fee, politics, price} Relation {fortifi cation, radiator, stockade} Artifact

( )

( ) (

,

)

ub C x support x C

μ = ′ (8.11)

covering all generalizations of one or more concepts in C′ and including all concepts that generalize all of C′ (including the topmost concept Top) as full members.

Leastness can be defined on top of a function that expresses how close a con-cept is to a set of concon-cepts C′ such as dist(C′ ,y) = minx∈C′ dist(x, y), where dist(x, y) expresses the shortest path upwards6 from x to y, as follows:

( )

( ) ( )

re-strictive version of leastness and with the other extreme λ = 0 corresponding to no restriction at all (all upper concepts become full members).

A simple soft least upper bound fl ub can now be defined as the product be-tween μlu and μub

( )

( )

( )

( )

( )

( )

lub , , *

f C λ x lu C λ x ub C x

μ = μ μ (8.13)

Notice that a lub for C is not necessary a best candidate among the fl ub el-ements. Thus, again with a division (crisp clustering) of C into {C1,... ,Ck}, the basis for the summary here is the set of fuzzy sets {fl ub(C1), ..., fl ub(Ck)}leading to the

As in the lub-based case the summarizers should, in addition, be weighted by support. Thus, we can weight all elements in each {fl ub(C1), ..., fl ub(Ck)} with the

where ⊗ is a t-norm, probably with product as an appropriate choice. Finally this set should obviously be restricted by some appropriate threshold α:

{ }

(

C1, ,Ck

) {

m x m x

( {

C1, ,Ck

} )

m

}

δα  = ∈δ  ∧ >α (8.16)

6. Upward refers to the idea that only paths consisting soely of edges in the direction of should be taken into account, and it should be strictly emphasized that the graph considered correspoinds to the transitively reduced ontology.

8.5 Conclusion 181

Given the previous example of noisy crisp clustering the use of a fl ub-based summary7 will lead to

{.19/cost + .15/outgo + .15/relation+

.15/person + .13/organization + .13/government+

.11/fi nacialloss + .11/artifact+

.11/defensivestruture + ...}

while the lub based summary was {.13/magnitude + .19/organization+

.25/person + .25/relation + .19/artifact}

In the fl ub based summary, cost has a high degree of membership, due to the fact that it is a very good description of three of the four elements in the cluster {cost, price, fee, politics}. Thus, the introduction of the fl ub reduces the effect of noise caused by the noisy element politics. Also, the degree of membership of ar-tifact is comparable to the degree of membership of defensive structure, which is the immediate generalization of fortress and stockade. Again, the result of using a fl ub based summary is that the effect of the noisy element radiator in the cluster {fortress, stockade, radiator} is reduced.

8.5 Conclusion

In this chapter we have considered how to use ontologies to provide data summa-ries, with a special focus on textual data. Such summaries can be used in a querying approach where concepts describing documents, rather than documents directly, are retrieved as query answers. The summaries presented are conceptual, due to fact that they exploit concepts from the text to be summarized, and ontology-based because these concepts are drawn from a reference ontology.

The principles for conceptual summarization are presented here, as related to so-called instantiated ontologies—a conceptual structure reflecting the content of a given document collection, and therefore, particularly well suited as a target for conceptual querying; however, the summaries introduced are not dependent on this notion.

We have discussed some possible directions and basic principles for sum-marizations. It is obvious that development of more specific turnkey methods for summarization should be guided by profound experiments within a frame work that includes a realistic general world knowledge resource. Initial studies have been done with a WordNet-based general ontology and SEMCOR as corpus (Informa-tion base); and preliminary experiments have been done with an UMLS-based gen-eral ontology and with a selection of abstracts from MedLine [14] as corpus; how-ever, much more thorough experimental work needs to be done.

In addition, summary evaluation principles should be taken into account, first of all, as complement in guiding the development of summary principles. However, specific characteristics encircling good summaries may also be used in solutions to

7. With λ being 1.

the stopping problem—on deciding when, in the process of incremental contrac-tion of (potential summary) sets, the best candidate has been found.

Initial considerations on the quality of summaries can be found in [19], but the issue is also an obvious direction for further work in continuation of what has been described here.

References

[1] Jensen, P. A. and J. F. Nilsson, “Ontology-Based Semantics for Prepositions, in Syntax and Semantics of Prepositions,” P. Paint-Dizier (ed.), Text, Speech and Language Technology, Vol. 29, Springer, 2006.

[2] Aronson, S.R., “Effective mapping of biomedical text to the UMLS Metathesaurus: the MetaMap Program,” Proc AMIA Symp, 2001, pp. 17–21.

[3] Abney, S., “Partial Parsing via Finite-State Cascades,” Proc of the ESS LLI ’96 Robust Pars-ing Workshop, 1996.

[4] Bodenreider, O., “The Unified Medical Language System (UMLS): Integrating Biomedical Terminology,” Nucleic Acids Research, Vol. 32, 2004, pp. D267-D270.

[5] Hahn, U., and I. Mani, “The Challenges of Automatic Summarization,” Computer, No-vember 2000.

[6] Melli, G., et. al., “Description of SQUASH, the SFU Question Answering Summary Han-dler for the DUC-2005 Summarization Task,” Proc of DUC-2005, Van couver, Canada, 2005, pp. 103–110.

[7] Shi, Z., et al., “Question Answering Summarization of Multiple Biomedical Docu ments,”

Lecture Notes in Computer Science, Springer Berlin/Heidelberg Vol. 4509/2007, Advances in Artificial Intelligence, 2007, pp. 284-295.

[8] Andreasen, T., H. Bulskov, “Conceptual Querying Through Ontologies,” Fuzzy Sets and Systems, Vol. 160, No. 51.

[9] Nilsson, J.F., “A Logico-Algebraic Framework for Ontologies—ONTOLOG,” in Jensen, P.A., Skadhauge, P., (eds.), First Int. OntoQuery Workshop, University of Southern Den-mark, 2001.

[10] Miller, G.A., et. al., “Using a Semantic Concordance for Sense Identification,” Proc. of the ARPA Human Language Technology Workshop, 1994, pp. 240–243.

[11] Miller, G.A., “Wordnet: A Lexical Database for English,” Commun. ACM, Vol. 38, No.

11, 1995, pp. 39–41.

[12] Rada, R., et. al., “Development and Application of a Metric on Semantic Nets,” IEEE Trans. on Systems, Man, and Cybernetics, Vol. 19, No. 1, January 1989, pp. 17–30.

[13] Unified Medical Language System, U.S. National Library of Medicine, http://www.nlm.nih.

gov/research/umls/

[14] Medical Literature Analysis and Retrieval System Online, U.S. National Library of Medi-cine, http://www.ncbi.nlm.nih.gov/pubmed/.

[15] Bulskov, H., R. Knappe, and T. Andreasen, “On Measuring Similarity for Conceptual Querying,” FQAS ’02: Proc. of the 5th Inter. Conf. on Flex ible Query Answering Systems, Springer-Verlag, 2002, pp. 100–111.

[16] Resnik, P., “Semantic Similarity in a Taxonomy: An Information-Based Measure and Its Application to Problems of Ambiguity in Natural Language,” 1999.

[17] Andreasen, T., R. Knappe, H. Bulskov, “Domain-Specific Similarity and Retrieval,” Proc.

IFSA 2005, Tsinghua University Press, 2005, pp. 496–502.

[18] Andreasen, T., et. al., “Content-Based Text Querying with Ontological Descriptors,” Data Knowledge Engineering, Elsevier Science Publishers, Vol. 48, No. 2, pp. 199–219.

8.5 Conclusion 183

[19] Yager, R. R., F. E. Petry, “A Multicriteria Approach to Data Summarization Using Concept Hierarchies,” IEEE Trans. on Fuzzy Sys., Vol. 14, No. 6, 2006.

[20] Zhou, X., H. Han, “Survey of Word Sense Disambiguation Approaches,” 18th FLAIRS Conference, 2005.

[21] Francis, W. N. and H. Kucera, Brown Corpus Manual, Department of Linguistics, Brown University, 1964.

Dans le document Data Mining in Biomedicine Using Ontologies (Page 194-200)