Sørensen–Dice coefficient
The Sørensen–Dice index, also known by other names (see Names, below), is a statistic used for comparing the similarity of two samples. It was independently developed by the botanists Thorvald Sørensen[1] and Lee Raymond Dice,[2] who published in 1948 and 1945 respectively.
Contents
Name
The index is known by several other names, usually Sørensen index or Dice's coefficient. Both names also see "similarity coefficient", "index", and other such variations. Common alternate spellings for Sørensen are Sorenson, Soerenson index and Sörenson index, and all three can also be seen with the –sen ending.
Other names include:
- Czekanowski's binary (non-quantitative) index[3]
Quantitative version
The expression is easily extended to abundance instead of presence/absence of species. This quantitative version is known by several names:
- Quantitative Sørensen–Dice index[3]
- Quantitative Sørensen index[3]
- Quantitative Dice index[3]
- Bray-Curtis similarity (1 minus the Bray-Curtis dissimilarity)[3]
- Czekanowski's quantitative index[3]
- Steinhaus index[3]
- Pielou's percentage similarity[3]
- 1 minus the Hellinger distance[4]
Formula
Sørensen's original formula was intended to be applied to presence/absence data, and is
where A and B are the number of species in samples A and B, respectively, and C is the number of species shared by the two samples; QS is the quotient of similarity and ranges between 0 and 1.[5]
It can be viewed as a similarity measure over sets:
Similarly to Jaccard, the set operations can be expressed in terms of vector operations over binary vectors A and B:
which gives the same outcome over binary vectors and also gives a more general similarity metric over vectors in general terms.
For sets X and Y of keywords used in information retrieval, the coefficient may be defined as twice the shared information (intersection) over the sum of cardinalities :[6]
When taken as a string similarity measure, the coefficient may be calculated for two strings, x and y using bigrams as follows:[7]
where nt is the number of character bigrams found in both strings, nx is the number of bigrams in string x and ny is the number of bigrams in string y. For example, to calculate the similarity between:
night
nacht
We would find the set of bigrams in each word:
- {
ni
,ig
,gh
,ht
} - {
na
,ac
,ch
,ht
}
Each set has four elements, and the intersection of these two sets has only one element: ht
.
Inserting these numbers into the formula, we calculate, s = (2 · 1) / (4 + 4) = 0.25.
Difference from Jaccard
This coefficient is not very different in form from the Jaccard index. However, since it doesn't satisfy the triangle inequality, it can be considered a semimetric version of the Jaccard index.[3]
The function ranges between zero and one, like Jaccard. Unlike Jaccard, the corresponding difference function
is not a proper distance metric as it does not possess the property of triangle inequality.[3] The simplest counterexample of this is given by the three sets {a}, {b}, and {a,b}, the distance between the first two being 1, and the difference between the third and each of the others being one-third. To satisfy the triangle inequality, the sum of any two of these three sides must be greater than or equal to the remaining side. However, the distance between {a} and {a,b} plus the distance between {b} and {a,b} equals 2/3 and is therefore less than the distance between {a} and {b} which is 1.
Applications
The Sørensen–Dice coefficient is mainly useful for ecological community data (e.g. Looman & Campbell, 1960[8]). Justification for its use is primarily empirical rather than theoretical (although it can be justified theoretically as the intersection of two fuzzy sets[9]). As compared to Euclidean distance, Sørensen distance retains sensitivity in more heterogeneous data sets and gives less weight to outliers.[10] Recently the Dice score (and its variations, e.g. logDice taking a logarithm of it) has become popular in computer lexicography for measuring the lexical association score of two given words.[11]
See also
- Correlation
- Czekanowski similarity index
- Jaccard index
- Hamming distance
- Horn’s index
- Hurlbert’s index
- Kulczyński similarity index
- Pianka's index
- MacArthur and Levin's index
- Mantel test
- Morisita's overlap index
- Most frequent k characters
- Overlap coefficient
- Renkonen similarity index (due to Olavi Renkonen)
- Simplified Morisita’s index
- Tversky index
- Universal adaptive strategy theory (UAST)
References
<templatestyles src="Reflist/styles.css" />
Cite error: Invalid <references>
tag; parameter "group" is allowed only.
<references />
, or <references group="..." />
External links
![]() |
The Wikibook Algorithm implementation has a page on the topic of: Dice's coefficient |
- Open Source Dice / Sorensen Scala implementation as part of the larger stringmetric project
- ↑ Lua error in package.lua at line 80: module 'strict' not found.
- ↑ Lua error in package.lua at line 80: module 'strict' not found.
- ↑ 3.0 3.1 3.2 3.3 3.4 3.5 3.6 3.7 3.8 3.9 Gallagher, E.D., 1999. COMPAH Documentation, University of Massachusetts, Boston
- ↑ Lua error in package.lua at line 80: module 'strict' not found.
- ↑ http://www.sekj.org/PDF/anbf40/anbf40-415.pdf
- ↑ Lua error in package.lua at line 80: module 'strict' not found.
- ↑ Lua error in package.lua at line 80: module 'strict' not found.
- ↑ Looman, J. and Campbell, J.B. (1960) Adaptation of Sorensen's K (1948) for estimating unit affinities in prairie vegetation. Ecology 41 (3): 409–416.
- ↑ Roberts, D.W. (1986) Ordination on the basis of fuzzy set theory. Vegetatio 66 (3): 123–131.
- ↑ McCune, Bruce & Grace, James (2002) Analysis of Ecological Communities. Mjm Software Design; ISBN 0-9721290-0-6.
- ↑ Rychlý, P. (2008) A lexicographer-friendly association score. Proceedings of the Second Workshop on Recent Advances in Slavonic Natural Language Processing RASLAN 2008: 6–9