- Research article
- Open Access
Large scale study of multiple-molecule queries
© Nasr et al; licensee BioMed Central Ltd. 2009
- Received: 01 June 2009
- Accepted: 04 June 2009
- Published: 04 June 2009
In ligand-based screening, as well as in other chemoinformatics applications, one seeks to effectively search large repositories of molecules in order to retrieve molecules that are similar typically to a single molecule lead. However, in some case, multiple molecules from the same family are available to seed the query and search for other members of the same family.
Multiple-molecule query methods have been less studied than single-molecule query methods. Furthermore, the previous studies have relied on proprietary data and sometimes have not used proper cross-validation methods to assess the results. In contrast, here we develop and compare multiple-molecule query methods using several large publicly available data sets and background. We also create a framework based on a strict cross-validation protocol to allow unbiased benchmarking for direct comparison in future studies across several performance metrics.
Fourteen different multiple-molecule query methods were defined and benchmarked using: (1) 41 publicly available data sets of related molecules with similar biological activity; and (2) publicly available background data sets consisting of up to 175,000 molecules randomly extracted from the ChemDB database and other sources. Eight of the fourteen methods were parameter free, and six of them fit one or two free parameters to the data using a careful cross-validation protocol. All the methods were assessed and compared for their ability to retrieve members of the same family against the background data set by using several performance metrics including the Area Under the Accumulation Curve (AUAC), Area Under the Curve (AUC), F1-measure, and BEDROC metrics.
Consistent with the previous literature, the best parameter-free methods are the MAX-SIM and MIN-RANK methods, which score a molecule to a family by the maximum similarity, or minimum ranking, obtained across the family. One new parameterized method introduced in this study and two previously defined methods, the Exponential Tanimoto Discriminant (ETD), the Tanimoto Power Discriminant (TPD), and the Binary Kernel Discriminant (BKD), outperform most other methods but are more complex, requiring one or two parameters to be fit to the data.
Fourteen methods for multiple-molecule querying of chemical databases, including novel methods, (ETD) and (TPD), are validated using publicly available data sets, standard cross-validation protocols, and established metrics. The best results are obtained with ETD, TPD, BKD, MAX-SIM, and MIN-RANK. These results can be replicated and compared with the results of future studies using data freely downloadable from http://cdb.ics.uci.edu/.
- Molecular Fingerprint
- Label Path
- Inactive Molecule
- Tanimoto Similarity
- Query Molecule
The rapid search of large repositories of molecules is a fundamental task of chemoinformatics. In a typical search, the molecules in a repository are ranked by their similarity to a single molecule query. If an appropriate similarity metric is employed, the molecules most similar to the query are most likely to exhibit physical, chemical, or biological properties similar to the query molecule. There is extensive body of literature on how to search based on this type of single-molecule query ([1–5] and references therein).
Multiple-molecule searches [4–8] have been less studied than single-molecules searches. Furthermore, the published studies of multiple-molecule searches suffer from critical weaknesses: reliance on proprietary data, poorly characterized background databases, and non-standardized performance metrics. Furthermore, some of these studies do not appropriately cross-validate their results.
Our study directly rectifies these deficiencies by using public data and better evaluation methodology. We both introduce novel multiple-molecule query methods and compare these methods with previously described multiple-molecule query methods using public activity data sets, a public background database, a strict cross-validation protocol, and reproducible evaluation methodology.
In what follows, many of the disparate methods from the literature are organized into a common framework.
We begin by describing molecular representations and different metrics used to quantify the similarity between pairs of molecules. Multiple-molecule methods are often built on the pairwise-similarities between molecules.
There are many ways of measuring similarities between chemicals ([9–13] and references therein), but we focus on similarities measured between molecular fingerprints because they are the most commonly used in chemical search engines and the most studied.
Molecular fingerprint notation
Molecular fingerprints are binary vectors where each component is a bit associated with the presence or absence of a particular feature (for example a functional group, path, tree, or some other graphical substructure) of labeled atoms and bonds present in the molecule. These vectors are typically long and sparse, on the order of 100, 000 or more bits long and with only hundreds of 1-bits for a given molecule.
Let denote a molecule, and = (A i ) the corresponding fingerprint of length N, and A, the number of 1-bits in . Likewise, ℬ denotes another molecule, = (B i ) its corresponding fingerprint, and B the number of 1-bits in . Then A ∪ B (resp. A ∩ B) denotes the total number of 1-bits in OR (resp. AND ). We can also define A ⊕ B as the number of bits that are exclusively in either or but not both (the XOR), noting that this is exactly equivalent to (A ∪ B) - (A ∩ B).
For efficiency of space and time, fingerprints are often 'folded' to a much shorter length (N = 512 or N = 1024) using a lossy-compression algorithm. Each bit in the short, compressed fingerprint is indexed by i and each bit in the long, uncompressed fingerprint is indexed by j. The compression algorithm works by setting a 1-bit for each bit i in the short fingerprint if and only if there is at least a single 1-bit in one of the positions in the uncompressed, longer fingerprint which satisfies j mod N = i. This is the algorithm used by many chemoinformatic systems, including the Daylight system .
One can then compute similarity between either the longer, uncompressed fingerprints or the shorter, compressed fingerprints. Most similarity metrics are functions of A, B, (A ∩ B), (A ∪ B), and N, which can be directly computed from any pair of compressed or uncompressed fingerprints and then used to compute a similarity.
In choosing between compressed and uncompressed fingerprints, there is a tradeoff between efficiency and accuracy. The compressed fingerprints are faster and smaller but introduce systematic bias and random noise into any similarity computed from them. The folding compression algorithm artificially biases the similarities measured between folded-fingerprints to be higher than similarities computed between the corresponding unfolded-fingerprints. We have shown the bias introduced into some similarites by the folding-compression algorithm is both systematic and mathematically correctable ; using only information from the shorter, folded fingerprints, one can estimate the similarity between the unfolded fingerprints. This method improves on the tradeoff between efficiency and accuracy, yielding a system with high speed, low space, and high accuracy.
List of method abbreviations.
The multiple-molecule methods we will describe are generated by mathematically combining pairwise-molecular metrics. So, we must first define pairwise-molecular metrics before we define multiple molecule methods. We use the term 'metric' to include both similarities (denoted with an S) and dissimilarities (denoted with a D). Both can be used for searching by either selecting the most similar or the least dissimilar molecules from a query.
The Tanimoto similarity is identical to the Tversky similarity when α = β = 1.
is equivalent to N minus the Simple Matching similarity. In the case of binary vectors, this is also known as the Manhattan or Hamming distance.
is equivalent to one minus the Tanimoto similarity.
These two dissimilarities are also mathematical distances, and each is associated with a specific similarity. The association between a similarity and a dissimilarity is important in some of the following definitions.
Additional metrics can be generated by considering monotone transformations of previously defined formulas. In the case of single-molecule queries, the rankings produced by monotone transformations of the same metric are exactly equivalent. However, in the case of multiple-molecule methods, where pairwise-metrics are mathematically combined to generate a composite score, monotone transformations applied to the pairwise-metric can substantially affect the final rankings.
This equation is parameterized by the exponent, α, and is a monotonic transformation so long as α ≠ 0.
where S( , ) is the similarity between A and ℬ, D( , ) = C S - S( , ) is the associated dissimilarity between and ℬ, and C S is a constant fixed at the maximum possible value of the similarity S. The bandwidth parameter, λ, and the shape parameter, k, can be adjusted to fine tune this transformation to particular problems. Algebraically, this transformation is always monotone.
We refer to this metric (Equation 11) as the Exponential Hamming (EH) metric, to emphasize it can be viewed as the Hamming distance (Equation 6) transformed by Equation 10.
This variation is termed the Exponential Tanimoto (ET) metric.
All three of these similarities, EH, ET, and TP are monotonic transformations of their underlying metrics (Tanimoto or Simple Matching). This means that the rankings for single molecule queries produced by ET and TP are exactly equivalent to those produced by Tanimoto similarity or distance. Likewise, the rankings produced by EH for single molecule queries are equivalent to the rankings produced by the Simple Matching similarity or the Hamming distance.
The multiple-molecule methods we describe are based on multiple-molecule similarities defined as between the query molecules and each molecule ℬ in the database. These similarities are the foundation of multiple-molecule search methods; they are used to rank the database and find the most relevant chemicals to a given query.
We can generate different multiple-molecule similarities by three general strategies: (1) aggregating the separately ranked results from single-molecule queries, (2) aggregating the individual pairwise similarity measures, and finally (3) aggregating the fingerprints into profiles and computing similarities between these profiles and single molecules.
In the first class of strategies, the idea is to aggregate the ranked results of single-molecule queries based on a pairwise-similarity, using each separately. So rather than mathematically combining individual-pairwise similarities, we combine the ranked lists that these similarities generate. We must extend our notation to describe this class of approaches. While denotes the similarity of molecule and ℬ, we use to denote the rank of molecule ℬ in the database ordered by each molecule's pairwise similarity to .
In the top hits, molecules with the exact same similarity to the query occur frequently enough to require clear specification of 's treatment of molecules with tied score. In our formulation, ties are assigned cautious ranks that favor the inactive molecules. In other words, if active and inactive molecules are scored identically, the inactive molecules are assigned better ranks. Other ways of handling ties, which include the inverse of the former as well as assigning equal average ranks, yield very close results that are not significantly different.
Using a threshold of -K on the final formula (minimum of the ranks) is equivalent to pooling the top K hits, and removing redundant hits, from each single-molecule query.
In some applications, in addition to scoring the database with information from active compounds, we can leverage information about compounds known to be inactive. Often, we can accurately infer that most database molecules are inactive and randomly select tens or hundreds from the database to use as a set of inactive molecules.
This can be naturally varied by substituting different similarity metrics and aggregations, as has been done by at least one study . The most important variation can be derived by substituting the corrected Exponential-Tanimoto (Equation 12) for the Exponential-Hamming, we call this method the Exponential Tanimoto Discriminant (ETD) or by substituting the Tanimoto-power (Equation 9) for the Exponential-Hamming, we call this method the Tanimoto Power Discriminant (TPD).
Within this class of approaches, we can use a fingerprint profile to summarize the family in a more detailed manner than a consensus fingerprint. A fingerprint profile summarizes the information in a set of fingerprints, very much like a sequence profile or a position specific scoring matrix (PSSM) summarizes the information in a set of aligned sequences in bioinformatics. A fingerprint profile stores the relative frequencies that each bit is set to one in a set of fingerprints. If a given bit position is set to one in half the fingerprints, the corresponding component in the profile is set to 0.5. In addition, we can also assign different weights to each molecule or fingerprint in the family. In this most general case, . For proper scaling, it is desirable to use a convex combination with .
where i iterates over the fingerprints in the set of query molecules, and j iterates over the components of each fingerprint.
In the following sections, for brevity and clarity, abbreviations are used to references each of the different of multiple molecule query methods. These names are tabulated in Table 1.
Some of these query methods presume the choice of a pairwise-similarity from which to derive the multiple-molecule similarity. Reported results use the Tanimoto similarity as this underlying pairewise-metric where possible, because it is the most commonly used and best performing pairwise-similarity. The exceptions to this decision are the BKD and SUM-EH methods, where the Hamming and Simple Matching metrics are used in order to replicate results from the literature.
Although not reported in this study, all experiments were replicated using three other pairwise-metrics known to perform well on chemical fingerprints: the Overlap, Mismatch/Euclidian, and Cosine metrics. Very rarely, the Cosine metric slightly outperforms the Tanimoto metric, but in all other cases the Tanimoto metric always outperforms other metrics. Most importantly, the particular single-molecule measure used did not affect the performance patterns of the multiple-molecule query methods. If a particular multiple-molecule method worked best for Tanimoto, it would also work best when other metrics were used.
Data set characteristics.
The multiple-molecule queries are benchmarked on a total of forty one data sets, divided into two groups. The first group of benchmarks use a background of negative examples randomly selected from the ChemDB. The second group of benchmarks uses data sets carefully derived from high throughput screening experiments specifically to more accurately assess query performance. This second group is expected to be more difficult than the first group.
First, we used twenty-four active data sets against a background database of molecules selected from the ChemDB. The data sets include thirteen WOMBAT data sets , six data sets from Stahl et al. , four data sets from Sutherland et al.  and the compounds with confirmed activity against HIV from the National Cancer Institute's (NCI) high throughput screen . The background consists of 175, 000 randomly selected molecules. Based on the number of molecules in the active set, | |, Trunchon et. al.  derived a minimum size for the background database which ensures that there will be enough information for metrics to provide meaningful discrimination between compared methods. A background database size of 175, 000 corresponds to a maximum active set size of 860 actives: more than large enough to effectively evaluate methods in our largest data set, suth-dhfr with 722 chemicals. This background database is available for download at http://cdb.ics.uci.edu.
Second, we also used seventeen Maximum Unbiased Validation (MUV) data sets provided by Rohrer et al.  against their corresponding background databases. Each of these data sets include both 30 maximally dissimilar active compounds and a background set of 15,000 decoys similar to the actives in regards to low-dimension properties like solubility, volume, and surface area. These data sets were constructed specifically to avoid artificially high screening performance caused by using an inappropriate decoy data set. By selecting decoy data sets in this manner, it is assumed that the decoy chemicals are inactive even while there remains a small chance that they are active. In the case of the MUV data sets, the background decoys are selected from chemicals which have screened negatively and have, therefore, preliminary experimental evidence of inactivity. In contrast, by selecting the ChemDB background from untested chemicals, a stronger assumption that none of the untested chemicals are actives. There is a small chance that, for a small number of molecules, this assumption is not valid, and they are therefore incorrectly labeled as negative examples. This chance is sufficiently rare that it would not affect the performance metrics by an appreciable amount.
All of the forty-one data sets represent groups of diverse molecules with similar activity. In some cases, the molecules of each group are known to interact with the same protein. In other cases, they are known to elicit the same biological effect. Data sets include active molecules against HIV, different steroid receptors activists and antagonists, enzymes, and steroid families. See Table 2 for a full listing of the data sets.
For each data set we generate fingerprints with the same protocol using an in-house program written in Python. Fingerprints are associated with labelled paths of length up to 8 (i.e. 9 atoms and 8 bonds). In this case, the total number of observed labelled paths is about 150, 000. Compression is done using the lossy fingerprint folding algorithm. Results are reported for fingerprints of length N = 1024.
Different multiple molecule queries were assessed using a standardized protocol to calculate unbiased, quantitative measures of performance.
The fundamental, user-level goal of a multiple-molecule search is to identify additional molecules with the same biological activity as the query molecules. The performance at this specific task can be directly assessed. Using a standard leave-one-out (LOO) cross-validation protocol, we assign each molecule in the data set a LOO score, obtained by leaving it out from the classifier. So, for each molecule, its LOO score is defined as the score computed by ignoring any information known about just this molecule's class.
where T P, F P, and F N are the number of true positives, false positives, and false negatives at a particular threshold. For this study, the threshold measure which optimized F1 was chosen. The definitions of the other metrics can be found in the references.
All of the performance metrics require a fully ordered ranking to be computed, and cannot, therefore, handle ties appropriately. Of note, some methods output the same score for several chemicals, yielding ranked lists with ties. In these situations, the performance is computed by averaging twenty ranked lists sampled so as to randomly resolve ties differently each time. The performance across these samples is averaged and reported.
In total, fourteen similarity methods are evaluated with four performance metrics on 41 data sets resulting in 2296 measurements. For brevity, only summarized BEDROC, F1 and AUC performances are discussed. The complete results, including the AUAC performance, are included as additional files. [See Additional files 1, 2, 3].
In addition to reporting the performance of each method, the Receiver Operating Characteristic (ROC) curves and pROC curves  are used to display performance levels graphically. We use these established information-retrieval measures, using repeatable methodology, so our results can be directly compared to future studies.
Mean performance of similarity methods.
0.981265 ± 0.004540
0.480749 ± 0.019308
0.915781 ± 0.012475
0.633951 ± 0.038076
0.030289 ± 0.009587
0.209204 ± 0.049180
0.840620 ± 0.032520
0.128860 ± 0.032596
0.490227 ± 0.066911
0.973312 ± 0.005642
0.484504 ± 0.022397
0.893180 ± 0.015592
0.717104 ± 0.034874
0.053943 ± 0.015827
0.284041 ± 0.058423
0.914782 ± 0.018230
0.341373 ± 0.032341
0.719190 ± 0.041269
0.907632 ± 0.019609
0.327810 ± 0.037178
0.696596 ± 0.044681
0.910909 ± 0.017386
0.149837 ± 0.033785
0.581197 ± 0.050633
0.980763 ± 0.004517
0.501197 ± 0.019569
0.890840 ± 0.017745
0.987087 ± 0.002653
0.508081 ± 0.020886
0.922371 ± 0.011330
0.986616 ± 0.002649
0.451587 ± 0.025795
0.906017 ± 0.014098
0.935054 ± 0.010774
0.296279 ± 0.032160
0.699456 ± 0.037583
0.974831 ± 0.005798
0.491106 ± 0.022314
0.897401 ± 0.015760
0.974963 ± 0.005751
0.490653 ± 0.022311
0.897621 ± 0.015771
Mean performance of similarity methods across MUV data sets.
0.731133 ± 0.030578
0.149965 ± 0.023025
0.345171 ± 0.042642
0.509469 ± 0.020590
0.017739 ± 0.004382
0.061569 ± 0.010419
0.598784 ± 0.030562
0.021604 ± 0.005490
0.104799 ± 0.022261
0.714848 ± 0.028352
0.156955 ± 0.025644
0.312150 ± 0.041033
0.533202 ± 0.025204
0.020921 ± 0.004781
0.070374 ± 0.008572
0.617073 ± 0.034809
0.052993 ± 0.021674
0.153437 ± 0.040308
0.644467 ± 0.032684
0.061232 ± 0.022654
0.177026 ± 0.040264
0.642907 ± 0.031377
0.041962 ± 0.011625
0.176723 ± 0.039162
0.784118 ± 0.025509
0.145667 ± 0.027758
0.354250 ± 0.044227
0.785944 ± 0.025833
0.141997 ± 0.026715
0.356733 ± 0.043394
0.775774 ± 0.025289
0.152530 ± 0.023448
0.352975 ± 0.041960
0.679485 ± 0.030601
0.100943 ± 0.025459
0.230528 ± 0.046206
0.733893 ± 0.027988
0.155680 ± 0.026528
0.324622 ± 0.042935
0.729849 ± 0.028113
0.157239 ± 0.026507
0.323068 ± 0.042618
Mean difference in performance of similarity methods.
0.005821 ± 0.002344
0.027332 ± 0.006678
0.006590 ± 0.003236
0.353136 ± 0.036644
0.477792 ± 0.020875
0.713166 ± 0.043907
0.146467 ± 0.030849
0.379221 ± 0.028884
0.432144 ± 0.059736
0.013775 ± 0.003534
0.023577 ± 0.005890
0.029191 ± 0.007341
0.269982 ± 0.033128
0.454138 ± 0.023192
0.638330 ± 0.052591
0.072305 ± 0.016279
0.166707 ± 0.025321
0.203181 ± 0.032996
0.079455 ± 0.017748
0.180270 ± 0.031082
0.225774 ± 0.036813
0.076178 ± 0.016079
0.358244 ± 0.032005
0.341173 ± 0.045043
0.006324 ± 0.002360
0.006884 ± 0.008400
0.031531 ± 0.008401
0.000471 ± 0.000443
0.056493 ± 0.010274
0.016353 ± 0.004338
0.052032 ± 0.009065
0.211802 ± 0.026177
0.222915 ± 0.031105
0.012256 ± 0.003648
0.016975 ± 0.006151
0.024970 ± 0.007082
0.012124 ± 0.003584
0.017428 ± 0.004657
0.024749 ± 0.007079
Mean difference in performance of similarity methods across MUV data sets.
0.054811 ± 0.007343
0.007274 ± 0.009481
0.011562 ± 0.013673
0.276475 ± 0.021613
0.139500 ± 0.023056
0.295164 ± 0.036878
0.187161 ± 0.021317
0.135635 ± 0.020176
0.251934 ± 0.029684
0.071096 ± 0.008692
0.000284 ± 0.005657
0.044583 ± 0.013589
0.252742 ± 0.020037
0.136318 ± 0.021634
0.286359 ± 0.031277
0.168871 ± 0.018754
0.104246 ± 0.014075
0.203296 ± 0.017032
0.141477 ± 0.019134
0.096007 ± 0.014050
0.179707 ± 0.018725
0.143037 ± 0.016147
0.115277 ± 0.014736
0.180010 ± 0.012881
0.001826 ± 0.001822
0.011572 ± 0.009888
0.002483 ± 0.005203
0.015242 ± 0.010510
0.010170 ± 0.003688
0.004709 ± 0.007578
0.003758 ± 0.007367
0.106459 ± 0.013059
0.056296 ± 0.007868
0.126204 ± 0.012854
0.052051 ± 0.010111
0.001559 ± 0.004737
0.032111 ± 0.012709
0.056095 ± 0.009880
0.033664 ± 0.012780
In the next subsections, we explore graphical comparisons between methods within the same class of approaches. All ROCs and pROCs are leave-one-out (LOO) cross-validated and aggregated over all data sets with the ChemDB background. The plots visually illustrate qualitative comparisons of each method.
Aggregating similarities (parameter-free)
We next examine the different parameter-free methods of similarity aggregating; the methods which require fitting parameters from the data are not discussed here, but in a later section.
Profile similarity and aggregating numerators and denominators
For Tanimoto similarity, MinMax similarity computed between a profile vector and a database fingerprint (Equation 22) exactly corresponds with separately aggregating the numerators and denominators (NUMDEN-SIM, Equation 19). So, in the case of binary fingerprints, the only distinct profile similarity method we describe is BAYES method (Equation 24).
Aggregating similarities (with learned parameters)
In this section, we consider methods which learn one or two parameters from the data. These methods were easily trained using a standard leave-one-out procedure to select the best parameterization from an exhaustive parameter sweep. Although one can imagine more efficient means of training these models, this is consistent with the literature. In turn, we first evaluate SUM-ET, SUM-EH, and SUM-TP, which learn using only an active set of molecules, and then ETD, BKD, and TPD, which use additional information from inferred inactive molecules randomly chosen from the background database.
From the figure, we observe that ETD consistently outperforms SUM-ET by a small margin. There is little gain in performance from the information the inferred inactive compounds are adding. MINRANK consistently outperforms MAX-SIM and outperforms ETD in the early section of the curves (Figure 7(b)). NUMDEN-SIM's performance falls behind the other methods'.
In most cases, we expect MAX-SIM and MIN-RANK to be of the most use because of their high performance and simple implementation. However, the stronger performance of the methods with learned parameter(s) suggests a small performance gain can be realized from the added effort of tuning a query to each data set.
Here we have described and evaluated a large number of both novel and established methods of performing multiple-molecule queries using publicly available data, cross-validation protocol, and the best performance metrics.
Our benchmarks indicate that MAX-SIM and MIN-RANK, are the best performing methods which do not require parameters to be learned from the data. This is consistent with previous studies. Furthermore, the ETD is the best parameterized multiple-molecule querying method. Although the higher results are not always by a significant margin over TPD and BKD, it is consistent across our data sets and, to some extent, theoretically justifiable from previous work, but it is the first study of its kind on publicly available data and the first study to compare methods using bias-corrected similarity. All performance metrics were computed by evaluating the retrieval of chemicals with similar biological activity from a diverse random background and a background of decoys.
BAYES did not perform as well as expected. This could be because we fully cross-validated these methods when prior studies may not have done so. Hopefully, by releasing our data to the scientific community our results can be confirmed by other laboratories.
We can explain the improved performance of ETD over BKD on the ChemDB and MUV data. Recall that the Tanimoto metrics used by ETD are corrected while the Hamming distance and Simple Matching similarity, used by BKD, are not corrected. This would be expected to produce a small performance gain. Experiments not reported here show that, indeed, the performance of the Tanimoto-based methods degrades slightly when the uncorrected version is used in place of the corrected version. This, however, is not enough to fully explain the performance gain; these uncorrected versions still outperform BKD. This suggests that part of the performance gain is a direct result of building the method on the Tanimoto similarity rather than the Hamming distance.
Furthermore, we can explain the strong performance of the top methods by noting that most other methods implicitly assume that the chemicals in a given activity class are all similar to one another, that they are all in the same cluster of similar molecules. For example, summing or averaging similarities together assumes that the most likely chemicals are similar to all the chemicals in the query and by averaging similarities we are measuring their membership in this 'one-cluster' in a cleaner way. In contrast, for some parameters, the BKD, TPD and ETD a molecule's score can be large if only one query molecule is very similar, regardless of how dissimilar the other query molecules are.
It may be more appropriate to first divide up the query into clusters, and then score the database by their distance from the closest cluster. For example, to reformulate the BAYES method which typically makes one model of the activity class, we could first segment the query into clusters and model each cluster using a separate Bayesian model. This type of mixture modeling could yield some small performance gains but substantially increases the complexity of the method.
The MAX-SIM method has a nice interpretation which does not make the 'one-cluster' assumption. If we assume (1) the probability a given molecule (ℬ) has the same activity (A) as a query molecule ( ) is some positive-monotonic function, f, of their similarity , i.e. P(A|S( , ℬ)) = f(S( , ℬ)), and (2) this function, f, does not vary significantly across chemical space, then sorting chemicals by their maximum similarity to a set of query molecules is a natural way of finding additional molecules with the highest probability of having the same activity.
This interpretation helps us direct future work. For example, by relaxing the second assumption; perhaps the similarity can be tuned by additional indicators, such as the size of the molecule, its solubility, etc., to create a more explicit model of this probability function which would vary appropriately across chemical space. It is also possible that a more rigorous probabilistic framework could further improve retrieval accuracy.
Furthermore, in one of our data sets, the HIV screening data, we only use the 423 confirmed active data points of a 42,682 data point set which also includes 41,175 inactive molecules and 1,081 moderately-active molecules. It should be possible to integrate all this information, to make even better methods for querying databases for molecules with similar activity. Of course, when our query set has been increased to include thousands of molecules of multiple classes, we must design algorithms which not only produce results of biological relevance but also do so efficiently. A robust, efficient algorithm could extrapolate high throughput screening data (which is becoming increasingly available) to help annotate larger databases like PubChem and the ChemDB, and generate hypotheses for additional biological experiments.
Knowing that these are the best multiple-molecule methods helps direct future algorithmic work. The fastest database search methods [30, 31] are sub-linear in the size of the chemical database, but remain linear in the number of the query molecules. It should be possible to further decrease the search complexity by using a similar bounding technique to index the query molecules in addition to indexing the database.
Work supported by an NIH Biomedical Informatics Training grant (LM-07443-01) and an NSF MRI grant (EIA-0321390) to PB, by the UCI Medical Scientist Training Program, and by a Harvey Fellowship to SS. We would like also to acknowledge the OpenBabel project and the excellent comments from our anonymous reviewers.
- Fligner MA, Verducci JS, Blower PE: A Modification of the Jaccard/Tanimoto Similarity Index for Diverse Selection of Chemical Compounds Using Binary Strings. Technometrics. 2002, 44 (2): 110-119. 10.1198/004017002317375064.View ArticleGoogle Scholar
- Flower DR: On the Properties of Bit String-Based Measures of Chemical Similarity. Journal of Chemical Information and Computer Science. 1998, 38: 379-386.View ArticleGoogle Scholar
- James CA, Weininger D, Delany J: Daylight Theory Manual. 2004, [http://www.daylight.com/dayhtml/doc/theory/theory.toc.html]Google Scholar
- Xue L, Godden JF, Stahura FL, Bajorath J: Profile scaling increases the similarity search performance of molecular fingerprints containing numerical descriptors and structureal keys. J Chem Inf Comput Sci. 2003, 43 (4): 1218-1225.View ArticleGoogle Scholar
- Xue L, Stahura FL, Bajorath J: Similarity search profiling reveals effects of fingerprint scaling in virtual screening. J Chem Inf Comput Sci. 2004, 44 (6): 2032-2039.View ArticleGoogle Scholar
- Hert J, Willett P, Wilton DJ, Acklin P, Azzaoui K, Jacoby E, Schuffenhauer A: Comparison of Fingerprint-Based Methods for Virtual Screening Using Multiple Bioactive Reference Structures. J Chem Inf Comput Sci. 2004, 44 (3): 1177-1185. 10.1021/ci034231b.View ArticleGoogle Scholar
- Hert J, Willett P, Wilton DJ, Acklin P, Azzaoui K, Jacoby E, Schuffenhauer A: Enhancing the effectiveness of similarity-based virtual screening using nearest-neighbor information. Journal of Medicinal Chemistry. 2005, 48 (3): 7049-54. 10.1021/jm050316n.View ArticleGoogle Scholar
- Hert J, Willett P, Wilton D, Acklin P, Azzaoui K, Jacoby E, Schuffenhauer A: New Methods for Ligand-Based Virtual Screening: Use of Data Fusion and Machine Learning to Enhance the Effectiveness of Similarity Searching. J Chem Inf Model. 2006, 46 (2): 462-470. 10.1021/ci050348j.View ArticleGoogle Scholar
- Rupp MSG, Proschak E: Reasoning about molecular similarity and properties. J Chem Inf Model. 2007, 47 (6): 2280-2286. 10.1021/ci700274r.View ArticleGoogle Scholar
- Singh R: Reasoning about molecular similarity and properties. Proc IEEE Comput Syst Bioinform Conf. 2004, 266-277.Google Scholar
- Singh R: Surface similarity-based molecular query-retrieval. BMC Cell Biol. 2007, 8 (Suppl 1): S6-10.1186/1471-2121-8-S1-S6.View ArticleGoogle Scholar
- Swamidass SJ, Chen J, Bruand J, Phung P, Ralaivola L, Baldi P: Kernels for Small Molecules and the Prediction of Mutagenicity, Toxicity, and Anti-Cancer Activity. Bioinformatics. 2005, 21 (Supplement 1): i359-i368. 10.1093/bioinformatics/bti1055.View ArticleGoogle Scholar
- Jin B, Zhang Y, Wang B: Granular kernel trees with parallel genetic algorithms for drug activity comparisons. Int J Data Min Bioinform. 2007, 1: 270-285. 10.1504/IJDMB.2007.011613.View ArticleGoogle Scholar
- Swamidass SJ, Baldi P: Mathematical Correction for Fingerprint Similarity Measures to Improve Chemical Retrieval. J Chem Inf Model. 2007, 47 (3): 952-964. 10.1021/ci600526a.View ArticleGoogle Scholar
- Holliday JD, Hu CY, Willett P: Grouping of coefficients for the calculation of inter-molecular similarity and dissimilarity using 2D fragment bit-strings. Comb Chem High Throughput Screen. 2002, 5 (2): 155-66.View ArticleGoogle Scholar
- Harper G, Bradshaw J, Gittins JC, Green DVS, Leach AR: Prediction of Biological Activity for High-Throughput Screening Using Binary Kernel Discrimination. J Chem Inf Comput Sci. 2001, 41 (5): 1295-1300.View ArticleGoogle Scholar
- Chen B, Harrison RF, Pasupa K, Willett P, Wilton DJ, Wood DJ, Lewell XQ: Virtual Screening Using Binary Kernel Discrimination: Effect of Noisy Training Data and the Optimization of Performance. J Chem Inf Model. 2006, 46 (2): 478-486. 10.1021/ci0505426.View ArticleGoogle Scholar
- Klon AE, Glick M, Davies JW: Combination of a naive Bayes classifier with consensus scoring improves enrichment of high-throughput docking results. J Med Chem. 2004, 47 (18): 4356-4359. 10.1021/jm049970d.View ArticleGoogle Scholar
- Klon AE, Glick M, Thoma M, Acklin P, Davies JW: Finding more needles in the haystack: A simple and efficient method for improving high-throughput docking results. J Med Chem. 2004, 47 (11): 2743-2749. 10.1021/jm030363k.View ArticleGoogle Scholar
- Oprea T, Davis A, Teague S, Leeson P: Is There a Difference between Leads and Drugs? A Historical Perspective. J Chem Inf Comput Sci. 2001, 41 (5): 1308-1315.View ArticleGoogle Scholar
- Stahl M, Rarey M: Detailed Analysis of Scoring Funtions for Virtual Screening. Journal of Medicinal Chemistry. 2001, 44 (7): 1035-1042. 10.1021/jm0003992.View ArticleGoogle Scholar
- Sutherland JJ, O'Brien LA, Weaver DF: Spline-Fitting with a Genetic Algorithm: A Method for Developing Classification Structure-Activity Relationships. Journal of Chemical Information and Computer Sciences. 2003, 43 (6): 1906-1915.Google Scholar
- The Developmental Therapeutics Program NCI/NIH. [http://dtp.nci.nih.gov/]
- Truchon JF, Bayly CI: Evaluating virtual screening methods: good and bad metrics for the 'early recognition' problem. Journal of Chemical Information and Modeling. 2007, 47 (2): 488-508. 10.1021/ci600426e.View ArticleGoogle Scholar
- Rohrer SG, Baumann K: Maximum Unbiased Validation (MUV) Data Sets for Virtual Screening Based on PubChem Bioactivity Data. J Chem Inf Model. 2009, 49 (2): 169-184. 10.1021/ci8002649.View ArticleGoogle Scholar
- Clark RD, Webster-Clark DJ: Managing bias in ROC curves. J Comput Aided Mol Des. 2008, 22 (3-4): 141-146. 10.1007/s10822-008-9181-z.View ArticleGoogle Scholar
- Glick M, Klon AE, Acklin P, Davies JW: Enrichment of extremely noisy high-throughput screening data using a naive Bayes classifier. J Biomol Screen. 2004, 9: 32-36. 10.1177/1087057103260590.View ArticleGoogle Scholar
- Heyer LJ, Kruglyak S, Yooseph S: Exploring expression data: identification and analysis of coexpressed genes. Genome Res. 1999, 9 (11): 1106-1115. 10.1101/gr.9.11.1106.View ArticleGoogle Scholar
- Muchmore SW, Debe DA, Metz JT, Brown SP, Martin YC, Hajduk PJ: Application of Belief Theory to Similarity Data Fusion for Use in Analog Searching and Lead Hopping. J Chem Inf Model. 2008, 48 (5): 941-948. 10.1021/ci7004498.View ArticleGoogle Scholar
- Swamidass SJ, Baldi P: Bounds and Algorithms for Fast Exact Searches of Chemical Fingerprints in Linear and Sublinear Time. J Chem Inf Model. 2007, 47 (2): 302-317. 10.1021/ci600358f.View ArticleGoogle Scholar
- Baldi P, Hirschberg DS, Nasr RJ: Speeding Up Chemical Database Searches Using a Proximity Filter Based on the Logical Exclusive OR. J Chem Inf Model. 2008, 48 (7): 1367-1378. 10.1021/ci800076s.View ArticleGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.