Skip to main content

The algebraic extended atom-type graph-based model for precise ligand–receptor binding affinity prediction

Abstract

Accurate prediction of ligand-receptor binding affinity is crucial in structure-based drug design, significantly impacting the development of effective drugs. Recent advances in machine learning (ML)–based scoring functions have improved these predictions, yet challenges remain in modeling complex molecular interactions. This study introduces the AGL-EAT-Score, a scoring function that integrates extended atom-type multiscale weighted colored subgraphs with algebraic graph theory. This approach leverages the eigenvalues and eigenvectors of graph Laplacian and adjacency matrices to capture high-level details of specific atom pairwise interactions. Evaluated against benchmark datasets such as CASF-2016, CASF-2013, and the Cathepsin S dataset, the AGL-EAT-Score demonstrates notable accuracy, outperforming existing traditional and ML-based methods. The model’s strength lies in its comprehensive similarity analysis, examining protein sequence, ligand structure, and binding site similarities, thus ensuring minimal bias and over-representation in the training sets. The use of extended atom types in graph coloring enhances the model’s capability to capture the intricacies of protein-ligand interactions. The AGL-EAT-Score marks a significant advancement in drug design, offering a tool that could potentially refine and accelerate the drug discovery process.

Scientific Contribution

The AGL-EAT-Score presents an algebraic graph-based framework that predicts ligand-receptor binding affinity by constructing multiscale weighted colored subgraphs from the 3D structure of protein-ligand complexes. It improves prediction accuracy by modeling interactions between extended atom types, addressing challenges like dataset bias and over-representation. Benchmark evaluations demonstrate that AGL-EAT-Score outperforms existing methods, offering a robust and systematic tool for structure-based drug design.

Introduction

In the rapidly evolving field of structure-based drug design, the precise prediction of ligand-receptor binding affinity stands as a cornerstone of success [1,2,3,4]. This critical aspect determines the efficacy of a drug (ligand) in interacting with its target, typically a protein, shaping the outcome of drug discovery processes. Central to advancing these predictions is the application of graph theory, a vital branch of discrete mathematics that provides a structured framework for modeling complex relationships in molecular interactions [5,6,7,8,9,10,11,12].

Graph theory, with its diverse branches including geometric, algebraic, and topological graph theory, has revolutionized the way scientists approach ligand-receptor binding affinity. Geometric graph theory focuses on spatial connectivity, capturing the essence of geometric relationships within molecular structures [13, 14]. Algebraic graph theory, on the other hand, delves into the realm of algebraic connectivity, utilizing matrix representations like adjacency and Laplacian matrices to elucidate molecular interactions [15, 16]. Topological graph theory bridges graphs and topological spaces, offering insights into the more abstract aspects of molecular configuration [17, 18]. These methodologies have not only enhanced drug discovery but also found widespread applications in biomedical sciences [19,20,21], chemical analysis [22,23,24,25], molecular property evaluation [26, 27], and drug repurposing [28, 29].

Scoring functions (SFs) are computational methods used to evaluate protein-ligand interactions and are crucial in structure-based drug design for differentiating between viable and non-viable hypotheses. These scoring functions, based on their theoretical underpinnings, can generally be categorized into the following types: physics-based scoring functions [30,31,32,33,34], empirical scoring functions [35,36,37], knowledge-based scoring functions [38, 39], and the increasingly prominent machine learning (ML)-based scoring functions [40, 41]. ML-based SFs, in particular, have garnered attention for their superior performance, driven by extensive datasets, comprehensive molecular descriptors, and advanced machine learning algorithms. However, the efficacy of these functions is often contingent on the size of the training set and the similarity between the training and test sets, a challenge that has been the focus of several recent studies [42,43,44,45,46,47].

Significant strides in developing machine learning-based scoring functions have been made by utilizing three distinct types of descriptors. These include physics-based descriptors, which cover aspects like electrostatic binding energies and atomic interactions (Coulombic and van der Waals) [48]; descriptors based on geometric graph theory [13]; and those derived from algebraic topology [49]. The core idea behind these methodologies is the assumption that the essential physical phenomena are typically found within low-dimensional spaces or manifolds, even though they exist in a broader, high-dimensional data space. This concept, while recognized in the field of manifold learning, presents a major challenge: effectively translating critical physical information from a high-dimensional context into a practical, low-dimensional format for molecules and their complexes. A notable approach to tackle this challenge is the application of multiscale weighted colored subgraphs (MWCS) [50]. In this approach, a protein’s structural graph is colored based on the types of interactions between its nodes, leading to the formation of distinct subgraphs. This method stands out for its simplicity, low-dimensional nature, and robustness. A key advantage is its minimal data input requirements for binding affinity predictions, which only need atomic names and coordinates. This simplicity allows the method to circumvent complex data processing and parameterization steps, eliminating the need for molecular mechanical force fields, like charges, bond measurements, van der Waals parameters, and others. This not only streamlines the process but also reduces errors often associated with parameterization.

In our prior research [14], we developed the sybylGGL-Score, a sophisticated geometric graph-based method using extended multiscale weighted colored subgraphs for protein-ligand complexes. This approach leveraged graph coloring techniques based on protein atom names and ligand SYBYL atom types [51]. While sybylGGL-Score demonstrated exceptional efficacy in predicting protein-ligand binding affinity, surpassing other advanced methods, it did not fully explore the potential of algebraic graph theory within the extended MWCS framework.

In our current study, we take this concept further by developing an algebraic graph-based MWCS with extended atom-type graph coloring, known as the AGL-EAT-Score. This model employs both the Laplacian and adjacency matrices to represent subgraphs, characterizing molecules and their interactions through eigenvalues and eigenvectors. The effectiveness of AGL-EAT-Score has been rigorously evaluated using benchmark datasets like CASF-2016, CASF-2013, and the Cathepsin S dataset. To enhance our model’s robustness, we conducted a similarity search to eliminate redundant complexes from our training sets, ensuring a more reliable analysis. A summary of the models used for performance comparison, including key features and algorithms, is presented in Table 1.

Table 1 Some state-of-the-art machine learning-based scoring functions

Materials and methods

Extended atom-type multiscale weighted colored subgraphs

In this section, we explore the development of comprehensive graph theory descriptors for a biomolecule or molecular complex. A biomolecular graph, denoted as \({\mathcal {G}}({\mathcal {V}}, {\mathcal {E}})\), is composed of vertices \({\mathcal {V}}\) and edges \({\mathcal {E}}\), providing a powerful tool for representing non-covalent interactions among atoms within the molecule. This graph theory representation is further enriched through the technique of graph coloring, which assigns distinct labels to various types of elements. This coloring process creates a graph that encodes different atomic interactions, enabling the construction of an inclusive and simplified representation of the dataset. Within this framework, atoms in the molecule, identified by these labels, are organized into subgraphs, and the colored edges signify element-specific interactions.

In our previous studies [13, 15], the classification of interactions relied on combinations of element symbols of protein-ligand atoms involved, such as C-O, C-N, etc. Following our most recent work [14], bipartite-colored subgraphs are defined for protein-ligand complexes, where graph coloring is based on extended atom types for proteins and SYBYL atom types for ligands. Protein atom types are identified by their names within the protein data bank (PDB) structure such as carbon alpha (CA), carbon beta (CB), carbon delta-1 (CD1), etc. These atom names serve as identifiers for specific positions within a protein’s three-dimensional arrangement. They help define the individual atoms that constitute amino acids, the building blocks of proteins, and provide crucial information about their spatial arrangement and chemical properties. A total of 37 distinct atom names are considered that are frequently found in protein structures within the PDB database. In ligand Tripos Mol2 structure, SYBYL atom types classify atoms based on their chemical attributes and surroundings within molecular structures, aiding in the identification of diverse atom categories, considering factors like hybridization state, bonding characteristics, and potential interactions. The incorporation of SYBYL atom types enables precise classification, including distinct subtypes for Carbon (C) elements, such as C.1, C.2, C.3, C.ar, and C.cat. The set \({\mathcal {A}}_p\) represents atom names of proteins,

$$\begin{aligned} {\mathcal {A}}_p&= \{C, CA, \ldots , N, ND1, \ldots , O,OD1, \ldots , SD, SG\} \end{aligned}$$
(1)

And, the set \({\mathcal {A}}_l\) represents atom types of ligands,

$$\begin{aligned} {\mathcal {A}}_l&= \{As, B, Be, \ldots , C.1, C.2, \ldots , N.1, N.2, \ldots , V, Zn\} \end{aligned}$$
(2)

For convenience, we define \({\mathcal {A}}\) as the collection of all atom types within a given molecular dataset as described above, where \({\mathcal {A}}_k\) denotes the atom type at the ith position within the set. We further symbolize the subgraph vertices as \({\mathcal {V}}\), which are characterized by the atom coordinates \(r_i\) and their corresponding atom types \(\alpha _i\):

$$\begin{aligned} {\mathcal {V}}=\{ ({\textbf{r}}_i,\alpha _i)| {\textbf{r}}_i\in {\mathbb {R}}^3; \alpha _i\in {\mathcal {A}}; i=1,2,\cdots ,N\} \end{aligned}$$
(3)

Additionally, we symbolize the subgraph edges as \({\mathcal {E}}\) and defined as follows:

$$\begin{aligned} {\mathcal {E}}&= \{ \Phi (\Vert {{\textbf{r}}_i-{\textbf{r}}_j}\Vert ;\eta _{kk'})| \alpha _i={\mathcal {A}}_k,\, \alpha _j = {\mathcal {A}}_{k'}; \nonumber \\&\quad i,j=1,2,\cdots ,N;\, \Vert {{\textbf{r}}_i-{\textbf{r}}_j}\Vert \le c \}, \end{aligned}$$
(4)

where \(\Vert {{\textbf{r}}_i-{\textbf{r}}_j}\Vert \) denotes the Euclidean distance between the ith and jth atoms, and c represents a specified cutoff distance that defines the binding site between the atoms of types \({\mathcal {A}}_k\) and \({\mathcal {A}}_{k'}\). While c is a learnable parameter that can be optimized through cross-validation, we have chosen \(c=12\) Å  for this work, as it has proven effective in our previous studies [14, 61] and is also utilized in RF-Score [62]. We calculate the edge weights based on the characteristics distance \(\eta _{kk'}\) between pairs of atom types \({\mathcal {A}}_k\) and \({\mathcal {A}}_{k'}\) using the subgraph weight function \(\Phi \). The weight function \(\Phi \) assesses the interaction strength between atoms, taking into account their Euclidean distances, and it satisfies the following conditions:

$$\begin{aligned} \Phi (\Vert {{\textbf{r}}_i-{\textbf{r}}_j}\Vert ;\eta _{kk'})&=1, \quad \textrm{as}\; \Vert {{\textbf{r}}_i-{\textbf{r}}_j}\Vert \rightarrow 0,\nonumber \\ \Phi (\Vert {{\textbf{r}}_i-{\textbf{r}}_j}\Vert ;\eta _{kk'})&=0, \quad \textrm{as}\; \Vert {{\textbf{r}}_i-{\textbf{r}}_j}\Vert \rightarrow \infty , \alpha _i={\mathcal {A}}_k,\, \alpha _j = {\mathcal {A}}_{k'}. \end{aligned}$$
(5)

Often, a popular selection for \(\Phi \) is the generalized exponential function or the generalized Lorentz function denoted as follows:

$$\begin{aligned} \Phi _E(\Vert {{\textbf{r}}_i-{\textbf{r}}_j}\Vert ;\eta _{kk'}) = e^{-(\Vert {{\textbf{r}}_i-{\textbf{r}}_j}\Vert /\eta _{kk'})^\kappa }, \quad \kappa >0, \end{aligned}$$
(6)

and

$$\begin{aligned} \Phi _L(\Vert {{\textbf{r}}_i-{\textbf{r}}_j}\Vert ;\eta _{kk'}) = \frac{1}{1+\left( \Vert {{\textbf{r}}_i-{\textbf{r}}_j}\Vert /\eta _{kk'}\right) ^\kappa }, \quad \kappa >0, \end{aligned}$$
(7)

The generated weighted colored subgraph \({\mathcal {G}}({\mathcal {V}}, {\mathcal {E}})\) offers a robust depiction of molecular attributes at the atomic scale. Analyzing this subgraph allows us to extract detailed molecular descriptors and explore the system’s multiscale behavior. This behavior is a result of considering various characteristic distances \(\eta _{kk'}\) for different atom type pairs, enabling the creation of diverse and scalable graph-based descriptors, including the geometric subgraph centrality (GSC), defined as

$$\begin{aligned} \textrm{GSC}(\eta _{kk'})&=\sum _i \mu _i(\eta _{kk'})=\sum _i \sum _j \Phi (\Vert {{\textbf{r}}_i-{\textbf{r}}_j}\Vert ;\eta _{kk'}),\nonumber \\&\quad \alpha _i = {\mathcal {A}}_{k},\, \alpha _j = {\mathcal {A}}_{k'}, \end{aligned}$$
(8)

This extends the concept of the bipartite subgraph we utilized in our prior research on predicting protein-ligand binding affinities and free energy ranking [13], where every edge connects an atom in the protein with an atom in the ligand. The matrix representations of such subgraphs offer a simple and expressive way to describe the interactions between subgraph elements using matrices. As following our previous study [15], we utilized two highly significant matrices: the Laplacian matrix and the adjacency matrix in our present work.

For each atom-type pair subgraph, we compute the Laplacian matrix \(L\left( \eta _{k k^{\prime }}\right) \), which is defined as follows:

$$\begin{aligned} L_{i j}\left( \eta _{k k^{\prime }}\right) =&{\left\{ \begin{array}{ll}-\Phi \left( \left\| {\textbf{r}}_i-{\textbf{r}}_j\right\| ; \eta _{k k^{\prime }}\right) & \text{ if } i \ne j, \alpha _i={\mathcal {A}}_k, \alpha _j={\mathcal {A}}_{k^{\prime }} \\ & \text{ and } \left\| {\textbf{r}}_i-{\textbf{r}}_j\right\| \le c; \\ -\sum _j L_{i j} & \text { if } i=j\end{array}\right. } \end{aligned}$$
(9)

It’s important to emphasize that all eigenvalues of the Laplacian matrix are nonnegative. Let us define the eigenvalues and eigenvectors of \(L\left( \eta _{k k^{\prime }}\right) \) as \({\lambda _j}^L\),\(j = 1,2,\ldots \) and \({u_j}^L\),\(j = 1,2,\ldots \).

For each atom-type pair subgraph, we also compute the adjacency matrix \(A\left( \eta _{k k^{\prime }}\right) \), which is defined as follows:

$$\begin{aligned} A_{i j}\left( \eta _{k k^{\prime }}\right) =&{\left\{ \begin{array}{ll}-\Phi \left( \left\| {\textbf{r}}_i-{\textbf{r}}_j\right\| ; \eta _{k k^{\prime }}\right) & \text{ if } i \ne j, \alpha _i={\mathcal {A}}_k, \alpha _j={\mathcal {A}}_{k^{\prime }} \\ & \text{ and } \left\| {\textbf{r}}_i-{\textbf{r}}_j\right\| \le c; \\ 0 & \text { if } i=j\end{array}\right. } \end{aligned}$$
(10)

Indeed, eigenvalue analysis is widely recognized as a computationally expensive task. However, our AGL-EAT-Score approach benefits from two crucial factors that enhance its computational efficiency. Firstly, we restrict matrix constructions to encompass solely those atoms located in the proximity of the protein-ligand binding site. To determine the binding site of the protein-ligand complex, we define a cubic domain extending 12 Å  from the ligand atoms. This approach involves selecting protein atoms that fall within this defined cubic region by applying the cutoff distance to the ligand’s maximum and minimum coordinates along each axis. Furthermore, our atom-type-specific criteria further narrow down the atoms involved in each matrix construction. Consequently, we work with numerous small matrices, which enables an efficient spectral approach for analyzing protein-ligand binding affinities.

Algebraic graph learning

We employed Machine learning algorithms to analyze the eigenvalue statistics descriptors generated from the weighted colored subgraph Laplacian matrix or adjacency matrix as discussed above. For a given protein-ligand complex, we considered 37 unique atom names in the protein and 45 SYBYL atom types for the ligand for graph coloring, resulting in \(37\times 45=1665\) unique colored subgraphs. From each subgraph, we extracted nine statistics of the positive eigenvalues: the sum, mean, median, minimum, maximum, standard deviation, variance, number of positive eigenvalues, and the sum of the squares of the positive eigenvalues, as well as the edge counts of the subgraph. This yielded a total of \(1665\times 10=16650\) features for a complex. These extracted features will map the high-dimensional structures of biomolecular complexes into low-dimensional representations while preserving the vital physical and chemical properties of the complexes.

Supervised machine learning algorithms can include both classification and regression tasks, the labeled dataset is divided into two subsets: a training set and a test set. Let us denote \({\mathcal {G}}({\mathcal {X}}_i, \lambda )\), a function encoding the geometric information of a molecule into suitable graph representations using \({\mathcal {X}}_i\), a labeled dataset corresponding to the ith data point in the training set and \(\lambda \), a set of kernel parameters. The following loss minimization problem further reformulates the optimization process for training a machine learning model,

$$\begin{aligned} \min _{\lambda , \theta } \sum _{i\in I} {\mathcal {L}}({\textbf{y}}_i,{\mathcal {G}}({\mathcal {X}}_i, \lambda ); \theta ) \end{aligned}$$
(11)

Here, \({\mathcal {L}}\) denotes a scalar loss function that needs to be minimized, and \({\textbf{y}}_i\) refers to the labels assigned to the ith sample in the training set I. The set \(\theta \) includes hyperparameters that are dependent on the chosen machine learning algorithm and are usually tuned to achieve optimal performance. While various machine learning algorithms, including random forest, gradient boosting trees, graph neural networks, and convolutional neural networks, can be applied alongside the graph descriptors introduced in our study, our primary focus is to assess the effectiveness of the proposed algebraic graph features. To achieve this, we emphasize the use of gradient boosting trees (GBTs) as a regression model, a machine learning algorithm recognized for its robustness against overfitting. The 16,650 features extracted from the eigenvalues of multiscale weighted colored subgraphs of a protein-ligand complex serve as the input to the regression model, while the output corresponds to the predicted binding affinity. The visual depiction of our algebraic graph-based learning approach is presented in Fig. 1.

In our study, the implemented GBDT module in scikit-learn version 0.24.1 with the parameters, \(\texttt {n\_estimators}=20 000\), \(\texttt {max\_depth} = 8\), min_samples_split = 2, learning_rate = 0.005, loss = ls, subsample = 0.7, and max_features = sqrt.

Fig. 1
figure 1

Visualization of the AGL-EAT graph learning approach. First column: binding site of the molecular complex identified by its PDBID: 3up2. Second column: three different kinds of protein-ligand atom-type pairs. Third column: weighted colored subgraph representations of the corresponding atom-pairs. Fourth column: eigenvalues of the subgraph Laplacian and adjacency matrices. Fifth column: various statistics of these eigenvalues. Fifth column: advanced machine learning models like gradient boosting trees to combine and process these statistics for training and making predictions

Datasets

To assess the validity of our proposed model, we employed two widely acknowledged PDBbind benchmark datasets: CASF-2013 and CASF-2016, and the CatS dataset, which is part of the D3R datasets, a global grand challenge in drug design.

The PDBbind database offers an extensive repository of experimentally determined binding affinity data for biomolecular complexes found within the Protein Data Bank (PDB) [63]. Every PDBbind benchmark dataset comprises three intersecting subsets: the general set, the refined set, and the core set, with the core set being a subset of the general set and the refined set. The core set serves as the testing dataset for the respective benchmark. More details of the PDBbind datasets can be explored in the PDBbind website http://www.pdbbind.org.cn/.

On the other hand, the Drug Design Data Resource (D3R) Grand Challenge [64, 65] focuses on specific datasets, each of which includes a single protein and multiple ligands, all accompanied by measured affinity data. The Cathepsin S (CatS) dataset of D3R [65] consists of 459 CatS inhibitors for binding affinity prediction. Detailed information about this dataset can be found on the official D3R website https://drugdesigndata.org/about/grand-challenge-4/cathepsin_s. A summary of all the datasets used in this study has been listed in Tables 2 and 3.

Table 2 Summary of PDBbind datasets used to validate our model
Table 3 Summary of CatS dataset used to validate our model

Evaluation metrics

In this research, we evaluated the performance of our model’s scoring power through various metrics, including including root mean squared error (RMSE), mean absolute error (MAE), standard deviation (SD), as well as Pearson’s Correlation Coefficient, Kendall’s Tau, and Spearman’s Rho between the experimental and predicted pK values. Pearson’s Correlation Coefficient measures the linear relationship between two variables, ranging from -1 to 1, where 1 indicates a perfect positive linear relationship, -1 indicates a perfect negative linear relationship, and 0 indicates no linear relationship, thereby assessing how well the model’s predictions correlate with actual values. Kendall’s Tau measures the ordinal association between two variables by evaluating the strength and direction of the association through concordant and discordant pairs, making it useful for datasets with ordinal data or non-linear relationships. Spearman’s Rho assesses the monotonic relationship between two variables, similar to Pearson’s but based on rank order rather than actual values, making it suitable for evaluating models where the relationship may not be strictly linear but still monotonic.

For comparison purpose, we employed Pearson’s Correlation Coefficient for the PDBbind datasets, while for the CatS dataset, we utilized both Kendall’s Tau and Spearman’s Rho. These specific metrics for each dataset were chosen based on their popularity and acceptance within the research community.

Results and discussion

In this section, we present the results of hyperparameter optimization and the performance of our proposed AGL-EAT-Score on various benchmark datasets. Furthermore, we conduct a rigorous similarity test analysis to validate the robustness of AGL-EAT-Score. In our study, we adopt the notation \(\text {AGL-EAT}_{\beta ,\kappa ,\tau }^{{\mathcal {B}}}\) to characterize algebraic graph learning for extended atom types features. Here, \({\mathcal {B}}\) represents the type of matrix used, specifically, \({\mathcal {B}} =\textrm{Adj}\) denotes the adjacency matrix, while \({\mathcal {B}} = \textrm{Lap}\) refers to the Laplacian matrix. The parameter \(\beta \) indicates the specific kernel types, while \(\kappa \) and \(\tau \) correspond to the respective kernel parameters. Specifically, we employ the generalized exponential kernel denoted by \(\beta =\Phi _E\), and the generalized Lorentz kernel represented by \(\beta = \Phi _L\), for generating AGL-EAT features. The parameter \(\tau \) is a scaling factor determining the characteristic distance \(\eta _{kk'}\) between atom type k and atom type \(k'\). This distance is calculated as \(\eta _{kk'}=\tau (r_k + r_{k'})\). Here, \(r_k\) and \(r_{k'}\) are the van der Waals radii of the atoms of type k and type \(k'\), respectively.

Hyperparameter optimization

Hyperparameter optimization of a machine learning model involves the search for the most suitable combination of hyperparameter values that yield the best performance on a specific dataset, all within a reasonable time frame. To make our AGL-EAT-Score model work best for each benchmark, we optimize two key parameters \(\kappa \) and \(\tau \) for a given kernel type \(\beta \) and matrix type \({\mathcal {B}}\). We use a five-fold cross-validation (CV) along with a grid search method to find the best values for \(\tau \), which we search within the range of 0.5 to 10, and for \(\kappa \), within the range of 0.5 to 20. We increment both parameters by 0.5 in the search. Higher values for the power parameter \(\kappa \) are chosen to approximate the ideal low-pass filter (ILF) [13].

We applied a five-fold cross-validation on the refined set excluding the core set to find the optimized kernel parameters for each of the PDBbind benchmark datasets. We train our model on both the PDBbind refined set and general set using the derived optimized hyperparameters and then evaluate the model’s performance on the corresponding test set. However, for the CatS dataset, we perform a five-fold cross-validation on the training set and evaluate the model’s performance on the CatS test set provided by the D3R database. A detailed discussion of optimized hyperparameters and the model’s performances on each of the datasets used in this study has been documented in Figs. S1, S2, and S3 in the Supporting Information.

CASF-2016 benchmark

For the CASF-2016 benchmark, the optimal kernel parameters with the Adjacency matrix are \(\beta =\Phi _E\), \(\kappa =16.5\), and \(\tau =3.0 \) with a median Pearson’s correlation coefficient \(R_p =0.796 \) and the optimal kernel parameters with the Laplacian matrix are \(\beta =\Phi _E\), \(\kappa =19.5\), and \(\tau =2.5 \) with a median \(R_p\) of 0.795.

After the best models have been identified for each benchmark, our goal is to assess their performance on the test set by calculating Pearson’s correlation coefficient between the predicted and experimental binding affinities. We first train each model using the refined set, and then make predictions on the test set. To ensure reliable predictions, we repeat this process up to 50 times and compute the average of all predicted values to obtain the final predicted set. Next, we train the model using the general set, excluding the CASF-2016 core set. Training on this larger dataset, despite the lower quality 3D structures, will validate the robustness of the proposed models against more diverse and potentially irrelevant data. Similarly, we repeat the model training 50 times to generate predicted values, and these values are then averaged to produce the final prediction.

A summary of the performances of the best AGL-EAT models on the CASF-2016 is presented in Table 4. The best model reported is \(\text {AGL-EAT}_{\Phi _E,16.5,3.0}^{\textrm{Adj}}\) with an \(R_p\) of 0.873. A comparison within the CASF-2016 benchmark is presented in Fig. 2b, showcasing our model’s superior performance as it ranks at the top among other models. The efficiency of the AGL-EAT-Score is also noteworthy, particularly regarding the running time for feature generation and model training. On average, generating features for a given protein-ligand complex takes less than one second. For model training, we utilized a high-performance computing cluster with one node, a single core, and a memory allocation of 10 GB, powered by an Intel(R) Xeon(R) Gold 6126 CPU running at 2.60 GHz. The total running time for training the AGL-EAT models with the PDBbind v2016 general set was approximately 50 min and 8 s, while training with the PDBbind v2016 refined set took about 12 min and 58 s. A comparison of models’ running time is provided in supplementary Table S1.

Table 4 Performance of various AGL-EAT-score models on CASF–2016 test set

CASF-2013 benchmark

Subsequently, we consider the CASF-2013 benchmark dataset from the PDBbind database. Since the CASF-2013 training set is smaller than the previously discussed CASF-2016 set, one might expect lower performance compared to the above results. However, the CASF-2013 benchmark will further confirm the robustness of the proposed model with less informative training data. The AGL-EAT model with the Adjacency matrix shows optimal kernel parameters \(\kappa = 5.5\) and \(\tau = 2.0\) with kernel type \(\beta =\Phi _E\), resulting in a median Pearson’s correlation coefficient \(R_p = 0.795\). On the other hand, the model with the Laplacian matrix has optimal kernel parameters \(\kappa = 4.5\) and \(\tau = 2.0\) with kernel type \(\beta =\Phi _E\), delivering a median \(R_p = 0.796\).

Once the top-performing models for this benchmark are identified, we evaluate the performance of our model on the corresponding test set by calculating Pearson’s correlation coefficient between predicted and experimental binding affinities. Following CASF-2016, a similar repetition of model training, totaling 50 cycles, both for the refined set and the general set is undertaken to generate predicted values, which are subsequently averaged to derive the final prediction. The performance summary of the top AGL-EAT models for the CASF-2013 benchmark is outlined in Table 5, with the \(\text {AGL-EAT}_{\Phi _E,5.5,2.0}^{\textrm{Adj}}\) model achieving the highest Pearson correlation coefficient (\(R_p = 0.845\)). A visual comparison in the benchmark, depicted in Fig. 2a, demonstrates our model’s leading performance against competing models, underscoring its effectiveness in the evaluation.

Table 5 Performance of various AGL-EAT-Score models on CASF–2013 test set
Fig. 2
figure 2

Performance comparison plot measured in Pearson’s correlation of our AGL-EAT-Score and other machine learning-based models on the a CASF-2013, and b CASF-2016 benchmark datasets. Our model AGL-EAT-Score (highlighted in red color) scores. In CASF-2013 benchmark, our model achieved \(R_p = 0.845\) and the results of other methods obtained from prior research [59, 60, 66]. In CASF-2016 benchmark, our model achieved \(R_p = 0.873\) and the results of other methods obtained from prior research [57, 59, 67]

CatS dataset

In the context of the CatS dataset, we employ Kendall’s tau correlation coefficient as the performance evaluation metric, which assesses the model’s ability to capture the ranking and correlation of predicted binding affinities with the actual values, providing a comprehensive evaluation of the model’s performance. The optimal kernel parameters for the Adjacency matrix are \(\kappa =12.5\) and \(\tau = 8.0\) with exponential kernel type producing a median Kendall’s tau of 0.57837. The optimal kernel parameters with the Laplacian matrix are \(\kappa =16.5\) and \(\tau = 10.0\) for \(\beta =\Phi _E\) with a median Kendall’s tau of 0.57305.

After having the best-optimized model for the CatS training set, we assess the performance on the test set by calculating Kendall’s tau correlation coefficient between the predicted and experimental binding affinities. We train each of these optimized models using the training set and subsequently generate predictions for the test set. We repeat this process up to 50 times and calculate the average of all predicted values to yield the final predicted set, from which we calculated Kentall’s tau correlation coefficient (Kendall’s \(\tau \)) and Spearman’s rho correlation coefficient (Spearman’s \(\rho \)). Table 6 reports the performance of our models for CatS dataset. The best-performing model for this dataset is the \(\text {AGL-EAT}_{\Phi _E,5.5,2.0}^{\textrm{L}}\), achieving a Kendall’s tau of 0.552, and a Spearman’s \(\rho \) of 0.742. Figure 3 provides a performance comparison between our AGL-EAT-Score and other machine learning-based models taken from the official results of the D3R GC4 [65]. The results show that our model outperforms the affinity ranking of 459 CatS compounds in terms of Kendall’s \(\tau \) and Spearman’s \(\rho \).

Table 6 Performance of various AGL-EAT-Score models on CatS data set
Fig. 3
figure 3

Performance comparison plot of our AGL-EAT-Score and other machine learning-based models on the Cathepsin S (CatS) dataset. Our model AGL-EAT-Score (highlighted in red color) scores Kendall’s \(\tau = 0.55\) and Spearman’s \(\rho = 0.74\)

Validation of the robustness of AGL-EAT model

Non-redundant training sets

The performance of machine learning scoring functions is known to be influenced by the size of the training set and the degree of similarity between the training set and the test set [68, 69]. However, robust and consistent machine learning-based scoring functions are expected to show a linear improvement with respect to the increment of relevant information in the training data [70,71,72]. To this end, we investigated our model’s performance on the multiple non-redundant training sets. These non-redundant training sets are datasets that have been carefully curated to ensure that the included complexes are distinct from the test set complexes and do not contain redundant or highly similar complexes with test complexes. These sets are designed to reduce bias and overfitting in machine learning models by providing diverse and representative training data.

Similarity computation

In our endeavor to construct a non-redundant training set, we explored the evaluation of three distinct similarity measurements between protein-ligand complexes. The first of these measurements, known as protein sequence similarity and denoted as \(P_s\), quantifies the likeness between protein sequences. The second metric, \(L_s\), deals with the structural resemblance of ligands. Lastly, the third measurement, \(BS_s\), concerns itself with the comparison of protein-ligand binding sites. These similarity metrics collectively underpin the process of generating our non-redundant training set, allowing us to make informed decisions regarding redundancy reduction in the dataset. To compute the sequence similarity of two protein structures, we used \(``ggsearch36''\) from FASTA (version 36.3.8) [73], which employs a global-global (Needleman-Wunsch) search algorithm. The ligand structure similarity was computed using ROCS (version 3.5.1.1) [74, 75], which employs a Gaussian function with smoothness characteristics to model the molecular volume, allowing for systematic optimization to achieve the most accurate global fit. The binding site similarity was computed using PocketMatch (version 2.1) [76, 77], which assesses the similarity of binding sites using structural descriptors like residue properties and interatomic distances. This tool can also provide atomic-level alignments derived from pairings of amino acid residues.

To initiate the search for non-redundant complexes, we eliminated the overlaps between the train and test sets. We will be adopting the terms “hard overlap” to refer to the complexes that overlap between the train set and test sets, and “soft overlap” to refer to the structurally similar complexes in the train set and test sets, as defined by Minyi Su et al [45].

The process of searching for non-redundant training sets has been illustrated in the following contexts and visually represented in Fig. 4.

Fig. 4
figure 4

Visualization of the strategy for searching a non-redundant training set for different similarity cutoffs. In the initial phase, we conducted calculations to determine the cross-similarity between the training and test sets, followed by a process to reduce redundancy. In the following phase, we performed internal similarity calculations on the training set and subsequently reduced redundancy to get the final non-redundant training sets

  1. 1.

    Training Set vs Test Set Similarity Evaluation: We start by eliminating any “hard overlap” complexes shared between the training set and the test set. We then proceed to evaluate the similarity between every complex in the training set against every complex in the test set by employing the three similarity metrics, \(P_s\), \(L_s\) & \(BS_s\) as discussed above.

  2. 2.

    Training Set vs Test Set Redundancy Reduction: A training complex was classified as redundant to the test set if all three similarity metrics were above the defined cutoff and further eliminated from the training set.

  3. 3.

    Training Set Internal Similarity Evaluation: Following that, we compute the similarities among the remaining complexes within the training set. If the measurements for all three similarity metrics between two complexes exceed the defined cutoff, we categorize these complexes as redundant to each other.

  4. 4.

    Training Set Internal Redundancy Reduction: Finally, we adopt a systematic approach to eliminate redundant samples from the training set to get the optimal training sets for different similarity cutoffs.


PDBbind v2016:

In our study, we explored the calculation of similarity for both the PDBbind v2016 general set and the refined set. We derived non-redundant training sets for six distinct similarity cutoff points: \(70\%\), \(75\%\), \(80\%\), \(85\%\), \(90\%\), and \(95\%\). The summary of non-redundant complexes for different similarity cutoffs is listed in Table S2.

However, it’s important to note that a similar comparison for the PDBbind v2015 dataset wasn’t conducted since nearly all its molecules are already included in the PDBbind v2016 dataset, with only a marginal difference of around \(10\%\).


CatS:

The CatS dataset has a limited size of training data, where the performance of the model largely depends on the quality of the information in the training data rather than its size. Therefore, we decided to explore the redundant complexes instead of non-redundant complexes, i.e. where multiple complexes shared common features or functions. Our approach to searching for redundant complexes in the CatS dataset closely paralleled the methodology we employed when investigating non-redundant complexes in the PDBbind dataset. We derived redundant training sets for 10 distinct similarity cutoffs, starting from 45 to 90%. The summary of redundant complexes for different similarity cutoffs for the CatS dataset is presented in Table S3.

Next, we evaluate the performance of our proposed model trained on these redundant datasets. By doing so, we aim to understand how well the model generalizes and performs across different levels of data redundancy. Redundant complexes often share similar features; however, the degree of similarity largely depends on the intrinsic information of the training data and its relevance to the test data. Therefore, we strive to uncover how the level of similarity between the training and test data affects the model’s accuracy.

Performances on PDBbind v2016 Non-redundant training sets

Indeed, it is widely recognized that the size of the training set and the degree of similarity between the training set and the test set have a profound impact on machine learning scoring functions [68, 69]. To this end, we aim to investigate our model’s performance on the multiple non-redundant training sets of PDBbind v2016 General set and Refined set with different levels of redundancy.

Fig. 5
figure 5

Performance comparison plot of non-redundant training sets of PDBbind v2016 RefinedSet and GeneralSet

Figure 5 visualizes the performance of our proposed model when trained on these non-redundant training sets. Interestingly, when we calibrate our model on a training set that shares a high level of similarity with the test set, for example, employing a \(95\%\) similarity cutoff from the General set, we achieved a Pearson correlation of 0.869. This is remarkably close to the Pearson correlation of 0.873 obtained when using the complete training set. These findings emphasize the consistent and robust nature of our model across varying non-redundant training sets, without experiencing significant drops in predictive capabilities. This robust performance underscores the model’s reliability and versatility in handling diverse datasets.

Investigating the performances across various non-redundant training sets reveals another interesting fact: data quality significantly influences the model’s effectiveness. Consider the refined non-redundant training set for the \(95\%\) similarity cutoff, notably smaller in size compared to the general non-redundant training set at the \(70\%\) similarity cutoff. However, the performance of the smaller refined non-redundant set substantially outperforms the larger set. This observation underscores the importance of data quality over quantity, highlighting how the focused, refined data yield more accurate models despite their smaller scale.

Performances on CatS redundant training sets

In exploring the CatS dataset’s redundant training complexes, we observed a nuanced impact of redundancy levels on our model’s performance. Training the model on sets with varying similarity cutoffs, we noted that higher redundancy maintained stable performance, as measured by Kendall’s Tau. As captured in Fig. 6, for \(45\%\) similarity cutoff we achieved a Kendall’s Tau = 0.5512, which remains quite stable up to \(75\%\) similarity cutoff with Kendall’s Tau = 0.5355. This stability suggests the model effectively leverages redundant training complexes without significant loss in predictive power. However, beyond this threshold, performance declined, highlighting the need to balance leveraging redundancy for rich feature extraction and avoiding diminishing returns from overly similar and less diverse training data. This emphasizes the importance of balancing diversity and similarity in collecting training samples.

Fig. 6
figure 6

Performance comparison plot of redundant training sets of CatS Dataset

Importance of extended atom types

We further performed an investigation by expanding our previous work, AGL-Score [15] to justify the impact of the protein-ligand extended atom type (EAT) features in the model’s performance. We follow a systematic approach throughout this investigation.

Fig. 7
figure 7

Feature importance of a AGL-Score [15] b our AGL-EAT-Score model on PDBbind v2016 general set using the mean decrease in impurity of the gradient boosting trees. The horizontal axis on the graph denotes the names of the features, which are represented as various statistical metrics (like sum, mean, median, and others) specific to the atom name groups in the ligand-protein interaction

We initiated the process by ranking the important features of both the AGL-EAT-Score and AGL-Score models, using the PDBbind v2016 general set features as a reference. Figure 7 illustrates the top ten important features of the AGL-EAT-Score and AGL-Score models in this dataset. As shown in Fig. 7a, some of the most important interactions of AGL-Score are C-C, O-C, N-C so on. For the first set of experiments, we replaced the C-C features in AGL-Score with the extended atom type features from AGL-EAT-Score, specifically the C-C.1, CA-C.1, CB-C.1, C-C.2, CA-C.2, CB-C.2, and so on. We then employed this modified feature set to predict the test dataset. To ensure the reliability of predictions, we repeated the process 50 times and obtained the final predicted set by averaging all the predicted values. In the case of the base AGL-Score model, we reported a performance of \(R_p = 0.8559\) on the PDBbind v2016 general set and evidently, incorporating the C-C extended atom type (EAT) features led to an improvement in performance, resulting in \(R_p = 0.8685\).

We performed similar feature replacements for O-C, N-C, C-O, and O-H interactions as well, and the modified feature sets were used for test dataset predictions. A performance comparison of these experiments is presented in Fig. 8. In our final set of experiments, we replaced all the base AGL-Score C-C, O-C, N-C, and O-H interaction features with the corresponding AGL-EAT-Score extended atom type features for prediction purposes, resulting in improved performance with an \(R_p = 0.8715\), named as AGL-Score-combined-eat features in Fig. 8, indicates a significant enhancement. These results demonstrate the benefits of incorporating the extra level of detail in atom-type interactions. The best performance of the model is achieved when all the extended atom types of all the element types are considered.

Fig. 8
figure 8

Performance comparison plot of base AGL-Score [15] (highlighted in yellow color), AGL-EAT-Score (highlighted in red color), and the modified AGL-Score with different extended atom type interaction features from AGL-EAT-Scre (highlighted in blue colors) on the PDBbind v2016 General set

Finally, we compare the AGL-EAT-Score with our previously developed GGL-Score [14] to assess the robustness of algebraic graph features against geometric graph features. AGL-EAT-Score shows notable advantages in predicting binding affinities for protein-ligand complexes. In the CASF 2013 benchmark, AGL-EAT-Score achieved a Pearson R of 0.845, closely matching GGL-Score’s 0.848. Both models performed equally well in the CASF 2016 benchmark, with a Pearson R of 0.873. However, AGL-EAT-Score outperformed GGL-Score in the CatS dataset, recording a Kendall’s Tau of 0.552 and Spearman’s Rho of 0.742, compared to GGL-Score’s 0.487 and 0.674, respectively. This indicates AGL-EAT-Score’s superior ability to capture complex molecular interactions. The foundation of AGL-EAT-Score in algebraic graph theory enables a deeper analysis of structural and algebraic connectivity, revealing insights that geometric representations may overlook. This enhanced modeling capability not only improves predictive accuracy in specific contexts but also underscores AGL-EAT-Score’s adaptability in addressing the intricacies of molecular interactions. Overall, these factors position AGL-EAT-Score as a valuable tool for binding affinity prediction.

Conclusion

In this work, we propose a novel model named Algebraic Graph Learning with Extended Atom-Type Scoring Function (AGL-EAT-Score), which exhibits high accuracy in capturing protein-ligand interaction information. This model is distinguished by its innovative integration of extended atom-type multiscale weighted colored subgraphs and algebraic graph learning, enabling a detailed and sophisticated representation of molecular interactions. The AGL-EAT-Score has demonstrated its efficacy in accurately predicting ligand-receptor binding affinities, showcasing superior performance compared to both traditional and contemporary machine learning-based scoring functions. This was evidenced through extensive evaluations using benchmark datasets such as CASF-2016, CASF-2013, and the CatS dataset.

To further validate the robustness and address concerns of overfitting in machine learning-based scoring functions, we tested the performance of the proposed AGL-EAT-Score against redundant and non-redundant data built on the PDBbind general set v2016 and the CatS dataset. The model’s performance, consistent with the level of training data information, confirms the necessity of incorporating extended atom-type information rather than relying solely on basic element types.

As the field of drug design continues to progress, the proposed AGL-EAT-Score is positioned as a robust, innovative, and essential tool for describing the complex landscape of molecular interactions, thereby contributing significantly to advancements in pharmaceutical research.

Availability of data and materials

The source code is available at the Github repository: https://github.com/MathIntelligence/AGL-ETA-Score-Open.

References

  1. Dhakal A, McKay C, Tanner JJ, Cheng J (2022) Artificial intelligence in the prediction of protein-ligand interactions: recent advances and future directions. Br Bioinform 23(1):476

    Article  Google Scholar 

  2. Kairys V, Baranauskiene L, Kazlauskiene M, Matulis D, Kazlauskas E (2019) Binding affinity in drug design: experimental and computational techniques. Expert Opin Drug Discov 14(8):755–768

    Article  CAS  PubMed  Google Scholar 

  3. Klebe G (2000) Recent developments in structure-based drug design. J Mol Med 78:269–281

    Article  CAS  PubMed  Google Scholar 

  4. Huggins DJ, Sherman W, Tidor B (2012) Rational approaches to improving selectivity in drug design. J Med Chem 55(4):1424–1444

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  5. Mason O, Verwoerd M (2007) Graph theory and networks in biology. IET Syst Biol 1(2):89–119

    Article  CAS  PubMed  Google Scholar 

  6. Vishveshwara S, Brinda K, Kannan N (2002) Protein structure: insights from graph theory. J Theor Comput Chem 1(01):187–211

    Article  CAS  Google Scholar 

  7. Aittokallio T, Schwikowski B (2006) Graph-based methods for analysing networks in cell biology. Br Bioinform 7(3):243–255

    Article  CAS  Google Scholar 

  8. Xia K, Wei G.-W (2016) A review of geometric, topological and graph theory apparatuses for the modeling and analysis of biomolecular data. arXiv preprint arXiv:1612.01735

  9. Samudrala R, Moult J (1998) A graph-theoretic algorithm for comparative modeling of protein structure. J Mol Biol 279(1):287–302

    Article  CAS  PubMed  Google Scholar 

  10. Jacobs DJ, Rader AJ, Kuhn LA, Thorpe MF (2001) Protein flexibility predictions using graph theory. Prot Struct Funct Bioinform 44(2):150–165

    Article  CAS  Google Scholar 

  11. Shen C, Zhang X, Deng Y, Gao J, Wang D, Xu L, Pan P, Hou T, Kang Y (2022) Boosting protein-ligand binding pose prediction and virtual screening based on residue-atom distance likelihood potential and graph transformer. J Med Chem 65(15):10691–10706

    Article  CAS  PubMed  Google Scholar 

  12. Cao D, Chen G, Jiang J, Yu J, Zhang R, Chen M, Zhang W, Chen L, Zhong F, Zhang Y et al (2024) Generic protein-ligand interaction scoring by integrating physical prior knowledge and data augmentation modelling. Nat Mach Intell. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s42256-024-00849-z

    Article  Google Scholar 

  13. Nguyen DD, Xiao T, Wang M, Wei G-W (2017) Rigidity strengthening: a mechanism for protein-ligand binding. J Chem Inf Model 57(7):1715–1721

    Article  CAS  PubMed  Google Scholar 

  14. Rana MM, Nguyen DD (2023) Geometric graph learning with extended atom-types features for protein-ligand binding affinity prediction. arXiv preprint arXiv:2301.06194

  15. Nguyen DD, Wei G-W (2019) Agl-score: algebraic graph learning score for protein-ligand binding scoring, ranking, docking, and screening. J Chem Inf Model 59(7):3291–3304

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  16. Chen D, Gao K, Nguyen DD, Chen X, Jiang Y, Wei G-W, Pan F (2021) Algebraic graph-assisted bidirectional transformers for molecular property prediction. Nat Commun 12(1):1–9

    Google Scholar 

  17. Wang R, Nguyen DD, Wei G-W (2020) Persistent spectral graph. Int J Numer Methods Biomed Eng 36(9):3376

    Article  Google Scholar 

  18. Meng Z, Xia K (2021) Persistent spectral-based machine learning (perspect ml) for protein-ligand binding affinity prediction. Sci Adv 7(19):5329

    Article  Google Scholar 

  19. Stam CJ, Reijneveld JC (2007) Graph theoretical analysis of complex networks in the brain. Nonlinear Biomed Phys 1:1–19

    Article  Google Scholar 

  20. Kocbek S, Kim J-D (2017) Exploring biomedical ontology mappings with graph theory methods. PeerJ 5:2990

    Article  Google Scholar 

  21. Pavlopoulos GA, Secrier M, Moschopoulos CN, Soldatos TG, Kossida S, Aerts J, Schneider R, Bagos PG (2011) Using graph theory to analyze biological networks. BioData Min 4:1–27

    Article  Google Scholar 

  22. Trinajstic N (2018) Chemical graph theory. Routledge, Oxfordshire

    Book  Google Scholar 

  23. Schultz HP (1989) Topological organic chemistry. 1. graph theory and topological indices of alkanes. J Chem Inf Comput Sci 29(3):227–228

    Article  CAS  Google Scholar 

  24. Janezic D, Milicevic A, Nikolic S, Trinajstic N (2015) Graph-theoretical matrices in chemistry. CRC Press, Boca Raton

    Book  Google Scholar 

  25. Angeleska A, Jonoska N, Saito M (2009) Dna recombination through assembly graphs. Discret Appl Math 157(14):3020–3037

    Article  Google Scholar 

  26. Basak SC, Niemi GJ, Veith GD (1990) A graph-theoretic approach to predicting molecular properties. Math Comput Model 14:511–516

    Article  Google Scholar 

  27. Randić M, Wilkins CL (1980) Graph-theoretical analysis of molecular properties. isomeric variations in nonanes. Int J Quantum Chem 18(4):1005–1027

    Article  Google Scholar 

  28. Gramatica R, Di Matteo T, Giorgetti S, Barbiani M, Bevec D, Aste T (2014) Graph theory enables drug repurposing-how a mathematical model can drive the discovery of hidden mechanisms of action. PloS ONE 9(1):84912

    Article  Google Scholar 

  29. Huang C-H, Chang PM-H, Hsu C-W, Huang C-YF, Ng K-L (2016) Drug repositioning for non-small cell lung cancer by using machine learning algorithms and topological graph theory. BMC Bioinform 17:13–26

    Article  Google Scholar 

  30. Huang N, Jacobson MP (2007) Physics-based methods for studying protein-ligand interactions. Curr Opin Drug Discov Dev 10(3):325

    CAS  Google Scholar 

  31. Beard H, Cholleti A, Pearlman D, Sherman W, Loving KA (2013) Applying physics-based scoring to calculate free energies of binding for single amino acid mutations in protein-protein complexes. PloS ONE 8(12):82849

    Article  Google Scholar 

  32. Huang N, Kalyanaraman C, Irwin JJ, Jacobson MP (2006) Physics-based scoring of protein- ligand complexes: enrichment of known inhibitors in large-scale virtual screening. J Chem Inf Model 46(1):243–253

    Article  CAS  PubMed  Google Scholar 

  33. Huang N, Kalyanaraman C, Bernacki K, Jacobson MP (2006) Molecular mechanics methods for predicting protein-ligand binding. Phys Chem Chem Phys 8(44):5166–5177

    Article  CAS  PubMed  Google Scholar 

  34. Raha K, Merz KM (2005) Large-scale validation of a quantum mechanics based scoring function: predicting the binding affinity and the binding mode of a diverse set of protein- ligand complexes. J Med Chem 48(14):4558–4575

    Article  CAS  PubMed  Google Scholar 

  35. Zheng Z, Merz KM Jr (2011) Ligand identification scoring algorithm (lisa). J Chem Inf Model 51(6):1296–1306

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  36. Verkhivker G, Appelt K, Freer S, Villafranca J (1995) Empirical free energy calculations of ligand-protein crystallographic complexes. i. knowledge-based ligand-protein interaction potentials applied to the prediction of human immunodeficiency virus 1 protease binding affinity. Prot Eng Des Sel 8(7):677–691

    Article  CAS  Google Scholar 

  37. Eldridge MD, Murray CW, Auton TR, Paolini GV, Mee RP (1997) Empirical scoring functions: I. the development of a fast empirical scoring function to estimate the binding affinity of ligands in receptor complexes. J Comput Aided Mol Des 11:425–445

    Article  CAS  PubMed  Google Scholar 

  38. Muegge I, Martin YC (1999) A general and fast scoring function for protein- ligand interactions: a simplified potential approach. J Med Chem 42(5):791–804

    Article  CAS  PubMed  Google Scholar 

  39. Velec HF, Gohlke H, Klebe G (2005) Drugscorecsd knowledge-based scoring function derived from small molecule crystal data with superior recognition rate of near-native ligand poses and better affinity prediction. J Med Chem 48(20):6296–6303

    Article  CAS  PubMed  Google Scholar 

  40. Ballester PJ, Mitchell JB (2010) A machine learning approach to predicting protein-ligand binding affinity with applications to molecular docking. Bioinformatics 26(9):1169–1175

    Article  CAS  PubMed  Google Scholar 

  41. Kinnings SL, Liu N, Tonge PJ, Jackson RM, Xie L, Bourne PE (2011) A machine learning-based method to improve docking scoring functions and its application to drug repurposing. J Chem Inf Model 51(2):408–419

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  42. Li Y, Yang J (2017) Structural and sequence similarity makes a significant impact on machine-learning-based scoring functions for protein-ligand interactions. J Chem Inf Model 57(4):1007–1012

    Article  CAS  PubMed  Google Scholar 

  43. Li H, Peng J, Leung Y, Leung K-S, Wong M-H, Lu G, Ballester PJ (2018) The impact of protein structure and sequence similarity on the accuracy of machine-learning scoring functions for binding affinity prediction. Biomolecules 8(1):12

    Article  PubMed  PubMed Central  Google Scholar 

  44. Li H, Peng J, Sidorov P, Leung Y, Leung K-S, Wong M-H, Lu G, Ballester PJ (2019) Classical scoring functions for docking are unable to exploit large volumes of structural and interaction data. Bioinformatics 35(20):3989–3995

    Article  CAS  PubMed  Google Scholar 

  45. Su M, Feng G, Liu Z, Li Y, Wang R (2020) Tapping on the black box: how is the scoring power of a machine-learning scoring function dependent on the training set? J Chem Inf Model 60(3):1122–1136

    Article  CAS  PubMed  Google Scholar 

  46. Ellingson SR, Davis B, Allen J (2020) Machine learning and ligand binding predictions: a review of data, methods, and obstacles. Biochimica et Biophysica Acta (BBA) Gen Sub 1864(6):129545

    Article  CAS  Google Scholar 

  47. Davis B, Mcloughlin K, Allen J, Ellingson S.R (2020) Quantifying overfitting potential in drug binding datasets. In: Computational Science–ICCS 2020: 20th International Conference, Amsterdam, The Netherlands, June 3–5, 2020, Proceedings, Part III 20, pp. 585–598 . Springer

  48. Wang B, Zhao Z, Nguyen DD, Wei G-W (2017) Feature functional theory-binding predictor (fft-bp) for the blind prediction of binding free energies. Theor Chem Acc 136(4):1–22

    Article  Google Scholar 

  49. Cang Z, Wei G-W (2017) Topologynet: topology based deep convolutional and multi-task neural networks for biomolecular property predictions. PLoS Comput Biol 13(7):1005690

    Article  Google Scholar 

  50. Bramer D, Wei G-W (2018) Multiscale weighted colored graphs for protein flexibility and rigidity analysis. J Chem Phys 148(5):054103

    Article  PubMed  Google Scholar 

  51. Neudert G, Klebe G (2011) fconv: format conversion, manipulation and feature computation of molecular data. Bioinformatics 27(7):1021–1022

    Article  CAS  PubMed  Google Scholar 

  52. Moon S, Zhung W, Yang S, Lim J, Kim WY (2022) Pignet: a physics-informed deep learning model toward generalized drug-target interaction predictions. Chem Sci 13(13):3661–3673

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  53. Sánchez-Cruz N, Medina-Franco JL, Mestres J, Barril X (2021) Extended connectivity interaction features: improving binding affinity prediction through chemical description. Bioinformatics 37(10):1376–1382

    Article  PubMed  Google Scholar 

  54. Nguyen DD, Wei G-W (2019) DG-GL: differential geometry-based geometric learning of molecular datasets. Int J Numer Methods Biomed Eng 35(3):3179

    Article  Google Scholar 

  55. Wójcikowski M, Kukiełka M, Stepniewska-Dziubinska MM, Siedlecki P (2019) Development of a protein-ligand extended connectivity (plec) fingerprint and its application for binding affinity predictions. Bioinformatics 35(8):1334–1341

    Article  PubMed  Google Scholar 

  56. Jiménez J, Skalic M, Martínez-Rosell G, De Fabritiis G (2018) K deep: Protein-ligand absolute binding affinity prediction via 3d-convolutional neural networks. J Chem Inf Model 58(2):287–296

    Article  PubMed  Google Scholar 

  57. Stepniewska-Dziubinska MM, Zielenkiewicz P, Siedlecki P (2018) Development and evaluation of a deep learning model for protein-ligand binding affinity prediction. Bioinformatics 1:9

    Google Scholar 

  58. Cang Z, Mu L, Wei G-W (2018) Representability of algebraic topology for biomolecules in machine learning based scoring and virtual screening. PLoS Comput Biol 14(1):1005929

    Article  Google Scholar 

  59. Wang C, Zhang Y (2017) Improving scoring-docking-screening powers of protein-ligand scoring functions using random forest. J Comput Chem 38(3):169–177

    Article  PubMed  Google Scholar 

  60. Li H, Leung K-S, Wong M-H, Ballester PJ (2015) Improving autodock vina using random forest: the growing accuracy of binding affinity prediction by the effective exploitation of larger data sets. Mol Inform 34(2–3):115–126

    Article  PubMed  Google Scholar 

  61. Rana MM, Nguyen DD (2022) Eisa-score: element interactive surface area score for protein-ligand binding affinity prediction. J Chem Inf Model 62(18):4329–4341

    Article  CAS  PubMed  Google Scholar 

  62. Ballester PJ, Schreyer A, Blundell TL (2014) Does a more precise chemical description of protein-ligand complexes lead to more accurate prediction of binding affinity? J Chem Inf Model 54(3):944–955

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  63. Liu Z, Li Y, Han L, Li J, Liu J, Zhao Z, Nie W, Liu Y, Wang R (2015) Pdb-wide collection of binding data: current status of the pdbbind database. Bioinformatics 31(3):405–412

    Article  CAS  PubMed  Google Scholar 

  64. Gaieb Z, Liu S, Gathiaka S, Chiu M, Yang H, Shao C, Feher VA, Walters WP, Kuhn B, Rudolph MG (2018) D3r grand challenge 2: blind prediction of protein-ligand poses, affinity rankings, and relative binding free energies. J Comput Aided Mol Des 32(1):1–20

    Article  CAS  PubMed  Google Scholar 

  65. Parks CD, Gaieb Z, Chiu M, Yang H, Shao C, Walters WP, Jansen JM, McGaughey G, Lewis RA, Bembenek SD (2020) D3r grand challenge 4: blind prediction of protein-ligand poses, affinity rankings, and relative binding free energies. J Comput Aided Mol Des 34(2):99–119

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  66. Li Y, Han L, Liu Z, Wang R (2014) Comparative assessment of scoring functions on an updated benchmark: 2. evaluation methods and general results. J Chem Inf Model 54(6):1717–1736

    Article  CAS  PubMed  Google Scholar 

  67. Su M, Yang Q, Du Y, Feng G, Liu Z, Li Y, Wang R (2018) Comparative assessment of scoring functions: the casf-2016 update. J Chem Inf Model 59(2):895–913

    Article  PubMed  Google Scholar 

  68. Li H, Lu G, Sze K-H, Su X, Chan W-Y, Leung K-S (2021) Machine-learning scoring functions trained on complexes dissimilar to the test set already outperform classical counterparts on a blind benchmark. Br Bioinform 22(6):225

    Article  Google Scholar 

  69. Ain QU, Aleksandrova A, Roessler FD, Ballester PJ (2015) Machine-learning scoring functions to improve structure-based binding affinity prediction and virtual screening. Wiley Interdiscip Rev Comput Mol Sci 5(6):405–424

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  70. Durant G, Boyles F, Birchall K, Marsden B, Deane C (2023) Robustly interrogating machine learning based scoring functions: what are they learning? bioRxiv. https://doiorg.publicaciones.saludcastillayleon.es/10.1101/2023.10.30.564251

    Article  Google Scholar 

  71. Shen C, Hu Y, Wang Z, Zhang X, Zhong H, Wang G, Yao X, Xu L, Cao D, Hou T (2021) Can machine learning consistently improve the scoring power of classical scoring functions? insights into the role of machine learning in scoring functions. Br Bioinform 22(1):497–514

    Article  Google Scholar 

  72. Tran-Nguyen V-K, Junaid M, Simeon S, Ballester PJ (2023) A practical guide to machine-learning scoring for structure-based virtual screening. Nat Protoc 18(11):3460–3511

    Article  CAS  PubMed  Google Scholar 

  73. Pearson WR, Lipman DJ (1988) Improved tools for biological sequence comparison. Proc Natl Acad Sci 85(8):2444–2448

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  74. OpenEye Scientific Software: ROCS 3.5.1.1. http://www.eyesopen.com. OpenEye Scientific Software, Santa Fe, NM (2023)

  75. Hawkins PC, Skillman AG, Nicholls A (2007) Comparison of shape-matching and docking as virtual screening tools. J Med Chem 50(1):74–82

    Article  CAS  PubMed  Google Scholar 

  76. Yeturu K, Chandra N (2008) Pocketmatch: a new algorithm to compare binding sites in protein structures. BMC Bioinform 9:1–17

    Article  Google Scholar 

  77. Nagarajan D, Chandra N (2013) Pocketmatch (version 2.0): A parallel algorithm for the detection of structural similarities between protein ligand binding-sites. In: 2013 National Conference on Parallel Computing Technologies (PARCOMPTECH), pp. 1–6 . IEEE

Download references

Acknowledgements

The authors thank the anonymous reviewers for their valuable suggestions. This work is supported in part by funds from the National Science Foundation (NSF: # 2053284, # 2151802, and # 2245903), the University of Kentucky Startup Fund, and Markey Cancer Research Informatics Shared Resource Facility (P30 CA177558). We would also like to thank OpenEye Scientific Software for providing ROCS software.

Author information

Authors and Affiliations

Authors

Contributions

F.T.M. and M.M.R. contributed equally to data curation, software development, conducting experiments, and writing the initial draft. A.M. verified the experiments. D.D.N. initiated the conceptualization, designed the research, secured funding, and reviewed and edited the paper. S.E. actively engaged in discussions and thoroughly reviewed the manuscript.

Corresponding author

Correspondence to Duc D. Nguyen.

Ethics declarations

Competing interests

The authors declare no Competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Mukta, F.T., Rana, M.M., Meyer, A. et al. The algebraic extended atom-type graph-based model for precise ligand–receptor binding affinity prediction. J Cheminform 17, 10 (2025). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s13321-025-00955-z

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s13321-025-00955-z

Keywords