Preprint
Article

An Optimized Deep Neural Network Framework for Classification of Drug–Drug Interactions

Altmetrics

Downloads

135

Views

90

Comments

0

Submitted:

10 June 2024

Posted:

12 June 2024

You are already at the latest version

Alerts
Abstract
Accurately predicting drug-drug interactions (DDIs) is crucial for preventing adverse drug events in clinical settings. However, existing methodologies often rely on complex models built from diverse data sources, posing challenges in computational drug discovery. To address the need for precise computational methods in predicting unknown DDIs, this study introduces a novel Deep Neural Network (DNN)-based approach. By leveraging a wide range of drug-related information, including substructure, targets, side effects, pathways, and indications, our method calculates multiple drug similarities. These similarities are then synthesized using a nonlinear fusion method to extract high-level features. Subsequently, a tailored neural network is deployed for interaction prediction. Comparative evaluation against three prominent machine learning classifiers—Extreme Gradient Boosting (XGBoost), Adaptive Gradient Boosting (AdaBoost), and Light Gradient-Boosting Machine (LGBM)—using three benchmark datasets demonstrates the superior performance of DNN. It achieves outstanding accuracy, precision, recall, and F1-score metrics, all reaching 94.6% in cross-validation. Additionally, case studies involving numerous drug pairs confirm the reliability of DNN in accurately predicting unknown DDIs. These findings underscore DNN as a potent and reliable method for DDI prediction, with promising implications for drug discovery and healthcare applications.
Keywords: 
Subject: Biology and Life Sciences  -   Biochemistry and Molecular Biology

1. Introduction

Polypharmacy, the administration of multiple drugs to treat complex diseases, is a common therapeutic approach [1]. However, it introduces the risk of drug-drug interactions (DDIs), which can alter the pharmacological effects of medications [2,3]. Studies indicate that DDIs contribute to approximately 30% of reported adverse drug events (ADEs), leading to increased incidence rates, mortality, and occasionally, the withdrawal of drugs from the market. This results in significant medical costs due to the stringent demands of drug development [4]. Therefore, the accurate identification and understanding of DDIs are essential for guiding polypharmacy prescriptions by clinicians and patients, as well as for informing drug development efforts in pharmaceutical companies. Despite the availability of in-vitro experiments and clinical trials to identify DDIs, the systematic screening of potential DDI candidates from a vast array of drugs remains challenging, time-consuming, and resource-intensive.
Accurate prediction and understanding of drug-drug interactions (DDIs) are pivotal for healthcare and pharmaceutical sectors due to their substantial impact on adverse drug events (ADEs) and associated costs. Despite the abundance of biomedical data, there exists a pressing need for computational methods to effectively navigate the extensive landscape of potential DDIs. While repositories such as DrugBank offer comprehensive repositories of documented DDIs, they represent only a fraction of the myriad of possible drug combinations. Consequently, numerous computational approaches have emerged in recent years to address this gap [5].
In recent years, there has been a significant surge in the availability of scientific literature, electronic medical records, adverse event reports, drug labels, and related data sources [6]. Researchers have actively pursued methods to extract information about drug-drug interactions (DDIs) from diverse textual sources and medical records using advanced natural language processing (NLP) techniques [7]. Additionally, efforts have been made to infer potential DDIs through similarity-based approaches, leveraging known interactions [4,6,7]. Moreover, various computational techniques including machine learning [8], network modeling [6], and knowledge graphs [2] have been applied to predict DDIs. However, many of these computational methods predominantly focus on identifying the occurrence of a DDI given a specific drug pair.
In recent years, there has been a significant surge in efforts to deepen our understanding of drug-drug interactions (DDIs) through the development of advanced multi-type DDI prediction methodologies. These methodologies aim to provide more comprehensive insights beyond simple predictions of DDI occurrences [7]. A crucial development in this field was the establishment of a gold standard DDI dataset sourced from DrugBank. This dataset encompasses 192,284 DDIs associated with 86 DDI types across 191,878 drug pairs [5]. It served as the foundation for multi-type DDI prediction tasks, formulated as multi-label classification challenges.
A noteworthy approach is the Deep-DDI method, which introduced deep neural networks (DNNs) to harness structural information from chemical compounds for each drug pair. Subsequent advancements have focused on enhancing prediction accuracy by incorporating diverse biological information, such as drug targets and enzymes, to characterize drug pairs. Additionally, these methods integrate structural information obtained from autoencoders or the encoder module of transformers to learn low-dimensional latent features, in conjunction with DNN algorithms for classification [8,9,10].
In recent research endeavors, numerous methods have emerged aiming to enhance the prediction accuracy and depth of understanding of drug-drug interactions (DDIs). One notable approach involves representing the feature vector of a drug through a similarity profile, which assesses the structural likeness of a given drug against others in the dataset. Moreover, recent advancements include the utilization of few-shot learning techniques, as demonstrated by Deng et al. [11], to improve prediction performance for rare types of DDIs with limited samples [12,13]. Additionally, the CSMDDI method introduced by Liu et al. [14] generates embedding representations of drugs and DDI types, followed by learning a mapping function to predict multi-type DDIs [15]. Another notable contribution is the deep MDDI model proposed by Feng et al. [16], which integrates an encoder utilizing deep relational graph convolutional networks and a tensor-like decoder to capture topological features and predict multi-label DDI types. Furthermore, Yang et al. [17] developed a substructure-aware graph neural network, employing a message-passing neural network with a novel substructure attention mechanism and a substructure-substructure interaction module for DDI prediction. These methodologies collectively contribute to advancing the field of DDI prediction and hold promise for improving drug safety and efficacy.
Some early approaches, such as those by Vilar et al. [18] and Lu et al. [19], focused solely on known DDIs, while newer similarity-based methods, like those by Vilar et al. [18] and Zhang et al. [20], utilized drug similarities to predict potential DDIs. Ensemble methods, introduced by Gottlieb et al. [21] and Zhang et al. [20], combined various drug similarity measures to improve prediction accuracy. Additionally, machine learning models like Dependency-based Convolutional Neural Network (DCNN) by Liu et al. and Deep DDI by Ryu et al. utilized text mining and deep learning techniques to predict DDIs based on drug structures and names [22]. A novel approach in this field is network pharmacology, exemplified by methods like NIMS proposed by Li et al. [14], which constructs disease-specific biological networks to identify synergistic drug interactions based on therapeutic targets. This approach has shown promise, as demonstrated by Guo et al. [23], who used synergistic modules in a differential gene interaction network to effectively inhibit Inflammation-Induced Tumorigenesis (IIT) [24].
In our pursuit of enhancing drug-drug interaction (DDI) prediction accuracy, we adopt a comprehensive approach that integrates similarity selection, fusion, and neural network techniques. This methodology aims to elevate the precision of DDI prediction, thus contributing to advancements in drug discovery and healthcare. Our similarity selection process draws from a heuristic method proposed by Olayan et al. [25], originally designed for drug-target interaction prediction, allowing us to focus on high-information subsets of the data. Additionally, we leverage Similarity Network Fusion (SNF), a well-established technique in biological research, to amalgamate diverse data types and bolster prediction accuracy.
Our method, termed DNN, combines a neural network model with techniques for similarity selection and fusion to enhance the precision of DDI prediction through nonlinear analysis and feature extraction. DNN operates through a structured process: initially, it aggregates information on diverse drug similarities, including chemical, target-based, and Gene Ontology similarities, sourced from multiple datasets. Subsequently, a heuristic process proposed by Olayan et al. [25] is applied to identify the most informative and least redundant subset of similarity types. This refined subset is then integrated using the Similarity Network Fusion (SNF) method. Following this, the amalgamated similarity matrix, alongside interaction data, serves as the input for training the neural network. Our assessment of DNN against three prominent machine learning classifiers—Extreme Gradient Boosting (XGBoost), Adaptive Gradient Boosting (AdaBoost), and Light Gradient-Boosting Machine (LGBM)—across three benchmark datasets involved stratified five-fold cross-validation [26]. Additionally, case studies conducted on various drugs underscore DNN's efficacy in predicting unknown DDIs. These findings validate the robustness and efficiency of DNN as an approach for precise DDI prediction [27,28,29].

2. Materials and Methods

2.1. Datasets

In our investigation, we utilize two datasets with varying sample scales. The first dataset, referred to as Dataset1, is a benchmark dataset assembled by Deng et al. [11]. It encompasses 445,390 drugs and 209,904 pairwise DDIs across 20 distinct DDI types. Each drug entry in Dataset1 is characterized by four fundamental types of features extracted from DrugBank [30], namely chemical substructures, targets, pathways, and enzymes.
Conversely, Dataset2, sourced from the research conducted by Lin et al. [31], comprises 222,695 drugs with 209,904 pairwise DDIs spanning the same 20 DDI types as Dataset1. However, Dataset2 offers a more extensive array of features for each drug entry. These features encompass various aspects such as QTc-prolonging activities, serum concentration, absorption, neural activity, stimulatory activities, hyperglycemic and hyponatremic activities, arrhythmia, adverse effects, metabolism, bronchoconstrictory and bronchodilatory activities, toxicity, central nervous system depressant activities, excretion, myopathy, therapeutic efficacy, myelosuppression, vasoconstricting and vasodilatory activities, potassium, calcium, and sodium concentration, as well as hypertensive and hypotensive activities, sedative activities, and miscellaneous interactions.
For situations involving multiple drugs, we assign a unique identifier to each situation as a DDI type and aggregate them based on features extracted from other datasets. This enables the application of machine learning classifier models. To delve into the specifics of DDI types, Figure 1 illustrates the correlation between different drug types in multidrug interactions, as shown in Figure 2, highlighting a range of highly and less correlated DDI types. To enhance precision in subsequent models, we introduce an innovative approach by combining positively correlated (Figure 3) and negatively correlated (Figure 4) DDI types. This combination enriches the feature set, facilitating more efficient feature representation.
In the DDI dataset, the five most frequently occurring drugs in the "Drug I " column are Fluvoxamine, Venlafaxine, Phenytoin, Ziprasidone, and Cyclosporine. Similarly, in the "Drug II" column, the top five drugs by frequency are Curcumin, Pitolisant, Rucaparib, Stiripentol, and Apalutamide. Regarding the interaction types represented in the "DDI’s types" column, the top five in terms of frequency are 'risk or severity of adverse effects', 'metabolism', 'serum concentration', 'therapeutic efficacy', and 'hypotensive activities'. These are provided valuable insights into the most prevalent drugs and interaction types present in the dataset, which can inform further analysis and modeling efforts in the study of DDI.

2.2. Dataset Splitting

For each dataset, we carefully selected a metric and devised a splitting pattern that best suits its unique characteristics. Given the diverse nature of our dataset's input parameters, which include both categorical factors like material and process types, as well as sub-categorical processes, along with numerical inputs, we performed scaling. Moreover, to ensure consistency, we normalized the output classes - DDI’s types. The normalization process follows the equation:
x = x m i n ( x ) max x m i n ( x )
Where, x is the scaled value, x is the original value, m i n   ( x ) is the minimum value in the dataset, and max x is the maximum value in the dataset.
In machine learning, it's imperative to split datasets into training and test subsets to gauge model performance on unseen data. Consequently, we divided our dataset into training and testing sets. The models were trained using the training data, while the test data remained separate for evaluating model performance. This division was executed post scaling of the datasets, as depicted in Figure 5.
To ensure a comprehensive evaluation of our model's performance and enhance its reliability, we integrated k-fold cross-validation into our methodology. This widely-adopted technique involves dividing the dataset into k equally-sized folds or subsets. The model is then trained on k-1 folds and validated on the remaining fold. This iterative process is repeated k times, with each fold serving as the validation set exactly once.
In addition to k-fold cross-validation, we implemented supplementary measures to validate the robustness of our model. These measures encompassed bootstrap resampling techniques and sensitivity analyses to assess the stability of our findings under diverse conditions. Such comprehensive validation approaches are crucial for establishing the credibility and generalizability of our predictive models in real-world scenarios.
The primary objective of employing cross-validation is to mitigate overfitting, wherein the model becomes excessively tailored to the training data, leading to poor generalization to unseen data. By evaluating the model across multiple validation sets, cross-validation provides a more accurate estimate of its ability to generalize. However, it's important to acknowledge that this approach may introduce increased variability in model testing, particularly when assessing against individual data points. Outliers within these data points can significantly impact testing variability.
Furthermore, k-fold cross-validation can be computationally intensive due to its iteration over the dataset's size. To address this concern and minimize overfitting, we selected k=11, as it demonstrated highly effective performance with minimal errors and overfitting, as illustrated in Figure 6.

2.3. ML Models

In the context of classifying DDI types in multidrug interactions, the importance of time and cost considerations is paramount, as highlighted in the introduction. Achieving precise outcomes while minimizing expenses, without resorting to clinical tests, underscores the significance of exploring machine learning (ML) models as feasible alternatives to conventional experimental and numerical methods. This study delves into the performance of various ML models using pre-segmented datasets, focusing on non-linear algorithms. Below, we provide a brief overview of these models, followed by an in-depth discussion on effective metric methodologies in the subsequent section. The results and discussions section will then unveil the outcomes of these models, followed by a comparative analysis in the metrics section to discern their respective efficiencies.

2.3.1. Non-Linear Classification Algorithm

In situations where classes cannot be distinctly separated by a linear boundary, the utilization of non-linear classifiers becomes imperative in machine learning models. These classifiers excel in addressing intricate classification challenges by discerning complex patterns and relationships within the data. Unlike linear models, non-linear classifiers demonstrate superior performance when dealing with complex datasets. Here, we introduce the most prevalent models applied to our DDI type classes, including tree-based algorithms and Neural Networks (NNs) algorithms, offering insights into the capabilities of non-linear algorithms.

2.3.1.1. Gradient Boosting Trees (GBT)

Gradient Boosting Trees (GBT) stands out as a robust tree-based algorithm that iteratively enhances a loss function by incorporating decision trees into the ensemble. With each iteration, a new tree is trained to forecast the residuals, representing the disparities between the actual and predicted values, of the prior trees. These forecasts are merged to derive the ultimate prediction. GBT demonstrates proficiency in managing intricate data relationships and is esteemed for its ability to yield remarkably precise predictions. It particularly thrives in scenarios where conventional machine learning algorithms encounter difficulties, such as navigating through noisy or high-dimensional datasets.

2.3.1.1.1. Extreme Gradient Boosting Machine (XGBM)

Extreme Gradient Boosting Machine (XGBoost) stands out as a refined version of the gradient boosting algorithm, celebrated for its exceptional speed and effectiveness. Crafted to enhance the gradient boosting process, XGBoost incorporates features such as parallel computing, regularization, and tree pruning, all aimed at boosting accuracy and efficiency.

2.3.1.1.2. Light Gradient Boosting Machine (LGBM)

The Light Gradient Boosting Machine (LGBM) mirrors XGBoost as a gradient boosting framework, esteemed for its speed, efficiency, and scalability. Particularly adept at managing large-scale datasets, LGBM has garnered widespread acclaim across diverse machine learning endeavors due to its stellar performance. Noteworthy for its swift training pace and efficacy, LGBM owes its attributes to its leaf-wise tree growth strategy and optimizations like Gradient-Based One-Side Sampling (GOSS) and Exclusive Feature Bundling (EFB). In contrast, XGBoost defaults to a depth-wise tree growth strategy, potentially leading to slower performance, especially with extensive datasets.

2.3.1.1.3. Adaptive Gradient Boosting Machine (AdaBoost)

AdaBoost, an abbreviation for Adaptive Boosting, stands as an ensemble learning technique crafting a robust classifier by amalgamating numerous weak classifiers. It operates by iteratively training a sequence of weak learners on weighted iterations of the training dataset. With each iteration, the algorithm prioritizes instances previously misclassified, thereby adaptively refining its strategy to enhance overall performance.
AdaBoost assigns a weight α t to each weak learner h t ( x ) , where t represents the iteration number. The final prediction H x is then obtained by summing the weighted predictions of all weak learners:
H x = s i g n ( t = 1 T α t h t ( x ) )
Here, T denotes the total number of weak learners. The sign function ensures that the final prediction is either +1 or -1, depending on the overall weighted sum of the weak learners' predictions.

2.3.1.2. Neural Networks (NNs)

Neural Networks (NNs) are versatile machine learning algorithms capable of handling a wide range of tasks, including regression and classification [32]. These networks are composed of individual neurons that process input data through both linear and nonlinear transformations. Through iterative backpropagation, the outputs generated by these neurons are refined, with adjustments made to the weights and biases to optimize the model's performance.

2.3.1.2.1. Deep Neural Networks (DNN)

Deep Neural Networks (DNNs) are characterized by their layered architecture, featuring interconnected neurons. Typically, these networks comprise an input layer, one or more hidden layers, and an output layer. Neurons within the network process input data via weighted connections and activation functions to produce output. DNNs are widely utilized across various machine learning tasks, including classification, regression, and pattern recognition, owing to their ability to capture complex data relationships. The behavior of a node in a DNN is governed by the following general equation:
a i = f j = 1 n w i j . x j + b i
Where, a i is the output of the i -th node in the layer, f is the activation function applied element-wise, w i j is the weight connecting the j -th input to the i -th node, x j is the j -th input to the node, b i is the i -th bias term for the node, and n is the number of inputs to the node.

2.4. Evaluation Metrics

To assess the performance of our machine learning models on new data, we implemented 11-fold cross-validation after randomizing our datasets. This method involves dividing the data into 11 subsets and conducting 11 rounds of training and validation, where each subset serves as the validation set once while the rest act as the training set. The overall accuracy of our models is then determined by averaging the accuracies obtained across all eleven iterations.
Our dataset focuses on a classification task aimed at identifying DDI’s types in multidrug impact. We employed various evaluation metrics, including accuracy, precision (micro, macro, and weighted), recall (micro, macro, and weighted), and F1 score (micro, macro, and weighted). These metrics offer comprehensive insights into the model's performance, considering factors such as class imbalance and overall effectiveness across all classes. Additionally, to visualize the learning progress of our models, we utilize learning curves. These curves depict how specific metrics, such as accuracy or loss, change throughout the model training process, providing insights into the model's performance over successive iterations.

3. Results and Discussion

In our classification task, the use of machine learning (ML) is crucial for predicting outcomes in large datasets efficiently, yielding precise results in less time and without the expenses associated with clinical investigations. To focus our classification efforts, we initially select the five most important classes of Drug-Drug Interactions (DDIs). These classes are pivotal in DDI research, and handling the high-volume datasets requires robust CPU or GPU systems.
The first dataset comprises Drug I, including ‘Fluvoxamine’, ‘Venlafaxine’, ‘Phenytoin’, ‘Ziprasidone’, ‘Cyclosporine’, ‘Pentobarbital’, ‘Bortezomib’, ‘Citalopram’, ‘Ticlopidine’, ‘Isradipine’, ‘Diltiazem’, ‘Sildenafil’, ‘Ranolazine’, ‘Clozapine’, ‘Indinavir’. The second dataset, Drug II, consists of ‘Curcumin’, ‘Pitolisant’, ‘Rucaparib’, ‘Stiripentol’, ‘Apalutamide’, ‘Vemurafenib’, ‘Lumacaftor’, ‘Ceritinib’, ‘Isavuconazole’, ‘Acetyl sulfisoxazole’, ‘Dabrafenib’, ‘Cobicistat’, ‘Asenapine’, ‘Lorpiprazole’, ‘Fosphenytoin’. These fifteen classes encompass ‘risk or severity of adverse effects’, ‘metabolism’, ‘serum concentration’, ‘therapeutic efficacy’, ‘hypotensive activities’, ‘QTc-prolonging activities’, ‘antihypertensive activities’, ‘central nervous system depressant (CNS depressant) activities’, ‘anticoagulant activities’, ‘vasoconstricting activities’, ‘hypoglycemic activities’, ‘risk or severity of hypotension’, ‘sedative activities’, ‘hypokalemic activities’, ‘absorption’. We utilize DNN, XGBoost, LGBM, and AdaBoost models to investigate and classify these classes, as depicted in Figure 1.
In our study, datasets were gathered from diverse sources, prioritizing specific features and parameters deemed crucial for our investigation. These datasets were consolidated into five columns, each representing an identifier associated with a set of DDI-related attributes, such as side effects and drug types, along with the types of DDIs. In our SQL database, sourcing datasets based on meaningful correlations and relevant features proved challenging. Despite efforts to collect data from reliable sources and recent studies, the complexity of feature correlations posed a significant obstacle. This complexity extends to machine learning models, making it challenging to optimize weights and biases for each node or branch in tree-based models.
Navigating the challenges of overfitting and underfitting adds complexity to our analysis. To mitigate these issues and ensure the reliability of our results, we employ various techniques such as comparing training-validation datasets, testing metrics like confusion matrices, and analyzing learning curves. These approaches offer insights into the efficiency of our model across a logical range of overfitting scenarios.
In scrutinizing the performance of our model in classifying DDI types, we begin with an assessment of the XGBoost model. By exploring different combinations of hyperparameters such as learning rate and min child weight, we can monitor both model performance and overfitting tendencies. Figure 7 illustrates that a higher min child weight beyond 6 does not significantly enhance accuracy. Additionally, a learning rate within the range of 0.95 demonstrates optimal performance for the XGBoost model.
The evaluation process continues with testing and assessing each group of labels using confusion matrices. Labeling risk or severity of adverse effects as label 0, metabolism as label 1, serum concentration as label 2, therapeutic efficacy as label 3, hypotensive activities as label 4, QTc-prolonging activities as label 5, antihypertensive activities as label 6, central nervous system depressant (CNS depressant) activities as label 7, anticoagulant activities as label 8, vasoconstricting activities as label 9, hypoglycemic activities as label 10, risk or severity of hypotension as label 11, sedative activities as label 12, hypokalemic activities as label 13, absorption as label 14, as depicted in Figure 8, the model achieves an accuracy of 85.8%, precision of 98%, recall of 95%, and an F1-score of 96%. This demonstrates the model's high performance, effectively balancing overfitting concerns without encountering underfitting issues, essential for real-world DDI recognition testing.
Figure 9 showcases the LGBM model's performance. Optimizing parameters such as a higher min child weight and learning rate efficiently enhances model accuracy, reducing the disparity between training and validation curves and indicating decreased overfitting. Test evaluations confirm the model's effectiveness, with metrics including accuracy, precision, recall, and F1-score all attaining 86%, as demonstrated in the confusion matrix in Figure 10, for the situation when 0.3 as learning rate and min child weight as 2. It also shows that for a more than an optimum point forementioned, model performance adversely moved down. The key point for this monitoring procedure is to avoid overfitting as the gap between the training and validation is so neglectable.
Advancing to the examination of AdaBoost in Figure 11, alterations in parameters, such as elevating the min child weight and learning rate, do not notably augment model performance across all samples. Nevertheless, exceeding this threshold, additional increments in samples do not result in significant enhancements, indicating diminishing returns. The evaluation metrics, depicted in the confusion matrix in Figure 12, validate the model's adeptness in classifying DDI types, with notable test evaluations affirming its effectiveness. Metrics encompassing accuracy, precision, recall, and F1-score all reach 46%.
Considering the complex features inherent in DDI classification, deep neural networks (DNNs) offer efficient solutions and can effectively address overfitting concerns. In our algorithm, incorporating four hidden layers, regularization (l1 and l2) for monitoring overfitting, and dropout mechanisms, along with batch normalization to handle batch size during model fitting, yield promising results. Figure 13 depicts the training and validation plot, indicating robust performance with minimal overfitting, evidenced by a logical gap between the two curves and their consistent movement within well-defined performance categories. The confusion matrix results in Figure 14 further validate this performance, with accuracy, precision, recall, and F1-score metrics all registering at 94.6%.
This study highlights the efficacy of DNNs in handling intricate datasets and correlations, surpassing traditional ML models. It paves the way for deeper exploration into DDI recognition, with the goal of minimizing testing expenses while improving precision. For additional insights into the performance of ML models, Figure 15 presents a comparative plot of all models utilized for DDI type classification.

4. Conclusions

We propose a novel approach for predicting multi-type drug-drug interactions (DDIs) using supervised contrastive learning and three-level loss functions. Through rigorous validation, we demonstrate the effectiveness of these techniques and their superiority over existing methods. Case studies further validate our model's ability to identify previously unknown DDIs, highlighting its practical significance in drug discovery and healthcare.
The prediction of DDIs holds significant implications for drug discovery, offering insights into risk reduction and interaction mechanisms. By integrating diverse drug data and harnessing multi-source information for prediction, our model surpasses current methodologies in terms of performance. Statistical analysis underscores the substantial performance enhancements achieved by our approach, indicating its potential for DDI prediction.
Key findings from our investigation include:
  • Deep neural networks (DNNs) outperform traditional machine learning (ML) models, exhibiting consistently higher accuracy, precision, recall, and F1-score metrics (94.6%) compared to XGBoost, LGBM, and AdaBoost models. This underscores the pivotal role of ML, particularly DNNs, in DDI classification tasks.
  • Optimization of ML, particularly DNN, results in highly precise predictions, with correlation metrics exceeding 50% and remaining below 40%, a level of precision not attained by other ML models. This method had an important effect on improvement of the model performance.
  • To address overfitting concerns, we employ various techniques such as learning curves, confusion matrices, and training-validation plots, alongside k-fold cross-validation (k=11), ensuring robust model performance across different datasets.
  • Extensive dataset collection, comprising over 5 million entries for each drug type and DDI type, facilitates comprehensive testing and validation, enabling thorough monitoring of overfitting and underfitting issues.
  • Comparative analysis of ML models, including tree-based models and DNNs, offers valuable insights into the impact of multidrug side effects, highlighting the efficiency and cost-effectiveness of ML in DDI analysis.
  • Monitoring ML model performance with two hyperparameters as learning rate and min child weight made an easier way to tune parameters for a higher performance and lower overfitting.
In conclusion, our study underscores the promise of advanced ML techniques, particularly DNNs, in enhancing DDI prediction accuracy and efficiency. By leveraging diverse data sources and innovative modeling approaches, we contribute to the advancement of DDI research and drug discovery processes.

Ethics and Patient Consent

This study involves computational modeling and does not directly involve human subjects or patient data. Therefore, ethical approval and patient consent were not required. The data used in this research were sourced from publicly available datasets that contain information on drug properties, interactions, and related biological data. These datasets are anonymized and do not include any personally identifiable information. The focus of this research is on developing and validating a deep neural network framework to improve the accuracy of drug-drug interaction (DDI) predictions. By utilizing diverse drug-related information and advanced machine learning techniques, the study aims to enhance the reliability of computational methods in drug discovery and interaction prediction. The ultimate goal is to contribute to safer and more effective clinical practices by providing tools that can help in anticipating and preventing adverse drug interactions. All data handling and analysis were conducted in accordance with relevant guidelines and regulations to ensure the integrity and reproducibility of the results. Furthermore, the study adheres to ethical standards for research in computational biology and bioinformatics, ensuring that all processes are transparent and can be independently verified. By emphasizing the ethical handling of data and the absence of direct human subject involvement, this study maintains a strong ethical foundation while advancing the field of computational drug discovery.

References

  1. Prueksaritanont T, Chu X, Gibson C, et al. Drug-drug interaction studies: regulatory guidance and an industry perspective. AAPS J 2013; 15: 629–645. [CrossRef]
  2. Shamami MA, Teimourpour B, Sharifi F. Community Detection on a Modified Adjacency Matrix: A Novel Network Approach in Drug-Drug Interaction. 2024 20th CSI International Symposium on Artificial Intelligence and Signal Processing (AISP) 2024; 1–5.
  3. Nagai N. Drug interaction studies on new drug applications: current situations and regulatory views in Japan. Drug Metab Pharmacokinet 2010; 25: 3–15. [CrossRef]
  4. Zhang W, Chen Y, Liu F, et al. Predicting potential drug-drug interactions by integrating chemical, biological, phenotypic and network data. BMC Bioinformatics 2017; 18: 1–12. [CrossRef]
  5. Wishart DS, Feunang YD, Guo AC, et al. DrugBank 5.0: a major update to the DrugBank database for 2018. Nucleic Acids Res 2018; 46: D1074–D1082. [CrossRef]
  6. Feng YH, Zhang SW, Shi JY. DPDDI: a deep predictor for drug-drug interactions. BMC Bioinformatics 2020; 21: 419. [CrossRef]
  7. Rohani N, Eslahchi C. Drug-Drug Interaction Predicting by Neural Network Using Integrated Similarity. Scientific Reports 2019 9:1 2019; 9: 1–11. [CrossRef]
  8. Mei S, Zhang K. A machine learning framework for predicting drug–drug interactions. Scientific Reports 2021 11:1 2021; 11: 1–12. [CrossRef]
  9. Xiong G, Yang Z, Yi J, et al. DDInter: an online drug–drug interaction database towards improving clinical decision-making and patient safety. Nucleic Acids Res 2022; 50: D1200–D1207. [CrossRef]
  10. Niazi-Ali S, Atherton GT, Walczak M, et al. Drug-drug interaction database for safe prescribing of systemic antifungal agents. Ther Adv Infect Dis; 8. Epub ahead of print 2021. [CrossRef]
  11. Deng Y, Qiu Y, Xu X, et al. META-DDIE: predicting drug-drug interaction events with few-shot learning. Brief Bioinform; 23. Epub ahead of print 1 January 2022. [CrossRef]
  12. Tatonetti NP, Denny JC, Murphy SN, et al. Detecting Drug Interactions From Adverse-Event Reports: Interaction Between Paroxetine and Pravastatin Increases Blood Glucose Levels. Clin Pharmacol Ther 2011; 90: 133. [CrossRef]
  13. Deng C, Wu J, Shao X. Reliability assessment of machining accuracy on support vector machine. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) 2008; 5315 LNAI: 669–678.
  14. Liu S, Chen K, Chen Q, et al. Dependency-based convolutional neural network for drug-drug interaction extraction. Proceedings - 2016 IEEE International Conference on Bioinformatics and Biomedicine, BIBM 2016 2017; 1074–1080.
  15. Chen L, Liu HG, Liu W, et al. [Analysis of clinical features of 29 patients with 2019 novel coronavirus pneumonia]. Zhonghua Jie He He Hu Xi Za Zhi 2020; 43: E005.
  16. Feng YH, Zhang SW, Zhang QQ, et al. deepMDDI: A deep graph convolutional network framework for multi-label prediction of drug-drug interactions. Anal Biochem; 646. Epub ahead of print 1 June 2022. [CrossRef]
  17. Yang Z, Zhong W, Lv Q, et al. Learning size-adaptive molecular substructures for explainable drug–drug interaction prediction by substructure-aware graph neural network. Chem Sci 2022; 13: 8693–8703. [CrossRef]
  18. Vilar S, Friedman C, Hripcsak G. Detection of drug-drug interactions through data mining studies using clinical sources, scientific literature and social media. Brief Bioinform 2018; 19: 863–877.
  19. Lü L, Pan L, Zhou T, et al. Toward link predictability of complex networks. Proc Natl Acad Sci U S A 2015; 112: 2325–2330. [CrossRef]
  20. Zhang Y, Qiu Y, Cui Y, et al. Predicting drug-drug interactions using multi-modal deep auto-encoders based network embedding and positive-unlabeled learning. Methods 2020; 179: 37–46. [CrossRef]
  21. Gottlieb A, Stein GY, Oron Y, et al. INDI: a computational framework for inferring drug interactions and their associated recommendations. Mol Syst Biol 2012; 8: 592. [CrossRef]
  22. Morteza Ghazali S, Alizadeh M, Mazloum J, et al. Modified binary salp swarm algorithm in EEG signal classification for epilepsy seizure detection. Biomed Signal Process Control 2022; 78: 103858. [CrossRef]
  23. Guo Y, Nie Q, MacLean AL, et al. Multiscale Modeling of Inflammation-Induced Tumorigenesis Reveals Competing Oncogenic and Oncoprotective Roles for Inflammation. Cancer Res 2017; 77: 6429–6441. [CrossRef]
  24. Hosseini Rad R, Baniasadi S, Yousefi P, et al. Presented a Framework of Computational Modeling to Identify the Patient Admission Scheduling Problem in the Healthcare System. J Healthc Eng; 2022. Epub ahead of print 2022. [CrossRef]
  25. Olayan RS, Ashoor H, Bajic VB. rratum: DDR: Efficient computational method to predict drug-target interactions using graph mining and machine learning approaches (Bioinformatics (2018) 34:7 (1164-1173) DOI: 10.1093/bioinformatics/btx731). Bioinformatics 2018; 34: 3779.
  26. Ilani MA, Amini L, Karimi H, et al. CNN-based Labelled Crack Detection for Image Annotation. Epub ahead of print 27 May 2024. [CrossRef]
  27. Ilani MA, Tehran SM, Kavei A, et al. Automatic Image Annotation (AIA) of AlmondNet-20 Method for Almond Detection by Improved CNN-Based Model. Epub ahead of print 20 May 2024. [CrossRef]
  28. Ilani MA, Tehran SM, Kavei A, et al. Exploring Machine Learning Models for Lung Cancer Level Classification: A comparative ML Approach. Epub ahead of print 21 May 2024. [CrossRef]
  29. Ilani MA, Tehran SM, Kavei A, et al. A Covid-19 Prediction Approach: Prediction of Infectious Disease Probability Using Machine Learning. Epub ahead of print 23 May 2024. [CrossRef]
  30. Wishart DS, Knox C, Guo AC, et al. DrugBank: a knowledgebase for drugs, drug actions and drug targets. Nucleic Acids Res; 36. Epub ahead of print January 2008. [CrossRef]
  31. Lin S, Chen W, Chen G, et al. MDDI-SCL: predicting multi-type drug-drug interactions via supervised contrastive learning. J Cheminform 2022; 14: 1–12. [CrossRef]
  32. Lecun Y, Bengio Y, Hinton G. Deep learning. Nature 2015 521:7553 2015; 521: 436–444.
Figure 1. drug-drug interaction (DDI) types featurized for ML models.
Figure 1. drug-drug interaction (DDI) types featurized for ML models.
Preprints 108892 g001
Figure 2. Correlation in multidrug features as drug-drug interaction (DDI).
Figure 2. Correlation in multidrug features as drug-drug interaction (DDI).
Preprints 108892 g002
Figure 3. Combination in correlation for less than 0.3.
Figure 3. Combination in correlation for less than 0.3.
Preprints 108892 g003
Figure 4. Combination in correlation for less than 0.4.
Figure 4. Combination in correlation for less than 0.4.
Preprints 108892 g004
Figure 5. Split DDI datasets as training and test for both input and output.
Figure 5. Split DDI datasets as training and test for both input and output.
Preprints 108892 g005
Figure 6. k-fold cross validation with k =11 was employed on DDI’ datasets as a preparatory step before applying machine learning models.
Figure 6. k-fold cross validation with k =11 was employed on DDI’ datasets as a preparatory step before applying machine learning models.
Preprints 108892 g006
Figure 7. Training and Validation plots under consideration of Min child Weight and Learning Rate in XGBoost Model.
Figure 7. Training and Validation plots under consideration of Min child Weight and Learning Rate in XGBoost Model.
Preprints 108892 g007
Figure 8. XGBoost Model Confusion Matrix.
Figure 8. XGBoost Model Confusion Matrix.
Preprints 108892 g008
Figure 9. Training and Validation plots under consideration of Min child Weight and Learning Rate in LGBM Model.
Figure 9. Training and Validation plots under consideration of Min child Weight and Learning Rate in LGBM Model.
Preprints 108892 g009
Figure 10. LGBM Model Confusion Matrix.
Figure 10. LGBM Model Confusion Matrix.
Preprints 108892 g010
Figure 11. Training and Validation plots under consideration of Min child Weight and Learning Rate in AdaBoost Model.
Figure 11. Training and Validation plots under consideration of Min child Weight and Learning Rate in AdaBoost Model.
Preprints 108892 g011
Figure 12. AdaBoost Model Confusion Matrix.
Figure 12. AdaBoost Model Confusion Matrix.
Preprints 108892 g012
Figure 13. Training plot in DNN on DDI dataset.
Figure 13. Training plot in DNN on DDI dataset.
Preprints 108892 g013
Figure 14. DNN Model Confusion Matrix.
Figure 15. Comparison of ML Performance on DDI's Type Classification.
Figure 15. Comparison of ML Performance on DDI's Type Classification.
Preprints 108892 g014
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated