## I. INTRODUCTION
Drug development and discovery is a time-consuming, costly, and intricate process that often takes decades and billions of dollars. The
Author p: UG Student, Siddhartha School of Pharmacy, Dehradun, Uttarakhand.
average cost of introducing a novel medication to market, including opportunity and direct expenses, is about $2.6 billion, according to the Tufts Centre for the Study of medicine Development [1]. Despite these substantial investments, the overall success rate of drug candidates from Phase I trials to market approval is below $10\%$ [2]. This high attrition rate in later phases, combined with escalating research and development (R&D) expenditures, has driven the pharmaceutical industry to seek more efficient and accurate methods.
Artificial intelligence (AI) has emerged as a disruptive force in this field thanks to its powerful computational capabilities. The methods of computing that are included in the artificial intelligence (AI) category include machine learning (ML), deep learning (DL), neural networks, reinforcement learning, and natural language processing (NLP). These technologies can process large biomedical datasets and find patterns that conventional techniques are unable to detect[3,4]. Many facets of drug development, including target identification, chemical creation, protein structure prediction, and biomarker analysis, are currently being addressed by these methods.
For instance, AlphaFold by DeepMind has demonstrated the potential of AI by accurately predicting protein 3D structures, which are critical in understanding biological functions and designing novel drugs [5]. Additionally, the integration of AI with screening at high throughput and omics technologies has enabled pharmaceutical companies to significantly shorten discovery cycles and reduce cost burdens.
This review aims to comprehensively examine the ways that AI is transforming the pharmaceutical industry, focusing on applications across all phases of the development and discovery of new drugs. We present an overview of the AI-powered drug discovery pipeline, analyze current tools and platforms, and explore into the main technical, ethical, and regulatory issues. We also provide insight into the future direction of AI in pharmaceutical R&D, highlighting its significance in precision medicine and collaborative innovation.
Optimising Clinical Trials with AI: Artificial intelligence is quickly changing how clinical trials are planned and carried out by enhancing patient recruitment, stratification, and real-time monitoring. Natural Language Processing (NLP) algorithms are being used to mine clinical notes and electronic health records (EHRs) in order to find qualified participants with high
precision, thereby addressing a significant bottleneck in clinical research: patient recruitment [55]. Additionally, AI makes predictive modelling for patient stratification possible, which enables researchers to divide up patients according to how they are likely to respond to treatment. This reduces trial size and increases statistical power [56]. For instance, IBM Watson Health and Deep6 AI have shown that they can use AI-based analytics to cut patient screening times by over $80\%$ [57]. AI-enabled wearable technology also makes it easier to monitor trial participants in real time, guaranteeing early adverse event detection and enhancing compliance by identifying behavioural patterns [58]. These applications make clinical trials more flexible and effective by lowering trial costs and schedules while simultaneously enhancing safety and data quality.
Pharmacovigilance and AI: In order to ensure long-term drug safety, post-marketing surveillance, also known as pharma vigilance, is essential. Artificial intelligence (AI) provides significant benefits in identifying adverse drug reactions (ADRs) from disparate data sources. Pharma vigilance now uses AI algorithms that can evaluate unstructured data from social media, electronic health records, and scientific literature to discover early warning signals, replacing its previous reliance on manual signal recognition via spontaneous reporting methods [48,59]. Traditional statistical techniques such as disproportionality analysis are not as sensitive or specific in identifying possible ADRs as machine learning classifiers and natural language processing tools [33,60]. Tools such as Med Watcher Social and FAERSmine, for example, have used AI to more quickly and thoroughly detect concealed safety signs [61]. In the era of expedited medication approvals and international pharmacovigilance regulations, artificial intelligence (AI) greatly speeds up regulatory response and improves patient safety by automating signal detection and prioritisation.
AI for Repurposing Drugs: Due to AI-driven approaches, the process of developing new beneficial applications for licensed drugs—known as drug repurposing—has quickened. Through the integration of omics data, clinical phenotypes, and chemical structural information, artificial intelligence systems have the ability to reveal previously undiscovered relationships between medications and illnesses [62]. To make highly accurate predictions about drug-disease correlations and biological pathways, platforms such as Benevolent AI and GENTRL use deep learning models [63]. AI-powered repurposing was particularly important during the COVID-19 pandemic, when medications such as baricitinib were discovered as possible treatments using machine learning algorithms just a few weeks after the virus was characterised [64]. These techniques shorten the discovery timeline and reduce the risk of failure by
focusing on compounds with known safety profiles. As a result, AI has become a strategic tool for extending the therapeutic utility of existing drug libraries in a cost-effective and timely manner.
## II. AI APPLICATIONS ACROSS THE DRUG DEVELOPMENT PIPELINE
### a) Target Identification and Validation
The drug development process begins with identifying and validating biological targets associated with disease mechanisms. Systems powered by artificial intelligence (AI) are demonstrating impressive abilities in examining enormous biological data repositories to find possible therapeutic targets with previously unheard-of speed and precision. By integrating and analyzing genomic, proteomic, and clinical data, AI tools can pinpoint molecules or biological pathways that play key roles in disease progression, providing valuable insights into potential therapeutic interventions.
Machine learning approaches can predict protein structures and functions, analyze protein-protein interactions, and identify disease-relevant biological pathways. These capabilities enable researchers to discover innovative targets and repurpose existing drugs for new indications. Additionally, AI algorithms can evaluate target druggability and prioritize candidates based on their potential clinical relevance, significantly streamlining the initial phases of discovering new drugs.
### b) Drug Design and Screening
AI streamlines the entire procedure of determining compounds that can efficiently modify prospective targets once they have been discovered. In conventional high-throughput screening, hundreds of chemicals are tested experimentally, which takes a lot of time and resources. Large libraries of compounds can be quickly scanned using AI-guided screening approaches to find the ones that have the best chance of binding to the target.
Beyond screening existing compounds, Al enables De novo drug design, such as creating novel chemical compounds structures optimized for specific targets and properties. Generative models can create new chemical entities with desired characteristics while maintaining structural validity and synthetic feasibility. Additionally, AI algorithms aids researchers in prioritising the most promising prospects for additional development by forecasting the pharmacokinetic and pharmacodynamic characteristics of drugs.
### c) Preclinical Research
Comprehensive safety and efficacy evaluations of possible medication candidates are part of preclinical research. In the past, this procedure has been costly, difficult, and frequently ineffective. This stage is being revolutionised by AI tools, which simplify data analysis,
anticipate drug interactions, and more effectively identify interesting molecules.
Potential drug candidates' toxicity can now be predicted using machine learning algorithms, which drastically reduces down the time and money needed for experimental testing. Preclinical testing time and expense can be significantly decreased by using Aldriven systems, enabling researchers to quickly screen thousands of potential compounds and rank those that seem to have the greatest potential. These predictive models can evaluate drug metabolism, absorption, distribution, and potential side effects, enabling earlier elimination of problematic compounds.
### d) Clinical Trials
The inefficiencies of traditional clinical trial designs frequently result in exorbitant expenses, protracted schedules, and perhaps ambiguous findings. Al can enhance the discovery of new therapies, trial results, and patient recruitment by customising treatment plans and inclusion criteria using predictive analytics.
Al-powered algorithms evaluate patient data to stratify patient populations and find pertinent biomarkers, improving trial design and raising the possibility of positive results. Additionally, prior to starting expensive and time-consuming clinical trials, researchers can test hypothetical trial situations, improve study protocols, and reduce risks using AI-driven simulations. Companies are developing patient-centric solutions using wearables and apps to gather real-world data, enabling more accurate tracking of drug safety and efficacy at the individual level.
### e) Post-market Surveillance and Personalized Medicine
AI continues to provide value after drug approval through post-market surveillance and personalized medicine applications. AI algorithms can monitor real-world data to detect previously unidentified side effects, evaluate long-term safety profiles, and identify additional therapeutic indications.
The development of personalised medicine is one of the most exciting uses of AI in medication development. Through better diagnoses, individualised data collection, and clinical decision support, AI can quicken this trend. AI systems are able to retain and examine patient data, such as lifestyle characteristics, clinical histories, and genetic profiles, enabling more specialised treatment strategies. Furthermore, AI technology can find biomarkers linked to medication
responses or the advancement of a disease, allowing for the creation of therapies with the highest possible efficacy and the fewest possible side effects.
## III. AI TECHNIQUES IN DRUG DISCOVERY
### a) Machine Learning Algorithms
Advanced artificial intelligence (AI)-based methods for drug development have replaced more conventional quantitative structure-activity relationship (QSAR) modelling techniques. Among these methods are decision trees, Support vector algorithms, random forests, and linear discriminant analysis (LDA), which can speed up QSAR analysis and increase prediction accuracy. Large datasets can be explored by algorithms for machine learning to discover links and patterns that traditional analysis techniques can omit.
### b) Deep Learning Models
Deep learning techniques have revolutionized drug discovery by enabling more complex pattern recognition and predictive capabilities. Recurrent neural networks deal with sequential data, whereas convolutional neural networks may analyse imaging data. Graph neural networks are particularly valuable for analyzing molecular structures and predicting compound properties. Deep learning models excel at integrating diverse data types and extracting meaningful features from complex biological datasets.
### c) Generative AI
Generative AI models represent a particularly exciting advancement in drug design. Novel structures for molecules with desired features can be generated by using these computational models, enabling the creation of entirely new chemical entities tailored to specific targets. Methodologies incorporating learning via reinforcement, generative networks with adversarial properties, and variational autoencoders allow researchers to explore chemical space more efficiently and design molecules with optimal characteristics.
### d) Large Language Models
Emerging large language models have recently been applied to drug discovery, offering new capabilities for analyzing scientific literature, predicting molecular interactions, and generating insights from unstructured data. These models can process vast amounts of textual information, enabling researchers to leverage knowledge contained in millions of scientific publications and databases to inform drug development decisions.
Table 1: Key AI Techniques and their Applications in Drug Discovery
<table><tr><td>AI Technique</td><td>Applications for Drug Discovery</td><td>Examples</td></tr><tr><td>Algorithms in machine learning</td><td>QSAR modeling, property prediction, virtual screening</td><td>Random Forest, SVM, Decision Trees</td></tr><tr><td>Deep Learning</td><td>Protein structure prediction, binding affinity estimation</td><td>Neural networks, CNN, RNN</td></tr><tr><td>AI generative</td><td>Designing de novo molecules and optimising leads</td><td>GANs, VAEs, Reinforcement Learning</td></tr><tr><td>Large Language Models</td><td>Literature analysis, knowledge extraction</td><td>Emerging models for scientific text processing</td></tr><tr><td>Computer Vision</td><td>High-throughput screening analysis, histopathology</td><td>Image recognition algorithms</td></tr><tr><td>Natural Language Processing</td><td>Mining scientific literature, patient data analysis</td><td>Text classification, named entity recognition</td></tr><tr><td>Reinforcement Learning</td><td>Molecular optimization, clinical trial design</td><td>Policy-gradient methods, Q-learning</td></tr></table>
## IV. OVERVIEW OF THE DRUG DEVELOPMENT PIPELINE
The traditional drug development pipeline is a multistep, iterative process designed to identify, develop, and bring new therapeutic agents to market.
The main goal is to ensure the efficacy, safety and quality of novel pharmaceutical compounds before they reach patients. Each stage presents unique challenges and opportunities for optimization, especially with taking into account of Artificial Intelligence (AI).
Table 2: AI Applications and Emerging Software in Drug Development Stages
<table><tr><td>Drug Development Stage</td><td>AI Techniques Used</td><td>Emerging Pharmacy Software / Tools</td><td>Recent Advancements / Key Features</td><td>Description</td></tr><tr><td>Target Identification & Validation</td><td>- Machine Learning (ML)
- Deep Neural Networks (DNN)
- NLP for literature mining</td><td>- IBM Watson Discovery
-BioXpress
- Open Targets Platform</td><td>- Predicting druggable targets from multi-omics data
- NLP-powered biomarker extraction
- Integration of CRISPR screening data</td><td>Identifying genes, proteins, or pathways associated with disease pathogenesis using techniques such as genomics, transcriptomics, and bioinformatics tools.</td></tr><tr><td>Hit Discovery / Screening</td><td>- Virtual Screening
- Reinforcement Learning
- Generative Adversarial Networks (GANs)</td><td>- Schrödinger Glide
-DeepChem
-Atomwise (AtomNet)</td><td>- AI-guided high-throughput screening
- Ligand-based drug design using DL
- Molecular fingerprint learning</td><td>Screening large chemical libraries (both physical and virtual) to find compounds that exhibit biological activity against the identified target.</td></tr><tr><td>Lead Optimization</td><td>- QSAR Modeling
- Active Learning
- Bayesian Optimization</td><td>-MOSES
-DeepMol
- Chemprop</td><td>- AI for ADMET prediction
- Property optimization with molecular generative models
- Predictive SAR mapping</td><td>Refining the chemical structure of hit compounds to enhance selectivity, potency, safety, metabolic stability. This often involves structure-activity relationship (SAR) studies.</td></tr><tr><td>Preclinical Development</td><td>- Supervised ML
- Toxicity Prediction Models
- Graph Neural Networks (GNNs)</td><td>-pkCSM
-DeepTox
-ADMETlab 2.0</td><td>- Predicting off-target toxicity
- Physiologically-based pharmacokinetics (PBPK) modeling
- GNN-based prediction of toxicity profiles</td><td>Conducting in vivo and in vitro experiments to evaluate pharmacokinetics (PK), pharmacodynamics (PD), and toxicity profile of the optimized compound. Animal models are used for predictive safety.</td></tr><tr><td>Clinical Trial Design & Execution</td><td>- AI-based patient stratification
- Predictive Modeling
- Digital Twins</td><td>- Deep 6 AI
- Unlearn. AI
- TriNetX</td><td>- Simulated control arms using digital twins
- AI for trial recruitment and dropout prediction
- Adaptive clinical trial design</td><td>Testing in human volunteers across three phases: Phase I (safety and dosing), Phase II (efficacy and side effects), and Phase III (confirmation of effectiveness and monitoring of adverse reactions).</td></tr><tr><td>Regulatory Submission & Market Access</td><td>- Natural Language Processing
- Data Harmonization
- Knowledge Graphs</td><td>- FDA's CDER AI-based systems
- Regulatory NLP Tools (e.g., Linguuamatics)</td><td>- Automated generation of eCTD documents
- Real-world evidence integration
- AI for pharmacovigilance signal detection</td><td>Submission of comprehensive data packages to agencies like the FDA or EMA to demonstrate safety, efficacy, and manufacturing quality for market authorization.</td></tr></table>
In recent years, AI has begun to redefine this traditional workflow. At the target identification stage, machine learning models analyze high-throughput omics data to uncover disease-relevant genes or pathways. For example, studies such as those by Zeng et al. (2020) have demonstrated how deep learning platforms like Deep Target can effectively predict molecular targets based on multi-omics integration.
For lead discovery, Al-driven virtual screening platforms such as AtomNet and DeepChem can evaluate billions of compounds in silico, drastically reducing the time and cost associated with physical screening (Wallach et al., 2015; Ramsundar et al.,
2019). These platforms use convolutional neural networks and graph-based learning algorithms to predict molecular binding affinities and activities with impressive accuracy. The use of molecular docking algorithms accelerated by AI further helps prioritize potential hits for synthesis and biological testing.
During lead optimization, AI tools predict physicochemical properties, simulate metabolism, and propose novel analogs through generative models like GANs and reinforcement learning. For instance, Insilico Medicine developed DDR1 kinase inhibitors using a deep generative approach that reached preclinical trials in less than 18 months (Zhavoronkov et al., 2019). AI
can also help guide Structure-Based Drug Design (SBDD) and Ligand-Based Drug Design (LBDD) by learning SAR patterns and predicting molecular modifications that may improve activity profiles.
In preclinical testing, AI supports toxicity prediction using in silico methods such as DeepTox and ProTox-II, trained on toxicogenomic databases. These systems can assess hepatotoxicity, cardiotoxicity, and off-target interactions, potentially reducing the need for animal testing (Mayr et al., 2016). Additionally, systems biology modeling, powered by AI, can simulate entire organ systems, helping to predict human-relevant responses without invasive methods.
For clinical trials, AI aids in optimizing trial design, predicting dropout rates, and stratifying patient populations. IBM Watson Health, for instance, has collaborated with healthcare institutions to improve patient matching for oncology trials, reducing recruitment time by over $50\%$ (IBM Watson Health, 2020). AI also assists in remote monitoring through wearable devices and real-time data integration. Furthermore, NLP algorithms process patient records and unstructured data to identify adverse drug reactions and eligibility criteria at scale.
Regulatory approval processes are increasingly incorporating AI-driven analyses. Regulatory bodies like the U.S. FDA have released frameworks supporting the use of machine learning in regulatory submissions, and the EMA's Big Data Task Force is developing standards for validation and traceability of AI-based outputs (FDA, 2021; EMA, 2020). These initiatives aim to ensure transparency, reproducibility, and clinical relevance of AI-supported decisions in healthcare product evaluations.
Overall, the integration of AI across the drug development pipeline enhances decision-making, reduces resource waste, and improves success rates by addressing bottlenecks that have historically hindered drug development timelines and outcomes. As more validated models and regulatory frameworks are established, AI will likely become a cornerstone of pharmaceutical R&D workflows.
### a) Recent Developments and Innovations (2023-2025)
The field of AI in drug discovery has witnessed significant advancements in recent years. In 2023, research company Startus Insights identified nine key trends where companies are breaking new ground, including AI & data analytics, patient-centric trials, and assay development.
Companies like Protai in Israel are leveraging AI to build drug discovery platforms, while Italian startup Metabolics is predicting drug effects by digitizing human cells. Other innovations include gut-on-chip testing solutions developed by Latvian startup Cellbox, which replicate human organs and run experiments on chips controlled by biosensors.
Researchers can now swiftly search through huge libraries of compounds for possible therapeutic candidates thanks to the combination of AI with high-throughput screening methods, greatly speeding up the early phases of drug discovery. AI analysis in conjunction with developments in proteomics, genomics, and other omics technologies has led to a better understanding of biological systems and disease pathways.
By 2025, the combination of both generative AI and enormously language modelshave further enhanced the efficiency and effectiveness of the drug development process, enabling more precise target identification, faster lead optimization, and improved clinical trial design.
### b) Challenges and Limitations
Despite remarkable progress, plenty of challenges still exist in the use of AI in drug development and development. Data quality issues represent a significant constraint, as the level of accuracy of the data determines how well AI models perform., they are trained on. Limited availability of high-quality, well-annotated datasets, particularly for rare diseases and novel targets, can restrict AI model performance.
Regulatory considerations pose another challenge, as regulatory frameworks are still adapting to AI-driven drug development approaches. Demonstrating the reliability and reproducibility of AI-generated results to regulatory authorities requires careful validation and transparency.
Ethical concerns must also be addressed, particularly regarding data privacy, bias in training datasets, and the appropriate use of AI-generated findings. Ensuring that AI applications in drug discovery adhere to ethical principles is essential for maintaining public trust and promoting responsible innovation.
Technical constraints continue to limit certain AI applications, especially for complex biological systems that are not fully understood. Interdisciplinary collaboration between computational experts and domain specialists is essential to overcome these limitations and realize the full potential of AI in pharmaceutical research.
## V. AI IN TARGET IDENTIFICATION AND VALIDATION
Target identification and validation represent the foundational step in drug discovery, where researchers aim to pinpoint disease-associated genes, proteins, or pathways that can be modulated to achieve therapeutic effects. Traditional approaches often rely on labour-intensive methods, including gene knockout studies, proteomics, and biochemical assays. However, these methods are limited by their scalability and complexity.
Artificial Intelligence (AI), particularly machine learning (ML) and deep learning (DL), is transforming this process by integrating and analyzing diverse biological datasets—including genomics, transcript-
tomics, proteomics, metabolomics, and interact omics—to identify promising therapeutic targets with higher accuracy and speed.
Table 3: Comprehensive Overview of AI in Target Identification & Validation
<table><tr><td>Category</td><td>Detailed Description</td></tr><tr><td>Objective</td><td>The goal of validation and target identification is to pinpoint biologically relevant molecules—typically proteins, genes, or non-coding RNAs—that are involved in disease pathways and are amenable to therapeutic modulation. AI enhances this process by mining vast biological datasets, integrating omics layers (genomics, transcriptomics, proteomics), and modeling biological networks to identify novel, high-confidence drug targets.</td></tr><tr><td>Key AI Techniques</td><td>- Machine Learning (ML): Supervised and unsupervised models analyze multi-dimensional datasets (e.g., gene expression, mutation frequencies) to classify potential targets based on disease relevance. - Natural Language Processing (NLP): Extracts knowledge from millions of biomedical documents (PubMed, patents, clinical trials) to detect implicit gene-disease relationships, using named entity recognition (NER) and relation extraction algorithms. - Graph Neural Networks (GNNs): Capture intricate relationships in protein-protein interaction (PPI) and gene regulatory networks. These models learn embeddings of network components to predict unknown interactions. - Deep Learning (DL): Autoencoders and convolutional networks integrate and reduce dimensionality of omics datasets (RNA-seq, ChIP-seq) to uncover latent patterns relevant to target biology. - Knowledge Graphs (KGs): Structure biological knowledge (e.g., disease-gene-pathway-drug linkages) into graph-based systems where AI infers novel associations using link prediction.</td></tr><tr><td>Emerging Pharmacy Software/Platforms</td><td>- IBM Watson Discovery: Uses NLP and cognitive computing to extract target-related evidence from biomedical text. - Open Targets Platform: A collaborative public-private partnership between EMBL-EBI, GSK, Biogen, and others; integrates genetics, expression, and literature data to rank potential drug targets by AI-scored evidence. - Pharos (NIH/IDG): Provides AI-curated target development levels (TDLs), integrating expression, structure, and binding data for over 20,000 human proteins. - BioXpress: A cancer-specific gene expression database powered by AI algorithms to identify over- or under-expressed genes from RNA-seq datasets. - DeepTarget: Utilizes transcriptomic and epigenetic features with deep learning to classify gene targets based on tissue-specific and cancer-relevant biomarkers. - TargetMine: Integrates functional genomics and protein information; uses AI for scoring and prioritization.</td></tr><tr><td>Recent Technological Advancements</td><td>- Multi-Omics Integration with AI: Tools like DeepOmix use variationalautoencoders to combine DNA methylation, proteomics, and transcriptomics data to discover central regulatory targets in diseases like cancer and neurodegeneration ([Vamathevan et al., 2019]). - CRISPR Screening Enhanced by AI: Deep learning interprets CRISPR-Cas9 knockout data to reveal genes essential for cell survival in disease contexts ([Carvalho et al., 2021]). - Target Druggability Prediction: Tools like TargetDB use ML models trained on known drug-target pairs to estimate the druggability of novel genes using sequence and structure features. - Literature-Based Target Extraction: NLP models like BERT and BioBERT extract disease-gene-drug triads from full-text articles, speeding up evidence-based hypothesis generation ([Lee et al., 2020]). - AI for Undruggable Proteins: Tools such as AlphaFold2 + DeepSite identify cryptic or allosteric binding pockets previously deemed inaccessible, opening new possibilities for target validation.</td></tr><tr><td>Advantages of AI Integration</td><td>- Data-Driven Discovery: AI leverages vast heterogeneous data (clinical, experimental, literature) to uncover hidden insights. - High Throughput and Scale: Enables simultaneous evaluation of thousands of potential targets across hundreds of disease states. - Precision in Target Selection: Prioritizes high-efficacy, low-toxicity targets using AI-based filtering. - Personalized Targeting: AI can stratify targets based on patient subgroups, making way for personalized therapies.</td></tr><tr><td>Limitations and Current Challenges</td><td>- Data Quality and Integration Issues: Inconsistent annotations, missing data, and batch effects hamper model training. - Black-Box Models: Deep learning approaches, while powerful, lack interpretability, which hinders regulatory acceptance. - Validation Bottleneck: Predicted targets require expensive and time-consuming wet-lab validation, slowing translation. - Biological Complexity: AI may struggle to fully capture nonlinear, context-specific biological interactions like feedback loops and epigenetic regulation. - Regulatory & Ethical Concerns: Use of AI in high-stakes decisions (e.g., oncology targets) needs ethical oversight and explainability frameworks.</td></tr></table>
### a) Machine Learning for Multi-Omics Integration
Machine learning approaches like random forests (RF), ensemble learning models, and support vector machines (SVM) are frequently employed to classify disease-relevant genes or proteins from massive multi-omics datasets. These methods can extract features, cluster disease subtypes, and prioritize potential drug targets based on their functional roles and network topologies [1,2].
For example, tools like DeepTarget leverage neural networks to combine gene expression, mutation frequency, and pathway involvement to predict viable targets in cancer therapy (Zeng et al., 2020) [3]. Similarly, PANDAomics by Insilico Medicine utilizes AI to rank targets based on disease association, biological relevance, druggability, and novelty [4].
### b) Protein-Protein Interaction (PPI) Networks and AI
Deep learning has shown promise in modeling protein-protein interactions, a crucial aspect of identifying nodes central to disease progression. Platforms such as STRING, BioGRID, and HINT offer curated PPI databases that, when coupled with graph neural networks (GNNs), can reveal hidden relationships within the proteome. These insights enable the identification of key regulatory proteins and interaction hubs [5,6].
### c) Predictive Modeling and Literature Mining
Models for natural language processing (NLP), among them Bidirectional Encoder Representations from Transformers (BERT) and SciBERT, are applied to mine biomedical literature and databases like PubMed, identifying emerging targets, associated pathways, and biomarkers [7]. AI-driven tools like IBM Watson Discovery can analyze thousands of scientific papers to extract meaningful patterns and hypotheses [8].
### d) Structural and Functional Annotation with AI
AI also contributes to functional annotation and structural prediction of potential targets. AlphaFold, developed by DeepMind, revolutionized protein structure prediction with over $90\%$ accuracy, enabling the visualization of binding sites and aiding in structure-based drug design. [9]
Case Example: A study by Aliper et al. (2016) demonstrated the use of DL models trained on transcriptomic profiles to distinguish between cancerous and non-cancerous cells and identify differential gene expression patterns that could serve as target leads for specific cancer types. These models significantly outperformed conventional clustering techniques [10].
Overall, AI empowers researchers to overcome data complexity and variability in biological systems, enhancing the efficiency and accuracy of target identification and validation. It paves the way for personalized therapeutic strategies by identifying targets that are context-specific and more likely to succeed in downstream development stages.
## VI. AI IN VIRTUAL SCREENING AND DRUG DESIGN
Virtual screening (VS) and drug design are critical components of the early drug discovery process. These techniques aim to identify potential lead compounds with high affinity for a target protein by screening large chemical libraries. Traditional approaches include ligand-based and structure-based virtual screening using molecular docking and pharmacophore modeling. However, these methods often suffer from limited accuracy, high false-positive rates, and significant computational burden.
Artificial Intelligence (AI) addresses these challenges by enabling more precise predictions of
molecular interactions, binding affinities, and drug- likeness properties. With advancements in deep learning, generative modeling, and reinforcement learning, AI has transformed both virtual screening and de novo drug design.
### a) AI-Driven Virtual Screening
Deep learning models such as convolutional neural networks (CNNs), graph neural networks (GNNs), and recurrent neural networks (RNNs) are employed to predict the bioactivity of compounds against a given target. Tools like AtomNet utilize 3D CNNs for structure-based virtual screening, enabling accurate identification of active compounds [11]. DeepChem and Chemprop are other open-source platforms that provide ML-based frameworks for property prediction, binding affinity estimation, and molecular classification [12,13].
Graph-based deep learning models excel at representing molecular structures and interactions. GNNs consider atom and bond features as nodes and edges, respectively, and can predict activity, toxicity, and solubility with high accuracy. These models significantly outperform traditional quantitative structure-activity relationship (QSAR) methods.
### b) Al in De Novo Drug Design
AI models like Generative Adversarial Networks (GANs), Variational Autoencoders (VAEs), and Reinforcement Learning (RL) are increasingly used to generate novel chemical structures with desired properties. The generative model learns the chemical space and generates synthetically feasible molecules optimized for drug-likeness, ADMET properties, and target binding.
One prominent example is Insilico Medicine, which designed and synthesized potent DDR1 kinase inhibitors using a generative pipeline combining RL and GANs in less than 18 months—a process that traditionally takes 4-6 years [14]. Similarly, BenevolentAI and Exscientia use AI for automated compound generation and optimization, achieving high hit-to-lead ratios.
### c) AI for Docking and Binding Affinity Prediction
AI-powered docking algorithms use DL models to predict ligand-protein binding poses and scoring functions. Tools like DeepDock, OnionNet, and KDEEP have been shown to outperform classical scoring functions in blind docking challenges [15,16]. These models learn spatial features from protein-ligand complexes and generalize across diverse targets.
### d) Case Studies and Applications
- COVID-19 Drug Discovery: AI-based drug screening was rapidly employed to identify potential inhibitors of SARS-CoV-2 proteins. Benevolent AI identified baricitinib as a repurposing candidate for COVID-19 treatment, later validated in clinical settings [17].
- AI-Based Fragment Screening: Schrödinger's Glide and DeepDock tools were used in campaigns for oncology and CNS diseases, integrating fragment-based drug discovery (FBDD) with DL models to improve early-stage hit generation.
Al's role in virtual screening and molecular design is revolutionizing lead identification and optimization. It offers unprecedented scalability, adaptability, and predictive accuracy, reducing time-to-hit and increasing the probability of downstream success.
## VII. AI IN PRECLINICAL TESTING AND SAFETY ASSESSMENT
Preclinical testing is a crucial phase in the drug development process, involving in vitro (cell culture) and in vivo (animal) studies to evaluate the safety, toxicity, pharmacokinetics (PK), and pharmacodynamics (PD) of drug candidates. This phase is vital for understanding how a drug behaves in a biological system and determining whether it is safe enough to progress to human clinical trials. However, traditional preclinical models are costly, time-consuming, ethically controversial, and not always predictive of human outcomes.
Artificial Intelligence (AI) is increasingly being applied to improve the predictive accuracy of preclinical assessments and reduce dependence on animal testing. AI enables the analysis of large-scale toxicogenomic, pharmacogenomic, and bioassay data to forecast potential adverse events, optimize compound dosing, and model drug metabolism.
### a) In Silico Toxicity Prediction
Machine learning models such as random forests, support vector machines (SVM), and deep neural networks (DNN) are widely used to predict toxicity endpoints including hepatotoxicity, cardiotoxicity, genotoxicity, and nephrotoxicity. Tools such as DeepTox, ProTox-II, and ADMETlab apply AI to analyze chemical structures and predict toxicological outcomes before any laboratory testing [18,19].
### b) Predictive Pharmacokinetics and Metabolism
AI models can simulate Absorption, Distribution, Metabolism, Excretion, and Toxicity (ADMET) profiles of drug candidates. For instance, the pkCSM tool uses graph-based signatures to forecast oral bioavailability, blood-brain barrier penetration, and cytochrome P450 interactions [20]. Deep learning approaches like those used by ADMET Predictor (Simulations Plus) further enhance prediction accuracy and guide compound optimization.
### c) Systems Biology and Organs-on-Chips
Systems biology modeling, powered by AI, integrates omics and physiological data to simulate
tissue-level responses to drugs. When combined with microfluidic technologies (organ-on-chip), AI can model complex biological interactions to predict human-relevant outcomes with greater reliability. For example, AI-enhanced liver-on-chip models can detect hepatotoxicity more accurately than standard animal models [21].
### d) Drug-Drug Interaction (DDI) Prediction
AI also plays a vital role in predicting potential drug-drug interactions, a common reason for postmarketing drug withdrawals. Deep learning models trained on electronic health records (EHRs), pharmacovigilance databases, and molecular data can anticipate adverse interactions early in the pipeline. DeepDDI is one such model using deep neural networks to identify clinically relevant DDIs [22].
### e) Case Studies
- Merck has integrated AI models with high-content screening data to predict neurotoxicity and prioritize safe leads earlier in the process.
- Novartis and Atomwise collaborate to use AI in predicting mitochondrial toxicity, one of the leading causes of late-stage failure.
Figure 3: AI in Preclinical Safety Workflow (A flowchart depicting data ingestion $\rightarrow$ feature extraction $\rightarrow$ toxicity/PK/PD modeling $\rightarrow$ output visualization and risk scoring.)
By leveraging AI in preclinical testing, pharmaceutical companies can identify safety liabilities earlier, minimize reliance on animal models, and accelerate regulatory submissions with higher confidence.
## VIII. AI IN CLINICAL TRIALS AND PATIENT STRATIFICATION
Clinical trials are essential for evaluating the safety and efficacy of new drugs in humans, but they are also among the most expensive and time-consuming phases of drug development. Traditional trial designs often suffer from high failure rates, low recruitment efficiency, and lack of personalized treatment strategies. Artificial Intelligence (AI) offers novel solutions to these challenges by enhancing patient recruitment, optimizing trial design, stratifying patients, and enabling real-time monitoring.
### a) Patient Recruitment and Matching
Recruiting eligible participants is a major bottleneck in clinical trials. AI-driven natural language processing (NLP) systems can analyze electronic health records (EHRs), clinical notes, and diagnostic reports to match patients with trial inclusion and exclusion criteria. For instance, IBM Watson for Clinical Trial Matching has shown a $70\%$ reduction in trial screening time by automatically identifying qualified participants from clinical databases [23].
### b) Trial Design Optimization
Machine learning models are used to simulate various trial scenarios, allowing researchers to predict potential outcomes and adapt protocols accordingly. Bayesian adaptive trial designs, supported by AI, can adjust randomization probabilities based on interim results, improving trial efficiency and ethical considerations. AI also helps in selecting endpoints, optimizing dosing regimens, and predicting dropout rates [24].
### c) Patient Stratification and Precision Medicine
AI can segment patients into subgroups based on biomarkers, genetic profiles, lifestyle data, and disease phenotypes. Unsupervised learning algorithms like clustering and t-SNE are used to identify hidden patient subpopulations that may respond differently to treatments. This stratification supports precision medicine, ensuring the right drug is given to the right patient at the right time [25].
### d) Real-Time Monitoring and Remote Data Collection
With the rise of wearable devices and digital health platforms, AI facilitates continuous remote monitoring of trial participants. Deep learning algorithms can analyze data from heart rate monitors, glucose sensors, and sleep trackers to detect early signs of adverse events or non-compliance. These tools enhance patient safety and reduce the need for in-person visits [26].
### e) Case Study: AI in Oncology Trials
Pfizer partnered with Concerto HealthAI to use real-world evidence and AI to optimize oncology trials. The collaboration helped improve trial feasibility, identify responsive patient groups, and reduce protocol amendments, which are typically time-consuming and costly.
By integrating AI into clinical trials, the pharmaceutical industry is moving toward more dynamic, data-driven, and patient-centric research models. These innovations can shorten trial durations, improve outcome predictability, and increase regulatory acceptance of trial results.
## IX. CHALLENGES, LIMITATIONS, AND ETHICAL CONSIDERATIONS
While the integration of Artificial Intelligence (AI) into drug discovery and development offers remarkable potential, several challenges and ethical concerns must be addressed to ensure responsible and effective implementation.
### a) Data Quality and Standardization
AI models rely heavily on high-quality, well-annotated datasets. Inconsistent data formatting, missing values, and heterogeneous sources (e.g., clinical, genomic, imaging) can reduce model accuracy and reproducibility. Moreover, lack of standardization in data collection protocols across institutions hinders the integration and generalization of AI models [27].
### b) Interpretability and Transparency
Most deep learning models function as "black boxes," making it difficult to interpret how specific decisions are made. This lack of transparency can hinder the trust of regulatory bodies, clinicians, and patients. Developing explainable AI (XAI) models is crucial to ensure traceability, accountability, and informed decision-making in healthcare applications [28].
### c) Regulatory and Validation Frameworks
AI-based tools used in drug discovery must adhere to stringent validation standards before regulatory acceptance. However, there is a lack of clear regulatory guidelines tailored specifically for AI systems. The U.S. FDA and European Medicines Agency (EMA) are actively developing frameworks, but harmonization across global agencies remains a challenge [29].
### d) Bias and Generalizability
AI models can inadvertently reflect biases present in training datasets, leading to unequal performance across different patient populations. This is especially concerning in precision medicine, where biased models can result in suboptimal or unsafe treatment recommendations for underrepresented groups [30].
### e) Ethical and Privacy Concerns
AI applications in healthcare often require access to sensitive patient data. Ensuring patient privacy, data ownership, and compliance with regulations such as the General Data Protection Regulation (GDPR) and Health Insurance Portability and Accountability Act (HIPAA) is essential. Additionally, ethical dilemmas arise when AI systems are involved in life-altering decisions without sufficient human oversight [31].
### f) Talent and Infrastructure Gaps
The implementation of AI technologies in drug development requires a skilled workforce proficient in data science, biology, and regulatory science. Many pharmaceutical companies face challenges in building interdisciplinary teams and developing the necessary computational infrastructure to support AI workflows [32].
### g) Cost and Resource Allocation
Although AI promises long-term savings, its initial implementation can be costly. Investments are needed in data infrastructure, high-performance computing, model training, and integration into existing R&D pipelines. For smaller firms and academic institutions, such costs may be prohibitive without collaborative partnerships.
Despite these challenges, continued innovation, policy development, and interdisciplinary collaboration can help address limitations and foster responsible AI use in drug discovery.
## X. FUTURE DIRECTIONS AND CONCLUSION
The application of Artificial Intelligence (AI) in drug discovery and development is rapidly evolving, offering transformative potential across the entire pharmaceutical value chain. Looking ahead, several promising directions are expected to redefine the landscape of biomedical research and therapeutic innovation.
The future of AI in drug discovery and development holds tremendous promise. Integration of multi-omics data-combining genomic, proteomic, metabolomic, and other biological information-will enable more comprehensive modeling of disease mechanisms and drug effects. This holistic approach will allow for more precise target identification and personalized treatment strategies.
Advances in quantum computing may further accelerate AI applications in drug discovery by enabling more complex molecular simulations and property predictions. Federated learning approaches could facilitate collaborative research while preserving data privacy, allowing organizations to collectively train AI models without sharing sensitive information.
The combination of AI with laboratory automation represents another promising direction. Al-guided robotic systems can design, execute, and analyze experiments with minimal human intervention, creating a closed-loop discovery process that iteratively improves based on experimental results.
As AI technologies continue to evolve, their integration into the drug development pipeline will likely become more seamless and comprehensive. This evolution will require ongoing collaboration between AI researchers, drug developers, clinicians, and regulatory authorities to ensure that AI-driven approaches deliver safe, effective, and accessible therapeutic innovations.
### a) Integration with Multi-Modal Data
The future of AI in drug discovery lies in the integration of diverse data modalities, including genomics, proteomics, transcriptomics, metabolomics, imaging, and electronic health records (EHRs). Multimodal AI models will enable a more holistic understanding of disease mechanisms and drug responses, allowing for precise, patient-specific therapeutic interventions. Tools like Google's DeepMind and Meta AI are advancing this capability with deep learning architectures capable of processing heterogeneous biomedical data at scale.
### b) Federated and Privacy-Preserving Learning
Federated learning models are gaining attention for their ability to train AI algorithms across decentralized datasets without sharing sensitive patient data. This approach helps overcome data privacy concerns and facilitates cross-institutional collaborations. As privacy regulations become more stringent, such models will play a critical role in clinical AI applications.
### c) AI-Augmented Drug Repurposing
AI is expected to significantly enhance drug repurposing strategies by identifying new therapeutic uses for existing drugs. This is particularly valuable in pandemic scenarios and rare diseases, where time and resources are limited. Platforms like BenevolentAI and Healx are already pioneering this field with AI models trained on biomedical knowledge graphs and real-world evidence.
### d) Digital Twins and Personalized Drug Testing
The concept of digital twins—virtual replicas of individual patients—combined with AI could revolutionize personalized medicine. These models simulate disease progression and drug response in silico, enabling personalized treatment regimens, dosing strategies, and adverse event prediction before administration.
### e) Regulatory Harmonization and Ethical AI
Future success of AI in drug development will depend on the establishment of globally harmonized regulatory frameworks that address data integrity, model validation, bias mitigation, and ethical considerations. Collaboration between regulators, industry stakeholders, and academic institutions will be essential to achieve trust and widespread adoption.
### f) Enhanced Human-AI Collaboration
Rather than replacing scientists, AI will increasingly serve as a decision-support partner, augmenting human capabilities. The synergy between domain experts and AI tools will accelerate hypothesis generation, streamline experimentation, and improve R&D productivity.
## XI. CONCLUSION
AI has emerged as a powerful force in transforming drug discovery and development, from target identification to clinical trials. By enabling faster, cheaper, and more accurate research processes, AI holds the promise of accelerating the delivery of safer and more effective therapies to patients. Despite existing challenges related to data quality, model transparency, and regulatory readiness, the future is optimistic. As technologies mature and collaborative ecosystems evolve, AI will become an indispensable pillar of precision medicine and pharmaceutical innovation.
Artificial intelligence has emerged as a transformative force in drug discovery and development, offering innovative solutions to longstanding challenges in pharmaceutical research. By accelerating target identification, enabling virtual screening of vast compound libraries, predicting molecular properties, optimizing clinical trials, and advancing personalized medicine, AI technologies are reshaping the entire drug development landscape.
While challenges related to data quality, regulatory considerations, and ethical concerns persist, the rapid pace of innovation in AI methodologies suggests that many of these limitations will be addressed in the coming years. The integration of diverse AI techniques-from machine learning and deep learning to generative AI and large language models provides a rich toolkit for addressing complex problems across the drug development pipeline.
The successful implementation of AI in pharmaceutical research requires interdisciplinary collaboration, combining expertise in computer science, biology, chemistry, medicine, and regulatory affairs[1]. By fostering such collaborative approaches and continuing to advance AI technologies, the field stands poised to dramatically accelerate the discovery and development of novel therapies, potentially transforming patient care and reducing the global burden of disease.
As we look to the future, the synergy between human expertise and artificial intelligence will likely define a new paradigm for drug discovery and development—one characterized by greater efficiency, reduced costs, higher success rates, and more personalized therapeutic approaches. This evolution promises to benefit not only the pharmaceutical industry but also healthcare systems and, most importantly, patients awaiting new treatments for challenging medical conditions.
Generating HTML Viewer...
References
72 Cites in Article
Indhupriya Subramanian,Srikant Verma,Shiva Kumar,Abhay Jere,Krishanpal Anamika (2020). Multi-omics Data Integration, Interpretation, and Its Application.
Christof Angermueller,Tanel Pärnamaa,Leopold Parts,Oliver Stegle (2016). Deep learning for computational biology.
X Zeng,X Song,T Ma,X Pan,Y Zhou,Y Hou,Z Zhang (2020). DeepTarget: End-to-end learning framework for predicting protein targets of small molecules.
A Zhavoronkov,Q Vanhaelen,T Oprea (2020). Will Artificial Intelligence for Drug Discovery Impact Clinical Pharmacology?.
D Szklarczyk,A Gable,D Lyon,A Junge,S Wyder,J Huerta-Cepas (2019). STRING v11: protein-protein association networks with increased coverage, supporting functional discovery in genome-wide experimental datasets.
C Stark,B Breitkreutz,T Reguly,L Boucher,A Breitkreutz,M Tyers (2006). BioGRID: a general repository for interaction datasets.
I Beltagy,K Lo,A Cohan SciBERT: A Pretrained Language Model for Scientific Text.
D Ferrucci (2012). Introduction to "This is Watson.
A Aliper,S Plis,A Artemov,A Ulloa,P Mamoshina,A Zhavoronkov (2016). Deep learning applications for predicting pharmacological properties of drugs and drug repurposing using transcriptomic data.
I Wallach,M Dzamba,A Heifets Atom Net: A deep convolutional neural network for bioactivity prediction in structure-based drug discovery.
B Ramsundar,P Eastman,P Walters,V Pande (2019). Deep Learning for the Life Sciences.
K Yang,K Swanson,Jin Coley,C Eiden,P Gao,H (2019). Analyzing Learned Molecular Representations for Property Prediction.
A Zhavoronkov,Y Ivanenkov,A Aliper (2019). Deep learning enables rapid identification of potent DDR1 kinase inhibitors.
Z Liu,Y Li,L Han,J Li,J Liu,Z Zhao (2015). PDB-wide collection of binding data: current status of the PDBbind database.
P Richardson,I Griffin,C Tucker,D Smith,O Oechsle,A Phelan (2020). Baricitinib as potential treatment for 2019-nCoV acute respiratory disease.
Andreas Mayr,Günter Klambauer,Thomas Unterthiner,Sepp Hochreiter (2016). DeepTox: Toxicity Prediction using Deep Learning.
P Banerjee,A Eckert,A Schrey,R Preissner (2018). ProTox-II: A webserver for the prediction of toxicity of chemicals.
Douglas Pires,Tom Blundell,David Ascher (2015). pkCSM: Predicting Small-Molecule Pharmacokinetic and Toxicity Properties Using Graph-Based Signatures.
B Zhang,A Korolj,Bfl Lai,M Radisic (2018). Advances in organ-on-a-chip engineering.
Jae Ryu,Hyun Kim,Sang Lee (2018). Deep learning improves prediction of drug–drug and drug–food interactions.
C Ross,I Swetlitz (2018). Watson Recommends Incorrect Cancer Treatments, System Training Questioned.
D Bhatt,C Mehta (2016). Adaptive Designs for Clinical Trials.
P Shah,F Kendall,S Khozin,R Goosen,J Hu,J Laramie (2019). Artificial intelligence and machine learning in clinical development: a translational perspective.
A Montalbano,C Toumazou (2020). Wearable sensors for real-time clinical trial monitoring: current trends and future opportunities.
A Beam,I Kohane (2018). Big data and machine learning in health care.
A Holzinger,C Biemann,C Pattichis,D Kell (2017). What do we need to build explainable AI systems for the medical domain?.
(2021). Artificial Intelligence and Machine Learning in Software as a Medical Device.
Milena Gianfrancesco,Suzanne Tamang,Jinoos Yazdany,Gabriela Schmajuk (2018). Potential Biases in Machine Learning Algorithms Using Electronic Health Record Data.
W Price,I Cohen (2019). Privacy in the age of medical big data.
E Topol (2019). High-performance medicine: the convergence of human and artificial intelligence.
Doug Meil (2025). IBM Watson Health.
(2021). Artificial Intelligence/Machine Learning-Based Software as a Medical Device (SaMD) Action Plan.
Ema (2020). European Medicines Agency (EMA), The.
Insilico Medicine (2021). AI-designed drug enters clinical trials.
Sarfaraz Niazi,Zamara Mariam (2020). Artificial intelligence in drug development: reshaping the therapeutic landscape.
J Nelson,B Smith,J Jared,J Younger (2011). Prospective trial of real-time electronic surveillance to expedite early care of severe sepsis.
C Tan,H Chen,C Xia (2008). The prediction of cardiovascular disease based on trace element contents in hair and a classifier of boosting decision stumps.
H Chen,C Tan (2011). Prediction of Type-2 diabetes based on several element levels in blood and chemometrics.
Michael Hooper,Lisa Weavind,Arthur Wheeler,Jason Martin,Supriya Gowda,Matthew Semler,Rachel Hayes,Daniel Albert,Norment Deane,Hui Nian,Janos Mathe,Andras Nadas,Janos Sztipanovits,Anne Miller,Gordon Bernard,Todd Rice (2012). Randomized trial of automated, electronic monitoring to facilitate early detection of sepsis in the intensive care unit*.
H Giannini,J Ginestra,C Chivers (2019). A machine learning algorithm to predict severe sepsis and septic shock: development, implementation, and impact on clinical practice.
S Naushad,Janaki Ramaiah,M Pavithrakumari,M (2016). Artificial neural network-based exploration of gene-nutrient interactions in folate and xeno-biotic metabolic pathways that modulate susceptibility to breast cancer.
J Vamathevan,D Clark,P Czodrowski (2019). Applications of machine learning in drug discovery and development.
Erik Gawehn,Jan Hiss,Gisbert Schneider (2016). Deep Learning in Drug Discovery.
Rafael Gómez-Bombarelli,Jennifer Wei,David Duvenaud,José Hernández-Lobato,Benjamín Sánchez-Lengeling,Dennis Sheberla,Jorge Aguilera-Iparraguirre,Timothy Hirzel,Ryan Adams,Alán Aspuru-Guzik (2018). Automatic Chemical Design Using a Data-Driven Continuous Representation of Molecules.
Gisbert Schneider (2018). Automating drug discovery.
R Miotto,L Li,B Kidd,J Dudley (2016). Deep Patient: an unsupervised representation to predict the future of patients from the electronic health records.
A Bate,M Lindquist,I Edwards (1998). A Bayesian neural network method for adverse drug reaction signal generation.
Elena Izmailova,John Wagner,Eric Perakslis (2018). Wearable Devices in Clinical Trials: Hype and Hypothesis.
Y Wang,D Zeng,L Wang (2018). Social media analytics for drug safety signal detection.
Rave Harpaz,Alison Callahan,Suzanne Tamang,Yen Low,David Odgers,Sam Finlayson,Kenneth Jung,Paea Lependu,Nigam Shah (2014). Text Mining for Adverse Drug Events: the Promise, Challenges, and State of the Art.
T Botsis,M Nguyen,E Woo (2011). Text mining for the biopharmaceutical domain: a systematic review of recent advances.
H Luo,J Wang,M Li (2016). Drug repositioning based on comprehensive similarity measures and birandom walk algorithm.
C Carvalho (2021). Functional genomics and CRISPR screens in drug discovery: Synergy with AI.
J Lee (2020). BioBERT: A pre-trained biomedical language representation model for biomedical text mining.
Yadi Zhou,Yuan Hou,Jiayu Shen,Yin Huang,William Martin,Feixiong Cheng (2020). Network-based drug repurposing for novel coronavirus 2019-nCoV/SARS-CoV-2.
H Chen,O Engkvist,Y Wang,M Olivecrona,T Blaschke (2018). The rise of deep learning in drug discovery.
Sean Ekins,Ana Puhl,Kimberley Zorn,Thomas Lane,Daniel Russo,Jennifer Klein,Anthony Hickey,Alex Clark (2019). Exploiting machine learning for end-to-end drug discovery and development.
J Stokes,K Yang,K Swanson (2020). A deep learning approach to antibiotic discovery.
Jessica Vamathevan,Dominic Clark,Paul Czodrowski,Ian Dunham,Edgardo Ferran,George Lee,Bin Li,Anant Madabhushi,Parantu Shah,Michaela Spitzer,Shanrong Zhao (2019). Applications of machine learning in drug discovery and development.
Han Altae-Tran,Bharath Ramsundar,Aneesh Pappu,Vijay Pande (2017). Low Data Drug Discovery with One-Shot Learning.
W Walters,R Barzilay (2021). Applications of deep learning in molecule generation and molecular property prediction.
A Zhavoronkov (2018). Artificial intelligence for drug discovery, biomarker development, and generation of novel chemistry.
M Ragoza,J Hochuli,E Idrobo,J Sunseri,D Koes (2017). Protein-ligand scoring with convolutional neural networks.
M Kuenemann,S Schmidl,A Korotcov (2021). Evaluation of graph neural networks for predicting drug-target interactions.
L Chen,A Cruz,S Ramsey (2020). The rise of generative models in drug design.
Q Liu,H Wu,X Wang (2021). Application of Artificial Intelligence in the Diagnosis and Treatment of COVID-19 Lung Disease.
Junshui Ma,Robert Sheridan,Andy Liaw,George Dahl,Vladimir Svetnik (2015). Deep Neural Nets as a Method for Quantitative Structure–Activity Relationships.
D Elton,Z Boukouvalas,M Fuge,P Chung (2019). Deep learning for molecular design-a review of the state of the art.
No ethics committee approval was required for this article type.
Data Availability
Not applicable for this article.
How to Cite This Article
Dr. Sadia Parveen. 2026. \u201cDrug Development and Discovery Considering Artificial Intelligence: A Through Analysis\u201d. Global Journal of Medical Research - B: Pharma, Drug Discovery, Toxicology & Medicine GJMR-B Volume 25 (GJMR Volume 25 Issue B1).
Explore published articles in an immersive Augmented Reality environment. Our platform converts research papers into interactive 3D books, allowing readers to view and interact with content using AR and VR compatible devices.
Your published article is automatically converted into a realistic 3D book. Flip through pages and read research papers in a more engaging and interactive format.
Our website is actively being updated, and changes may occur frequently. Please clear your browser cache if needed. For feedback or error reporting, please email [email protected]
Thank you for connecting with us. We will respond to you shortly.