| Home | E-Submission | Sitemap | Contact Us |  
top_img
Environ Anal Health Toxicol > Volume 40:2025 > Article
Kim and Choi: Big data and AI: Potential and challenges for digital transformation in toxicology

Abstract

Toxicology has evolved from an observational science to predictive science, driven by advances in computational methods and large-scale data generation. Advances in computing power and the rapid accumulation of toxicological big data have opened new opportunities to modernize chemical risk assessment through artificial intelligence (AI). This study explores the current status of toxicity databases and key methodologies of AI such as machine learning, deep learning, and large language models. The study further examines representative case studies, which leverage AI-based toxicity prediction models in chemical prioritization and others. Despite the advancements, critical challenges remain, including the limited availability of high-quality, homogeneous datasets and the black-box nature of AI models, which hinder regulatory acceptance. To address these issues, this study emphasizes the need for explainable AI and the integration of the adverse outcome pathway framework to enhance model interpretability. By outlining future research directions and advocating for transparent, reproducible AI models, this study contributes to advancing regulatory science, chemical safety assessment, and the broader adoption of AI as new approach methodologies (NAMs) for next generation risk assessment (NGRA).

Introduction

Digital technologies—computers, smartphones, the internet, and other devices that work with digital data—let us store, process, and share information electronically, making them the backbone of digitalization [1]. They power almost every field, from business and education to entertainment, and their influence on healthcare has grown rapidly. In medicine, digitalization appears in telemedicine, where video or phone consultations connect patients and clinicians without a physical visit—especially valuable for people in remote or underserved areas [2]. High-resolution imaging tools such as CT and MRI reveal detailed internal views that guide diagnosis and treatment, while medical robots can dispense drugs or assist in surgery. By integrating these technologies into care delivery and risk assessment, healthcare systems can improve outcomes, boost efficiency, and lower costs.
With the rise of digital technologies, the field of toxicology is moving toward digital toxicology [3]. Traditionally, toxicology has relied on small-scale experimental approaches to assess the adverse effects of chemicals on living organisms. These methods, while scientifically rigorous, are time-consuming, costly, and often limited in their ability to evaluate the growing number of chemicals in modern industrial and environmental contexts. Therefore, new digitalized methods for toxicity testing such as big data [4], artificial intelligence (AI) for toxicity prediction [5], molecular dynamics [6], Physiologically Based Pharmacokinetic modeling [6] are rapidly developed.
Among them, big data and AI have emerged as core digital toxicological techniques in recent years [7]. The proliferation of big data—driven by high-throughput screening (HTS), multi-omics platforms, environmental monitoring, and extensive toxicological databases—creates an unprecedented opportunity for data-driven toxicology. When coupled with AI, the field is shifting toward a more predictive, efficient, and systems-based discipline [8]. Machine learning (ML) algorithms and other AI techniques enable the analysis of high-dimensional datasets, uncovering patterns and relationships that may be invisible to traditional statistical methods [9]. From predicting toxicity of chemicals to identifying biomarkers of exposure, AI empowers researchers to extract actionable insights from big data [10,11].
In response, this review paper explores how big data and AI are used in toxicology, highlighting both current and emerging applications. It also identifies key challenges and future directions for AI-based toxicity prediction, ultimately contributing to the modernization of chemical risk assessment.

Big Data in Toxicology

The term "big data" refers to structured or unstructured datasets that encompass a wide variety of data types and are generated at high speed and in large volumes, often requiring high-performance computing and advanced computational approaches for effective analysis[12]. Large public databases store vast amounts of toxicity data, regulation, exposure, network information which are crucial for data-driven toxicological research (Figure 1).

In vitro mechanism data

Big data for in vitro mechanism data includes high-throughput and high-content screening data, as well as data generated through omics technologies and gene arrays, such as transcriptomics, metabolomics, proteomics, and microbiome studies. These data are often used to identify biomarker or mechanism-based chemical screening in the various studies (Ref). Moreover, one of the prerequisites for applying AI approaches is the availability of large-scale datasets, as a sufficiently large dataset is necessary to develop reliable models without overfitting. The diverse sources of in vitro mechanism data are used to develop toxicity prediction models. For example, MoleculeNet is widely used for training AI-based toxicity prediction models. One of its key characteristics is that, in addition to ensuring data homogeneity and sufficiency, it provides a standardized data format that is easily interpretable by modelers. As AI research in toxicology advances, even datasets that require specialized toxicological expertise are being utilized in increasingly diverse applications. Toxicogenomics datasets, such as those in ToxCast and Open TG-GATEs, are also widely used to develop toxicity prediction models, further enhancing the field of predictive toxicology.

In vivo animal testing data

Traditionally, large numbers of animals have been used to assess the safety of chemicals through testing. The significance of results from these animal studies extends beyond past risk assessments; they remain valuable for validating data from alternative methods, including new approach methodologies (NAMs) [19, 20]. Although animal test results are not the definitive gold standard and have clear limitations in validating NAMs, the current regulatory framework remains largely dependent on animal testing. Thus, comparing animal test results with NAMs-derived data can provide meaningful insights into the validation and refinement of NAMs [21].
Thus, systematically collecting, curating, and structuring existing in vivo experimental data into reference databases is essential. Efforts to make these databases publicly accessible align with the OECD’s Mutual Acceptance of Data (MAD) principle, which aims to reduce redundant testing and promote international data sharing [22].
To support this transition, large-scale initiatives in Europe and the U.S. have been actively developing in vivo reference databases (Table 3). For instance, ToxRefDB, launched by the U.S. Environmental Protection Agency (EPA), compiles in vivo toxicity data from thousands of studies, primarily focusing on pesticide chemicals. By providing key in vivo endpoints such as NOAELs (No Observed Adverse Effect Levels) and LOAELs (Lowest Observed Adverse Effect Levels), ToxRefDB serves as a critical benchmark for evaluating and validating NAM-based toxicity predictions.
Additionally, eChemPortal, a global regulatory toxicology hub managed by the Organisation for Economic Co-operation and Development (OECD), provides worldwide access to in vivo regulatory toxicology studies by integrating datasets from governmental and international agencies. By offering comprehensive information on physicochemical properties, environmental fate, and ecotoxicology, this platform plays a crucial role in harmonizing regulatory frameworks across different regions.

Artificial Intelligence in Toxicology

Various AI algorithms, ranging from conventional machine learning approaches to Large Language Models (LLMs), which have recently gained significant attention across various fields, are actively being used to develop toxicity prediction models(Figure 2). In this study, we provide a brief overview of current methodologies for AI-based toxicity prediction models while highlighting several limitations that need to be addressed.

Machine learning

Early in-silico toxicology relied on hand-crafted molecular descriptors combined with algorithms such as random forests, support vector machines, gradient boosting, and k-nearest neighbors [11]. These models rely on molecular fingerprints and physicochemical properties to predict toxicity endpoints, making them effective for analyzing structured datasets with well-defined chemical properties. One of the key advantages of machine learning in toxicology is its efficiency, as these models can perform well even with relatively small training datasets. Additionally, traditional machine learning models, such as decision trees and random forests, offer greater transparency and interpretability compared to complex deep learning models, making them more accessible for regulatory applications [29].
Despite these advantages, machine learning approaches have several limitations. Many models require manual feature engineering, which can be labor-intensive and may not fully capture complex biological interactions. Moreover, traditional machine learning models struggle with predicting the toxicity of novel chemicals, particularly those that interact with biological systems in ways that are not well represented in the training data. To overcome these limitations, hybrid approaches that integrate machine learning with deep learning and mechanistic toxicology are being explored to enhance predictive accuracy.

Deep learning

Deep learning has emerged as a powerful tool for toxicity prediction, particularly for modeling complex chemical features [30]. Unlike traditional machine learning models, deep learning techniques such as graph neural networks (GNNs) [31], convolutional neural networks (CNNs) [32], and recurrent neural networks (RNNs) [33] can automatically extract features from raw toxicological data without requiring extensive manual preprocessing. These models have shown superior performance in simplified molecular descriptor-based toxicity prediction.
Among deep learning models, GNNs have proven particularly effective for representing molecular structures as mathematical graphs, capturing intricate structural relationships that are difficult to model using conventional QSAR approaches. Additionally, generative models such as autoencoders and generative adversarial networks (GANs) [34] are being used to augment toxicology datasets by synthesizing artificial data to improve AI training.
However, despite their advantages, deep learning methods present challenges that must be addressed for broader adoption in regulatory toxicology. One of the most significant concerns is the "black box" nature of these models, which makes it difficult to understand how they arrive at specific toxicity predictions [35]. This lack of transparency raises concerns among regulatory agencies, as decision-making in chemical risk assessment requires a clear rationale for predictions. Furthermore, deep learning models require large, high-quality datasets for training, which are not always available in toxicology. Addressing these issues requires the development of explainable AI (XAI) techniques [36,37], such as attention mechanisms and Shapley additive explanations (SHAP) [38], which help interpret model predictions and provide insights into the key features driving toxicity assessments.

Generative artificial intelligence

Generative deep-learning frameworks have become a cornerstone of modern molecular design because they internalize the statistical signatures of experimentally characterized compounds and then propose entirely new structures that manifest comparable—or deliberately improved—properties [7]. Early successes came from recurrent-neural-network architectures trained on SMILES strings; when these sequence models are placed under a reinforcement-learning regime the network receives an immediate reward for every token it adds, allowing it to bias generation toward high-throughput-screening libraries that already meet predefined physicochemical envelopes or show predicted affinity for a chosen biological target [39]. Variational auto-encoders extend this idea by mapping molecules into a smooth, differentiable latent space and then performing gradient-guided walks through that space; as a result, they can be steered automatically toward lower lipophilicity, higher solubility or enhanced drug-likeness before decoding the optimized latent vectors back into valid chemical graphs [40,41]. Collectively, these methods not only expand the explorable chemical universe by orders of magnitude but also compress the design–make–test–learn cycle, enabling chemists to front-load toxicity considerations and converge more rapidly on candidate molecules with balanced profiles of potency, safety, and manufacturability.

Large language model (LLM)

Large language models (LLMs), such as GPT-4, BERT, and BioBERT, are increasingly being utilized in toxicology for tasks ranging from literature mining to predictive modeling [42]. These models have demonstrated the ability to extract toxicological insights from vast scientific corpora, enabling automated data curation and knowledge discovery. One of the key applications of LLMs in toxicology is their use in analyzing and summarizing toxicity-related research, facilitating the integration of diverse sources of information for risk assessment. LLMs can also generate molecular descriptions and predict toxicological profiles based on natural language descriptions of chemical properties, providing an alternative approach to structure-based modeling [43].While LLMs offer new opportunities for toxicological research, they also present challenges. One of the primary concerns is the potential for generating false or misleading information, commonly referred to as the "hallucination problem." [44]. Since LLMs rely on probabilistic language generation rather than direct chemical-mechanistic reasoning, their outputs may not always be scientifically accurate. Additionally, while these models excel in text-based knowledge retrieval, they are less effective in making precise numerical toxicity predictions compared to structured AI models. To mitigate these limitations, it is needed to explore hybrid approaches that integrate LLMs with structured toxicological databases, allowing for improved accuracy and regulatory compliance.

Application of Artificial Intelligence in Toxicology

Chemical prioritization and next generation risk assessment

The ultimate goal of developing AI-based toxicity prediction models is to apply them for chemical screening and prioritization while reducing unnecessary animal testing. Although AI-driven toxicity prediction models inherently have some level of uncertainty, this is a common challenge in all modeling approaches. As the well-known saying in modeling states, "all models are wrong, but some are useful," the key is to effectively leverage these models by incorporating accurate uncertainty quantification and evaluation [45]. Ensuring that these models provide reliable toxicity predictions despite their limitations is crucial for regulatory acceptance and practical application in chemical risk assessment.
In this study, we analyzed several representative research efforts that apply AI-based toxicity prediction models for chemical prioritization (Table 3). These studies utilized AI-driven approaches to assess various chemicals, including consumer product chemicals, microplastics, diesel particulate matter, per- and polyfluoroalkyl substances (PFAS), bisphenols, and nanoparticles. The applicability domain of AI-driven toxicity prediction has been steadily expanding, moving beyond organic compounds to include inorganic chemicals and nanomaterials, demonstrating the increasing versatility of these predictive models [46].
Currently, AI-based toxicity models are primarily used as low-tier screening tools within risk assessment frameworks. While these in silico approaches are not yet considered standalone replacements for traditional experimental methods, they enable the assessment of thousands of chemicals efficiently. By prioritizing chemicals for further experimental evaluation, AI-based models contribute to reducing the number of required in vivo and in vitro tests, thereby improving efficiency in chemical safety assessment.

Emerging applications of artificial intelligence in toxicology

Recent advances in artificial intelligence are opening new frontiers in toxicology beyond advancingtoxicity prediction models. Key areas include (1) large language models for literature mining and knowledge extraction, (2) image-based toxicity analysis using deep learning, and (3) precision toxicology integrating genomic and exposure data. Below, we review each of these areas, summarizing state-of-the-art approaches and findings.

Large language models for literature mining and knowledge extraction

In many cases, safety information about chemicals scattered, thus poses a challenge for risk assessors to synthesize evidence. One application is using LLMs to assist safety knowledge extraction from unstructured text. For example, the European Food Safety Authority’s AI4NAMS project evaluated GPT-based models for rapid evidence retrieval on chemicals of concern [53,54]. In a case study on Bisphenol A, a baseline GPT-3 model was compared to a fine-tuned version that had been trained on domain-specific toxicology data. The fine-tuned model substantially outperformed the generic model in accurately extracting and consolidating toxicological information from scientific publications. This demonstrates that with domain tuning, LLMs can reliably identify key toxicity findings (e.g., dose–response data, hazard classifications) and summarize them, thereby supporting risk assessors in reviewing evidence. LLMs are also being explored for automating systematic reviews and regulatory document preparation. Given a large set of study reports or publications, an LLM can be prompted to output structured summaries, highlight critical endpoints, and even draft sections of risk assessment reports. Early demonstrations suggest that a regulatory-focused LLM could significantly streamline tasks like compiling dossiers under REACH, by automatically extracting toxicity data from both structured databases (e.g., PubChem, ToxCast) and unstructured sources (journal articles, study reports). Importantly, the role of these models is to assist experts rather than replace them – for instance, flagging relevant studies and pulling out key results for human evaluation. Nonetheless, the efficiency gains are clear: an AI co-worker that can read and distill thousands of pages of toxicology information in minutes enables scientists to focus on interpretation and decision-making. However, challenges remain in ensuring accuracy (avoiding AI “hallucinations” or misinterpretation of data). A recent study demonstrated that introducing as little as 0.001% of false information into a training dataset could cause LLMs to generate medically harmful content while still passing standard performance benchmarks [55]. This phenomenon, known as data poisoning, raises serious concerns about LLMs trained on unverified toxicological sources, where subtle misinformation could lead to erroneous safety assessments or misclassification of chemical hazards. These risks highlight the urgent need for external validation, provenance tracking, and the use of curated knowledge when deploying LLMs in domains like toxicology that require high levels of reliability and accountability.In summary, LLMs offer a transformative approach to literature mining in toxicology, from building knowledge of chemical–effect relationships to drafting evidence-based assessments. As these models continue to improve and incorporate toxicology-specific training, they are expected to become invaluable for keeping pace with the ever-growing scientific knowledge base in environmental health.

Image-based toxicity analysis using deep learning

AI is also being harnessed to interpret complex biological images in toxicology. Deep learning models, particularly convolutional neural networks (CNNs), can automatically recognize morphological patterns indicative of toxicity in microscopy and histopathology images. This is enabling high-throughput image-based toxicity assays that historically required laborious human scoring. A prime example is the comet assay for DNA damage, where individual cell nuclei exhibit “comet tails” when DNA strand breaks occur. Recent studies applied CNNs to classify comet assay images and achieved accuracy levels on par with expert analysis [56]. In one study, a CNN was trained on ~800 labeled comet images (augmented to nearly 10,000) and learned to categorize DNA damage into four severity classes (healthy to highly damaged) with 96.1% overall accuracy. This automated image analysis removes the bottleneck of manual scoring and adds consistency in genotoxicity evaluation. Similarly, deep learning has been introduced to toxicologic histopathology – the examination of tissue slides from animal studies [57]. Whole-slide imaging and CNN-based image analysis are poised to assist pathologists by detecting microscopic lesions or classifying tissue abnormalities caused by toxicants. For instance, prototype deep neural networks have been trained on thousands of annotated histology images from rodent toxicity studies to learn features of “normal” vs. diseased tissue, facilitating objective identification of treatment-related changes. These approaches have demonstrated the potential to improve workflow efficiency and quantitative rigor in pathology assessment. Early applications include automated identification of liver and kidney lesions and mapping their distribution across a slide, which can help flag subtle toxic effects that might be missed by the human eye. In high-content cell imaging assays, deep learning-based phenotypic profiling is also used to predict toxic mechanisms. Imaging platforms like Cell Painting, which capture multichannel cellular morphology, combined with deep neural networks can cluster compounds by phenotypic signatures and even predict modes of action or toxicity endpoints [58]. For example, AI models integrating Cell Painting images have successfully distinguished specific forms of cell death (e.g., apoptosis vs. ferroptosis) and improved detection of mitochondrial toxicity by combining image features with gene expression data [59,60]. Overall, deep learning is enabling image-based toxicology to move from qualitative observations to quantitative, scalable analysis. By extracting rich feature patterns from cell and tissue images, these models enhance our ability to detect toxicity earlier (in vitro) and more reliably (in vivo), supporting more rapid and mechanistically anchored safety evaluations.

Precision toxicology integrating genomic and exposure data

Another frontier is the integration of toxicogenomic data and personal exposure information to enable precision or personalized toxicology. Just as precision medicine tailors treatments to a patient’s genetic makeup, precision toxicology aims to predict individual risk and susceptibilities to chemical exposures by considering omics data (genomics, transcriptomics, metabolomics) alongside environmental context. AI and machine learning are critical to analyze these high-dimensional datasets and uncover biomarker signatures predictive of toxicity. One approach uses toxicogenomic profiles (e.g. gene expression changes or protein biomarkers in cells exposed to chemicals) to forecast adverse outcomes in vivo. For example, Rahman et al. (2022) employed machine learning to identify an optimal panel of DNA damage response proteins that serve as predictors of genotoxicity [61]. By applying feature-selection algorithms to in vitro transcriptomic data, they discovered that only a handful of biomarkers (five genes/proteins) could accurately predict rodent carcinogenicity and Ames test mutagenicity (with ≥ 70% accuracy for both). Notably, these top biomarkers were all involved in conserved DNA repair pathways, providing mechanistic insight. This illustrates how integrating omics endpoints with AI can bridge molecular-level data to whole-animal outcomes, improving predictivity while reducing testing. Another pioneering strategy uses generative AI to simulate toxicogenomic responses. Chen et al. (2022) developed a GAN-based model (“Tox-GAN”) trained on a large toxicogenomics database of rat studies, which can generate in vivo gene expression profiles for untested chemicals across dose levels [62]. The synthetic profiles showed > 87% agreement with actual data in terms of affected pathways, pointing to a future where animal experiments could be augmented or partially replaced by AI-simulated omics data.
A hallmark of precision toxicology is accounting for inter-individual variability. Factors such as genetic polymorphisms, pre-existing health conditions, microbiome, and cumulative exposures can cause people to respond differently to the same chemical. AI-driven models are starting to integrate these layers. For instance, researchers are combining human genomic data (e.g. risk alleles), personal exposure biomarkers, and toxicogenomic readouts to build personalized risk assessment models. By incorporating individual-level data, these models can stratify populations into more or less susceptible groups and predict toxic responses with higher resolution. As described by Singh et al. (2023), an AI/ML precision toxicology framework might integrate a person’s omics profile (gene or protein expression signatures indicating sensitivity) with their exposure history to tailor risk estimates [63]. Such models have been piloted in studies of air pollutants and occupational chemicals, where omics-based biomarkers of exposure and early effect are used to forecast long-term health outcomes in specific subpopulations. Likewise, environmental epidemiology is adopting multi-omics and AI to identify vulnerable cohorts – for example, using gene expression changes in blood as sentinels of chemical exposure in a community, then applying machine learning to link those changes with health records and personal exposure measures. The PrecisionTox initiative (EU) is one large-scale effort applying such integrative approaches across species to improve cross-species extrapolation and human relevance of toxicity findings [64]. Overall, precision toxicology represents a convergence of big data from toxicogenomics, exposomics, and human genetics. AI methods are essential to sift through these data and identify the combinations of molecular markers that best predict adverse outcomes in defined populations or even individuals. The outcome is a more tailored risk assessment, where instead of a one-size-fits-all safety threshold, we might predict a range of responses and safeguard those most at risk. Though still in its infancy, this approach holds promise for more equitable and effective public health protection, as interventions could be directed to high-risk groups identified by their molecular and environmental profiles. In coming years, we can expect AI-fueled models to increasingly inform regulatory science by integrating diverse data streams – from genomics to wearable sensors – thus refining toxicity predictions to reflect real-world human variability.

Regulatory Applications of Toxicity Big Data and Artificial Intelligence

In an exceptional display of regulatory coordination, three leading U.S. agencies signaled an important turn away from animal models in the spring of 2025 [65]. On 10 April, FDA Commissioner Martin A. Makary released a Preclinical Safety Roadmap that commits the Agency to sharply decrease animal use, citing evidence that > 90 % of drugs successful in animal studies ultimately fail in humans because of unforeseen efficacy or safety issues. That same day, EPA Administrator Lee Zeldin revived—and strengthened—the Agency’s hidden guarantee to eliminate vertebrate testing by 2035, extending the target to both industrial chemicals and pesticides. Nineteen days later, NIH Director Dr Jay Bhattacharya announced the creation of the Office of Research Innovation, Validation and Application (ORIVA), a hub charged with developing, validating and mainstreaming non-animal NAMs across the NIH research portfolio.
Within this context, several regulatory agencies have published frameworks and guidance to support the integration of AI and big data into chemical risk assessment and regulatory decision-making. These efforts reflect a growing recognition of AI’s potential to enhance chemical safety evaluations, reduce reliance on animal testing, and improve the predictive accuracy and efficiency of toxicological assessments. For instance, the European Food Safety Authority (EFSA) has developed an extensive framework through the AI4NAMS initiative (Artificial Intelligence for New Approach Methodologies), published in 2024 [54]. This report outlines a seven-step AI workflow that includes literature search, AI-based ranking, structuring, harmonization, data extraction, and integration with adverse outcome pathways (AOPs). EFSA provides 96 technical recommendations across domains such as ontology-based term normalization, automated picklist generation for reporting templates (OHT201), and the use of LLMs for data extraction with embedded reliability scoring. These efforts are aligned with the SPIDO NAMs Roadmap, which outlines EFSA’s plan to transition the majority of chemical safety assessments to NAMs by 2027, supported by AI-enabled dashboards, repositories, and knowledge graphs. The U.S. Food and Drug Administration (FDA) has also explored the use of AI in regulatory science. Its 2023 draft guidance, Considerations for the Use of Artificial Intelligence to Support Regulatory Decision-Making for Drug and Biological Products, recommends a risk-based approach for evaluating the reliability of AI-generated outputs throughout the product lifecycle. It emphasizes methodological transparency, model documentation, performance monitoring, and change management procedures. The European Medicines Agency (EMA) has adopted a holistic and strategic approach to AI integration. Its Artificial Intelligence Workplan lays out a roadmap for embedding AI across regulatory domains, including pharmacovigilance and product evaluation. Complementing this, the 2024 guiding principles on the use of LLMs provide ethical and operational guidance for using advanced AI in regulatory science, highlighting issues such as data governance, bias mitigation, and continuous learning. In parallel, the OECD continues to lead in harmonizing international standards for computational toxicology. The QSAR Assessment Framework (QAF), originally published in 2007 and periodically updated, offers detailed criteria for evaluating the scientific validity and regulatory acceptability of predictive models. The OECD’s ongoing initiatives increasingly focus on transparent documentation, applicability domain (AD) definition, and cross-agency interoperability of AI tools. While these agencies share a commitment to transparency, risk assessment, and scientific rigor, their strategies differ in emphasis. EFSA focuses on AI-driven automation of hazard identification using real-world toxicological data, with a strong emphasis on workflow standardization and ontology harmonization. FDA’s approach centers on model credibility and risk management within regulated product development. EMA promotes a governance-based model that integrates policy, ethics, and stakeholder engagement. OECD, by contrast, supports international alignment of computational models through quality assurance frameworks and collaborative testing. Together, these regulatory efforts underscore the increasing role of AI and big data in toxicology and chemical management. As AI tools continue to evolve, sustained guidance, robust validation practices, and harmonized policy development will be critical to ensuring the responsible, reproducible, and ethical use of AI in regulatory science.

Challenges and Future Perspectives

The integration of big data and AI in toxicology presents significant opportunities for improving chemical risk assessment and reducing reliance on animal testing. However, several challenges must be addressed to fully realize the potential of these technologies. Two critical obstacles are (1) the limited availability of high-quality data and (2) the black box nature of AI models, both of which hinder regulatory acceptance and broader application in toxicology.

Limited high-quality and quantity of data

One of the primary challenges in AI-driven toxicology is the limited availability of high-quality and homogeneous datasets. Although vast amounts of toxicological data exist, they are often fragmented, inconsistent, or inaccessible due to proprietary restrictions. Many datasets originate from diverse sources, but differences in study design, data formats, and reporting standards make integration difficult and inefficient. This gap in data completeness limits the effectiveness of predictive models, as AI algorithms require rich, well-annotated datasets to achieve high reliability and generalizability.
To address this issue, efforts should focus on harmonizing data formats, improving data-sharing frameworks, and increasing data accessibility. These initiatives align with the OECD MAD System, which ensures that a test performed in one country is accepted in over 40 others, thereby reducing duplicative testing and enhancing international regulatory collaboration. Additionally, collaborative initiatives involving regulatory agencies, academia, and industry, such as the Accelerating the Pace of Chemical Risk Assessment (APCRA) project and the Toxicology in the 21st Century (Tox21) program, aim to improve data quality, availability, and integration across different platforms [19].
As advanced data processing capabilities continue to evolve, it is critical to establish effective strategies for managing, structuring, and sharing existing toxicological data. Developing unified data standards and centralized repositories will be essential for ensuring that AI-driven toxicology can be applied effectively and transparently in chemical risk assessment and regulatory decision-making.

The black box problem

Another significant challenge in applying AI to toxicology is the black box nature of many machine learning models, particularly deep learning algorithms. These models often lack interpretability and transparency, making it difficult for regulatory agencies and toxicologists to trust their predictions. Unlike traditional QSAR models, which provide intuitive prediction mechanisms, many AI-driven approaches generate output without clearly explaining how or why a specific toxicity prediction was made.
This lack of interpretability raises concerns in regulatory decision-making, risk assessment, and hazard identification, where explainability is crucial for ensuring scientific credibility. Without a clear understanding of how AI models arrive at their conclusions, regulatory bodies may hesitate to fully integrate AI-based predictions into risk assessment frameworks.
To overcome this challenge, researchers are actively developing explainable AI (XAI) techniques, such as attention mechanisms, feature attribution methods, and rule-based models, to improve model transparency. To address this challenge, explainable AI (XAI) techniques such as SHAP (Shapley Additive Explanations) and LIME (Local Interpretable Model-agnostic Explanations) have been introduced in toxicological research. These methods provide feature attribution scores, helping researchers identify which molecular descriptors or biological signals contributed most to a given prediction. For instance, SHAP has been used to interpret models trained on Tox21 data, offering insights into the physicochemical properties that drive compound activity in bioassays. Similarly, LIME has helped evaluate the local fidelity of predictions in QSAR toxicity classification tasks. However, while these methods represent meaningful progress toward model transparency, they remain partial solutions. They typically offer post hoc explanations that approximate the model’s logic rather than fully capturing its internal reasoning, and their reliability may vary depending on model complexity and input perturbation. Therefore, we believe that these techniques, though valuable, are not yet sufficient for establishing fundamental trust in high-stakes applications such as regulatory toxicology. To build more trustworthy AI systems, future research should focus on inherently interpretable models and hybrid approaches that incorporate mechanistic toxicological knowledge—such as the adverse outcome pathway (AOP) framework—directly into model design [11]. These efforts will be critical in developing reproducible, scientifically justifiable AI models capable of gaining regulatory acceptance and broader adoption in the field.

Moving forward

Despite these challenges, big data and AI have the potential to revolutionize toxicology by enabling faster, more cost-effective, and ethically responsible toxicity assessments. Addressing data limitations through standardization, improved data sharing, and data augmentation techniques will enhance the quality and quantity of available toxicological datasets. Additionally, advancing explainable AI approaches will help bridge the gap between machine learning predictions and regulatory requirements.
By fostering global collaboration, investing in data infrastructure, and developing interpretable AI models, the field of toxicology can fully harness the power of AI to improve chemical safety assessments, minimize reliance on animal testing, and advance the adoption of NAMs in regulatory science (Figure 3).

Notes

Acknowledgement
This work was supported by the Korean Ministry of Environment under the ‘Environmental Health R&D Program’ (2021003310005).
Conflict of interest
Please state any conflict of interest.
CRediT author statement
DK: Investigation, Formal analysis, Writing - Original Draft, Writing - Reviewing and Editing; JC: Conceptualization, Writing - Original Draft, Writing - Reviewing and Editing.

References

1. Vial G. Understanding digital transformation: A review and a research agenda. The Journal of Strategic Information Systems 2019;28(2):119-144 https://doi.org/10.1016/j.jsis.2019.01.003.
crossref
2. Singh AV, Bansod G, Mahajan M, Dietrich P, Singh SP, Rav K, et al. Digital transformation in toxicology: Improving communication and efficiency in risk assessment. ACS Omega 2023;8(24):21377-21390 https://doi.org/10.1021/acsomega.3c00596.
crossref pmid pmc
3. Haleem A, Javaid M, Singh RP, Suman R. Telemedicine for healthcare: Capabilities, features, barriers, and applications. Sens Int 2021;2: 100117 https://doi.org/10.1016/j.sintl.2021.100117.
crossref pmid pmc
4. Richarz AN. Big data in predictive toxicology: Challenges, opportunities and perspectives. In: Dhawan A, editor. Issues in toxicology. Royal Society of Chemistry; 2020. 1–37.

5. Cavasotto CN, Scardino V. Machine learning toxicity prediction: Latest advances by toxicity end point. ACS Omega 2022;7(51):47536-47546 https://doi.org/10.1021/acsomega.2c05693.
crossref pmid pmc
6. Upton RN, Foster DJR, Abuhelwa AY. An introduction to physiologically-based pharmacokinetic models. Paediatr Anaesth 2016;26: 1036-1046 https://doi.org/10.1111/pan.12995.
crossref pmid
7. Méndez-Lucio O, Baillif B, Clevert DA, Rouquié D, Wichard J. De novo generation of hit-like molecules from gene expression signatures using artificial intelligence. Nat Commun 2020;11: 10 https://doi.org/10.1038/s41467-019-13807-w.
crossref pmid pmc
8. Luechtefeld T, Rowlands C, Hartung T. Big-data and machine learning to revamp computational toxicology and its use in risk assessment. Toxicology Research 2018;7: 732-744 https://doi.org/10.1039/c8tx00051d.
crossref pmid pmc
9. Samanta IS, Mohanty S, Parida SM, Rout PK, Panda S, Bajaj M, et al. Artificial intelligence and machine learning techniques for power quality event classification: a focused review and future insights. Results in Engineering 2025;25: 103873 https://doi.org/10.1016/j.rineng.2024.103873.
crossref
10. Lin Z, Chou WC. Machine learning and artificial intelligence in toxicological sciences. Toxicol Sci 2022;189(1):7-19 https://doi.org/10.1093/toxsci/kfac075.
crossref pmid pmc
11. Jeong J, Choi J. Artificial intelligence-based toxicity prediction of environmental chemicals: Future directions for chemical management applications. Environ Sci Technol 2022;56(12):7532-7543 https://doi.org/10.1021/acs.est.1c07413.
crossref pmid
12. Kumar Y, Marchena J, Awlla AH, Li JJ, Abdalla HB. The AI-powered evolution of big data. Applied Sciences 2024;14(22):10176 https://doi.org/10.3390/app142210176.
crossref
13. Gaulton A, Bellis LJ, Bento AP, Chambers J, Davies M, Hersey A, et al. ChEMBL: A large-scale bioactivity database for drug discovery. 2012;40(Database issue):D1100-D1107 https://doi.org/10.1093/nar/gkr777.
crossref
14. Lamb J, Crawford ED, Peck D, Modell JW, Blat IC, Wrobel MJ, et al. The Connectivity Map: using gene-expression signatures to connect small molecules, genes, and disease. Science 2006;313(5795):1929-1935 https://doi.org/10.1126/science.1132939.
crossref pmid
15. Richard AM, Judson RS, Houck KA, Grulke CM, Volarath P, Thillainadarajah I, et al. ToxCast chemical landscape: paving the road to 21st Century toxicology. Chem Res Toxicol 2016;29(8):1225-1251.
crossref pmid pdf
16. Wu Z, Ramsundar B, Feinberg EN, Gomes J, Geniesse C, Pappu AS, et al. MoleculeNet: A benchmark for molecular machine learning. Chem Sci 2018;9(2):513-530 https://doi.org/10.1039/c7sc02664a.
crossref pmid
17. Igarashi Y, Nakatsu N, Yamashita T, Ono A, Ohno Y, Urushidani T, et al. Open TG-GATEs: A large-scale toxicogenomics database. Nucleic Acids Res 2015;43(Database issue):D921-D927 https://doi.org/10.1093/nar/gku955.
crossref pmid
18. Kim S, Chen J, Cheng T, Gindulyte A, He J, He S, et al. PubChem in 2021: New data content and improved web interfaces. Nucleic Acids Res 2021;49(D1):D1388-D1395 https://doi.org/10.1093/nar/gkaa971.
crossref pmid
19. Kim D, Choi J. Perspective of Next Generation Risk Assessment (NGRA) using New Approach Methodologies (NAMs): Review on Accelerating the Pace of Chemical Risk Assessment (APCRA) initiative. Journal of the Korean Chemical Society 2023;67(1):19-35 https://doi.org/10.5012/jkcs.2023.67.1.19 [Korean].
crossref
20. European Chemicals Agency. New approach methodologies in regulatory science - Proceedings of a scientific workshop, Helsinki, 19–20 April 2016; [cited 2025 Mar 11]. Available from: https://data.europa.eu/doi/10.2823/543644.

21. Thomas RS, Bahadori T, Buckley TJ, Cowden J, Deisenroth C, Dionisio KL, et al. The next generation blueprint of computational toxicology at the U.S. Environmental Protection Agency. Toxicol Sci 2019;169(2):317-332 https://doi.org/10.1093/toxsci/kfz058.
crossref pmid
22. Rasmussen K, Rauscher H, Kearns P, González M, Riego Sintes J. Developing OECD test guidelines for regulatory testing of nanomaterials to ensure mutual acceptance of test data. Regul Toxicol Pharmacol 2019;104: 74-83 https://doi.org/10.1016/j.yrtph.2019.02.008.
crossref pmid pmc
23. Judson R, Richard A, Dix D, Houck K, Elloumi F, Martin M, et al. ACToR - Aggregated Computational Toxicology Resource. Toxicol Appl Pharmacol 2008;233(1):7-13 https://doi.org/10.1016/j.taap.2007.12.037.
crossref pmid
24. Kim D, Cho S, Jeon JJ, Choi J. Inhalation toxicity screening of consumer products chemicals using OECD Test Guideline Data-based machine learning models. J Hazard Mater 2024;478: 135446 https://doi.org/10.1016/j.jhazmat.2024.135446.
crossref pmid
25. Feshuk M, Kolaczkowski L, Watford S, Friedman KP. ToxRefDB v2.1: update to curated in vivo study data in the Toxicity Reference Database. Front Toxicol 2023;5: 1260305 https://doi.org/10.3389/ftox.2023.1260305.
crossref pmid pmc
26. Williams AJ, Grulke CM, Edwards J, McEachran AD, Mansouri K, Baker NC, et al. The CompTox Chemistry Dashboard: A community data resource for environmental chemistry. J Cheminform 2017;9(1):61 https://doi.org/10.1186/s13321-017-0247-6.
crossref pmid pmc
27. Bitsch A, Jacobi S, Melber C, Wahnschaffe U, Simetska N, Mangelsdorf I. REPDOSE: A database on repeated dose toxicity studies of commercial chemicals–A multifunctional tool. Regul Toxicol Pharmacol 2006;46(2):202-10 https://doi.org/10.1016/j.yrtph.2006.05.013.
crossref pmid
28. Wexler P. TOXNET: an evolving web resource for toxicology and environmental health information. Toxicology 2001;157(1-2):3-10 https://doi.org/10.1016/s0300-483x(00)00337-1.
crossref pmid
29. Kim D, Jeong J, Choi J. Identification of optimal machine learning algorithms and molecular fingerprints for explainable toxicity prediction models using ToxCast/Tox21 bioassay data. ACS Omega 2024;9(36):37934-37941 https://doi.org/10.1021/acsomega.4c04474.
crossref pmid pmc
30. LeCun Y, Bengio Y, Hinton G. Deep learning. Nature 2015;521(7553):436-444 https://doi.org/10.1038/nature14539.
crossref pmid
31. Zhou J, Cui G, Hu S, Zhang Z, Yang C, Liu Z, et al. Graph neural networks: A review of methods and applications. AI Open 2020;1(1):57-81 https://doi.org/10.1016/j.aiopen.2021.01.001.
crossref
32. Yamashita R, Nishio M, Do RKG, Togashi K. Convolutional neural networks: an overview and application in radiology. Insights Imaging 2018;9(4):611-629 https://doi.org/10.1007/s13244-018-0639-9.
crossref pmid pmc
33. Che Z, Purushotham S, Cho K, Sontag D, Liu Y. Recurrent Neural Networks for multivariate time series with missing values. Sci Rep 2018;8(1):6085 https://doi.org/10.1038/s41598-018-24271-9.
crossref pmid pmc
34. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative adversarial networks. Communications of the ACM 2020;63(11):139-144 https://doi.org/10.1145/3422622.
crossref
35. Hassija V, Chamola V, Mahapatra A, Singal A, Goel D, Huang K, et al. Interpreting black-box models: a review on explainable artificial intelligence. Cognitive Computation 2024;16(1):45-74 https://doi.org/10.1007/s12559-023-10179-8.
crossref
36. Linardatos P, Papastefanopoulos V, Kotsiantis S. Explainable AI: a review of machine learning interpretability methods. Entropy (Basel) 2020;23(1):18 https://doi.org/10.3390/e23010018.
crossref pmid pmc
37. Ali S, Abuhmed T, El-Sappagh S, Muhammad K, Alonso-Moral JM, Confalonieri R, et al. Explainable Artificial Intelligence (XAI): what we know and what is left to attain Trustworthy Artificial Intelligence. Information Fusion 2023;99(3):101805 https://doi.org/10.1016/j.inffus.2023.101805.
crossref
38. Nohara Y, Matsumoto K, Soejima H, Nakashima N. Explanation of machine learning models using shapley additive explanation and application for real data in hospital. Comput Methods Programs Biomed 2022;214: 106584 https://doi.org/10.1016/j.cmpb.2021.106584.
crossref pmid
39. Popova M, Isayev O, Tropsha A. Deep reinforcement learning for de novo drug design. Sci Adv 2018;4(7):eaap7885. https://doi.org/10.1126/sciadv.aap7885.
crossref pmid pmc
40. Kadurin A, Nikolenko S, Khrabrov K, Aliper A, Zhavoronkov A. druGAN: An advanced generative adversarial autoencoder model for de novo generation of new molecules with desired molecular properties in silico. Mol Pharm 2017;14(9):3098-3014 https://doi.org/10.1021/acs.molpharmaceut.7b00346.
crossref pmid
41. Gómez-Bombarelli R, Wei JN, Duvenaud D, Hernández-Lobato JM, Sánchez-Lengeling B, Sheberla D, et al. Automatic chemical design using a data-driven continuous representation of molecules. ACS Cent Sci 2018;4(3):268-276 https://doi.org/10.1021/acscentsci.7b00572.
crossref pmid pmc
42. Chen Z, Mao H, Li H, Jin W, Wen H, Wei X, et al. Exploring the potential of large language models (LLMs) in learning on graphs. ACM SIGKDD Explorations Newsletter 2024;25(2):42-61 https://doi.org/10.1145/3655103.3655110.
crossref
43. Yang H, Xiu J, Yan W, Liu K, Cui H, Wang Z, et al. Large language models as tools for molecular toxicity prediction: AI insights into cardiotoxicity. Journal of Chemical Information and Modeling 2025;65(5):2268-2282 https://doi.org/10.1021/acs.jcim.4c01371.
crossref pmid
44. Sun Y, Sheng D, Zhou Z, Wu Y. AI hallucination: Towards a comprehensive classification of distorted information in artificial intelligence-generated content. Humanities and Social Sciences Communications 2024;11: 1278 https://doi.org/10.1057/s41599-024-03811-x.
crossref
45. Curchoe CL. All models are wrong, but some are useful. J Assist Reprod Genet 2020;37(10):2389-2391 https://doi.org/10.1007/s10815-020-01895-3.
crossref pmid pmc
46. Ahmadi M, Ayyoubzadeh SM, Ghorbani-Bidkorpeh F. Toxicity prediction of nanoparticles using machine learning approaches. Toxicology 2024;501: 153697 https://doi.org/10.1016/j.tox.2023.153697.
crossref pmid
47. Jeong J, Choi J. Development of AOP relevant to microplastics based on toxicity mechanisms of chemical additives using ToxCast™ and deep learning models combined approach. Environment International 2020;137: 105557 https://doi.org/10.1016/j.envint.2020.105557.
crossref pmid
48. Vignaux PA, Lane TR, Urbina F, Gerlach J, Puhl AC, Snyder SH, et al. Validation of acetylcholinesterase inhibition machine learning models for multiple species. Chem Res Toxicol 2023;36(2):188-201 https://doi.org/10.1021/acs.chemrestox.2c00283.
crossref pmid pmc
49. Singam ERA, Tachachartvanich P, Fourches D, Soshilov A, Hsieh JCY, La Merrill MA, et al. Structure-based virtual screening of perfluoroalkyl and polyfluoroalkyl substances (PFASs) as endocrine disruptors of androgen receptor activity using molecular docking and machine learning. Environ Res 2020;190: 109920 https://doi.org/10.1016/j.envres.2020.109920.
crossref pmid
50. Zhao Y, Zhang X, Zhang Z, Huang W, Tang M, Du G, et al. Hepatic toxicity prediction of bisphenol analogs by machine learning strategy. Sci Total Environ 2024;934: 173420 https://doi.org/10.1016/j.scitotenv.2024.173420.
crossref pmid
51. Jeong J, Bae SY, Choi J. Identification of toxicity pathway of diesel particulate matter using AOP of PPARγ inactivation leading to pulmonary fibrosis. Environ Int 2021;147: 106339 https://doi.org/10.1016/j.envint.2020.106339.
crossref pmid
52. Wu S, Li SX, Qiu J, Zhao HM, Li YW, Feng NX, et al. Accurate prediction of rat acute oral toxicity and reference dose for thousands of polycyclic aromatic hydrocarbon derivatives based on chemometric QSAR and machine learning. Environ Sci Technol 2024;58(34):15100-15110 https://doi.org/10.1021/acs.est.4c03966.
crossref
53. Sonnenburg A, van der Lugt B, Rehn J, Wittkowski P, Bech K, Padberg F, et al. Artificial intelligence-based data extraction for next generation risk assessment: Is fine-tuning of a large language model worth the effort? Toxicology 2024;508: 153933 https://doi.org/10.1016/j.tox.2024.153933.
crossref pmid
54. Blümmel T, Rehn J, Mereu C, Graf F, Bazing F, Kneuer C, et al. Exploring the use of Artificial Intelligence (AI) for extracting and integrating data obtained through New Approach Methodologies (NAMs) for chemical risk assessment. EFSA Supporting Publication 2024;21(1):8567E https://doi.org/10.2903/sp.efsa.2024.EN-8567.
crossref
55. Alber DA, Yang Z, Alyakin A, Yang E, Rai S, Valliani AA, et al. Medical large language models are vulnerable to data-poisoning attacks. Nature Medicine 2025;31(2):618-626 https://doi.org/10.1038/s41591-024-03445-1.
crossref pmid pmc
56. Atila Ü, Baydilli YY, Sehirli E, Turan MK. Classification of DNA damages on segmented comet assay images using convolutional neural network. Comput Methods Programs Biomed 2020;186: 105192 https://doi.org/10.1016/j.cmpb.2019.105192.
crossref pmid
57. Kriegsmann M, Kriegsmann K, Steinbuss G, Zgorzelski C, Albrecht T, Heinrich S, et al. Implementation of deep learning in liver pathology optimizes diagnosis of benign lesions and adenocarcinoma metastasis. Clin Transl Med 2023;13(7):1299 https://doi.org/10.1002/ctm2.1299.
crossref
58. Willis C, Nyffeler J, Harrill J. Phenotypic profiling of reference chemicals across biologically diverse cell types using the Cell Painting assay. SLAS Discovery 2020;25(7):755-769 https://doi.org/10.1177/247255522.
crossref pmid
59. Schorpp K, Bessadok A, Biibosunov A, Rothenaigner I, Strasser S, Peng T, et al. CellDeathPred: a deep learning framework for ferroptosis and apoptosis prediction based on cell painting. Cell Death Discov 2023;9(1):277 https://doi.org/10.1038/s41420-023-01559-y.
crossref pmid pmc
60. Seal S, Carreras-Puigvert J, Trapotsi MA, Yang H, Spjuth O, Bender A. Integrating cell morphology with gene expression and chemical structure to aid mitochondrial toxicity detection. Commun Biol 2022;5(1):858 https://doi.org/10.1038/s42003-022-03763-5.
crossref pmid pmc
61. Rahman SM, Lan J, Kaeli D, Dy J, Alshawabkeh A, Gu AZ. Machine learning-based biomarker identification from toxicogenomics – bridging to regulatory relevant phenotypic endpoints. J Hazard Mater 2022;423(Pt B):127141 https://doi.org/10.1016/j.jhazmat.2021.127141.
crossref pmid
62. Chen X, Roberts R, Tong W, Liu Z. Tox-GAN: an artificial intelligence approach alternative to animal studies – a case study with toxicogenomics. Toxicol Sci 2022;186(2):242-259 https://doi.org/10.1093/toxsci/kfab157.
crossref pmid
63. Singh AV, Chandrasekar V, Paudel N, Laux P, Luch A, Gemmati D, et al. Integrative toxicogenomics: Advancing precision medicine and toxicology through artificial intelligence and OMICs technology. Biomed Pharmacother 2023;163: 114784 https://doi.org/10.1016/j.biopha.2023.114784.
crossref pmid
64. PrecisionTox Consortium. The precision toxicology initiative. Toxicol Lett 2023;383: 33-42 https://doi.org/10.1016/j.toxlet.2023.05.004.
crossref pmid
65. Hartung T. Artificial intelligence as the new frontier in chemical risk assessment. Front Artif Intell 2023;6: 1269932 https://doi.org/10.3389/frai.2023.1269932.
crossref pmid pmc

Figure 1.
Database for big data in toxicology
eaht-40-Special_Issue-e2025s07f1.jpg
Figure 2.
Key methodologies of artificial intelligence in toxicology
eaht-40-Special_Issue-e2025s07f2.jpg
Figure 3.
Application and perspective of big data and artificial intelligence in toxicology.
eaht-40-Special_Issue-e2025s07f3.jpg
Table 1.
A list of databases relevant to in vitro mechanisms data (*Note: Several databases provide both in vitro and in vivo data).
Database Description Data size Data type Ref
ChEMBL An Open database containing binding, functional and ADMET information for a large number of drug-like bioactive compounds ~2.4 million compound records, 17 million bioactivity measurements Bioactivity data, chemical structures, molecular targets, and assay data [13]
Connectivity map A resource containing gene expression profiles of cells treated with bioactive small molecules, used for drug repurposing and understanding mechanisms of action ~1.3 million gene expression profiles Transcriptomics data (gene expression profiles) [14]
ToxCast A high-throughput screening database from the US EPA to assess the potential toxicity of chemicals based on in vitro assays ~9,000 chemicals tested across 1,500 assays High-throughput screening (HTS) data, bioassays, and endpoints [15]
MoleculeNet A benchmark dataset for molecular machine learning, including datasets on physical chemistry, biophysics, and physiology Varies depending on thedata type Molecular properties, chemical structures, bioactivity data [16]
Open TG-GATEs A toxicogenomics database containing gene expression profiles and histopathological data from animal studies for drug-induced liver injury ~170 compounds, multiple dose-response studies Gene expression profiles, histopathology, and clinical chemistry data [17]
PubChem A public database for chemical molecules, bioassay results, and associated information ~120 million chemical substances, ~300 million bioactivities Chemical structures, bioassay results, molecular properties [18]
Table 2.
A list of databases relevant to in vivo animal testing data (*Note: Several databases provide both in vitro and in vivo data).
Database Description Data size Data type Ref
Aggregated Computational Toxicology Resource (ACToR) A database aggregating data from various public sources on environmental chemicals, including their properties, bioactivity, and toxicity Over 800,000 chemicals for various assays Chemical structures, bioassay data, regulatory data, and toxicology endpoints [23]
eChemPortal A global gateway provides free access to information on the properties of chemicals as assessed by government or international bodies. Varies depending on the linked datasets from contributing organizations Regulatory toxicology studies, physicochemical properties, environmental fate, ecotoxicology, and human health effects [24]
Toxicity reference database (ToxRefDB) A comprehensive database developed by the US EPA, containing in vivo toxicity data for pesticide chemicals used to support risk assessment ~400 chemicals with thousands of toxicology studies In vivo toxicology study data (e.g., NOAELs, LOAELs, and endpoints such as reproductive or developmental toxicity) [25]
Toxicity value database (ToxValDB) A database containing curated toxicity values and reference data to support chemical risk assessment Varied; focuses on toxicity benchmarks like NOAEL, LOAEL, and reference doses Toxicity values, benchmark doses, and regulatory thresholds [26]
RepDose A curated database containing repeated-dose toxicity study data to support chemical risk assessments and QSAR modeling ~700 chemicals with over 1,400 studies Toxicological study data (e.g., organ toxicity, NOAEL/LOAEL values, duration-based endpoints) [27]
ToxNET A toxicology data network by the US National Library of Medicine, including information on chemicals, their effects, and environmental health Over 50,000 chemicals Literature references, chemical data, toxicology data [28]
Table 3.
Case studies applying AI-based toxicity prediction models for chemical prioritization and next-generation risk assessment.
Aim of study AI-based toxicity prediction models
Application
Ref
Training dataset
Model
Target chemicals
Database Endpoint Data size Feature Algorithms
To screen the consumer product chemicals for their inhalation toxicity potential eChemPortal DB Inhalation toxicity 1709 Various molecular fingerprints and descriptors Various algorithms Consumer product chemicals [24]
To identify the potential toxicity mechanisms of microplastic ToxCast Various biological response 8615 Morgan Fingerprints Multi-layer perceptron Microplastic [47]
To develop AChE inhibition prediction models for various species and applied them to screening HBM4 EU CEC Screen list ChEMBL BindingDB Tox21 Literatures Acetylcholi nesterase inhibition 4075 1444 2D descriptors and Morgan fingerprints Various algorithms Chemicals of emerging concern [48]
To screen 5,206 PFASs from the CompTox database against the different binding sites on the AR using both molecular docking and machine learning techniques. ToxCast Androgen receptor activity 4500 Various molecular fingerprints and descriptors AdaBoost, LR, RF, SVM, k-NN PFAS [49]
To develop a new strategy to rapidly predict bisphenols' hepatotoxicity using network analysis coupled with machine learning models. ToxCast Various biological response Not specified 2D molecular descriptors Various algorithms Bisphenols [50]
To identify toxicity pathway of diesel particular matter using deep learning models ToxCast Various biological response Not specified Morgan Fingerprints Multi-layer perceptron diesel particular matter [51]
To develop quantitative structure–activity relationship (QSAR) models using machine learning (ML) for predicting the toxicity of PAH derivatives PubChem, Web of Science, Google Scholar Acute oral toxicity 788 Various molecular descriptors Four tree-based machine learning models Polycyclic Aromatic Hydrocarbon Derivatives [52]
TOOLS
PDF Links  PDF Links
PubReader  PubReader
ePub Link  ePub Link
XML Download  XML Download
Full text via DOI  Full text via DOI
Download Citation  Download Citation
  Print
Share:      
METRICS
3
Crossref
0
Scopus
1,100
View
81
Download
Editorial Office
Division of Environmental Science and Ecological Engineering, Korea University
145 Anam-ro, Seongbuk-gu, Seoul 02841, Korea
E-mail: envitoxic@gmail.com
About |  Browse Articles |  Current Issue |  For Authors and Reviewers
Copyright © 2025 by The Korean Society of Environmental Health and Toxicology & Korea Society for Environmental Analysis.     Developed in M2PI