CoTox: Chain-of-Thought-Based Molecular Toxicity Reasoning and Prediction
Abstract
CoTox, a framework integrating LLMs with chain-of-thought reasoning, enhances multi-toxicity prediction by incorporating chemical structure data, biological pathways, and gene ontology terms, improving interpretability and predictive performance in drug development.
Drug toxicity remains a major challenge in pharmaceutical development. Recent machine learning models have improved in silico toxicity prediction, but their reliance on annotated data and lack of interpretability limit their applicability. This limits their ability to capture organ-specific toxicities driven by complex biological mechanisms. Large language models (LLMs) offer a promising alternative through step-by-step reasoning and integration of textual data, yet prior approaches lack biological context and transparent rationale. To address this issue, we propose CoTox, a novel framework that integrates LLM with chain-of-thought (CoT) reasoning for multi-toxicity prediction. CoTox combines chemical structure data, biological pathways, and gene ontology (GO) terms to generate interpretable toxicity predictions through step-by-step reasoning. Using GPT-4o, we show that CoTox outperforms both traditional machine learning and deep learning model. We further examine its performance across various LLMs to identify where CoTox is most effective. Additionally, we find that representing chemical structures with IUPAC names, which are easier for LLMs to understand than SMILES, enhances the model's reasoning ability and improves predictive performance. To demonstrate its practical utility in drug development, we simulate the treatment of relevant cell types with drug and incorporated the resulting biological context into the CoTox framework. This approach allow CoTox to generate toxicity predictions aligned with physiological responses, as shown in case study. This result highlights the potential of LLM-based frameworks to improve interpretability and support early-stage drug safety assessment. The code and prompt used in this work are available at https://github.com/dmis-lab/CoTox.
Community
Can LLM assess molecular toxicity? ๐๐
During drug development, it is crucial to identify whether the chemical compound is toxic or not.
We introduce CoTox, a novel framework that utilizes LLMs for Molecular Toxicity Prediction.
Unlike traditional models that rely solely on molecular structure, CoTox integrates chemical structures, biological pathways, and GO terms to predict six types of organ-specific toxicities, including cardiotoxicity, hepatotoxicity, and nephrotoxicity.
By using Chain-of-Thought prompting, CoTox generates step-by-step reasoning for each prediction, offering transparent and interpretable explanations for why a compound might be toxic.
Interestingly, we also found that IUPAC names work better than SMILES when interfacing with LLMs, thanks to their human-readable format.
Our findings position CoTox as an interpretable and practical tool for early-stage drug development.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Breaking Bad Molecules: Are MLLMs Ready for Structure-Level Molecular Detoxification? (2025)
- MolVision: Molecular Property Prediction with Vision Language Models (2025)
- Large Language Model Agent for Modular Task Execution in Drug Discovery (2025)
- Conditional Chemical Language Models are Versatile Tools in Drug Discovery (2025)
- MolProphecy: Bridging Medicinal Chemists'Knowledge and Molecular Pre-Trained Models via a Multi-Modal Framework (2025)
- TextOmics-Guided Diffusion for Hit-like Molecular Generation (2025)
- Towards Interpretable Drug-Drug Interaction Prediction: A Graph-Based Approach with Molecular and Network-Level Explanations (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper