Analyze Diet

A modern AI framework integrating deep imputation, synthetic data balancing, and explainable modeling for survival prediction in horse colic.

Abstract: Artificial intelligence (AI) has emerged as one of the most transformative tools for developing clinical decision-support systems in veterinary medicine. Despite its growing use, its full potential remains underutilized in equine medicine, an area of both high economic and clinical importance. Accurate survival prediction in horses with colic is crucial for timely intervention and improved clinical outcomes. Methods: This study aimed to predict survival outcomes in horse colic cases by developing models that combine traditional machine-learning algorithms (XGBoost, Light Gradient Boosting Machine [LightGBM], and Categorical Boosting [CatBoost]) with advanced deep-learning architectures (TabNet, Feature Tokenizer Transformer [FT_Transformer], and Neural Oblivious Decision Ensemble [NODE]). Missing clinical data were imputed using deep-learning-based approaches-Generative Adversarial Imputation Networks (GAIN-OneHot, GAIN-Emb) and Missing Data Imputation via Denoising Autoencoder (MIDAS). Class imbalance was addressed through Conditional Tabular Generative Adversarial Network (CTGAN) and Tabular Variational Autoencoder (TVAE). Model interpretability was assessed using the SHapley Additive exPlanations (SHAP)-based Explainable Artificial Intelligence (XAI) framework to identify the most influential features contributing to survival prediction. Results: Among the tested combinations, the TVAE-GAIN-OneHot-LightGBM pipeline achieved the highest classification performance, with an area under the curve (AUC) value of 0.928, outperforming conventional statistical and machine-learning baselines. SHAP analysis revealed that total_protein, abdomo_appearance, mucous_membrane, packed_cell_volume, and temp_of_extremities were the most decisive clinical variables influencing the model's predictions. Conclusions: The findings demonstrate that ensuring data integrity, optimizing model complexity, and integrating XAI-based interpretability substantially enhance the reliability and clinical applicability of AI-driven models in veterinary medicine. The proposed framework provides a pioneering and explainable approach for developing accurate prognostic systems in equine colic, paving the way for broader AI adoption in clinical veterinary practice.
Publication Date: 2025-12-04 PubMed ID: 41352488DOI: 10.1016/j.aanat.2025.152767Google Scholar: Lookup
The Equine Research Bank provides access to a large database of publicly available scientific literature. Inclusion in the Research Bank does not imply endorsement of study methods or findings by Mad Barn.
  • Journal Article

Summary

This research summary has been generated with artificial intelligence and may contain errors and omissions. Refer to the original study to confirm details provided. Submit correction.

Overview

  • This study developed an advanced artificial intelligence framework that integrates deep imputation techniques, synthetic data balancing, and explainable modeling to accurately predict survival outcomes in horses suffering from colic.
  • The proposed AI models demonstrated superior performance compared to traditional methods and revealed key clinical factors important for survival prediction, improving decision-support tools in equine veterinary medicine.

Background and Importance

  • Artificial intelligence (AI) is increasingly used to develop clinical decision-support systems in veterinary medicine, although its potential in equine medicine remains underexploited despite the economic and clinical significance of horse health.
  • Colic, a common and potentially life-threatening condition in horses, requires timely intervention. Accurate prediction of survival outcomes can guide clinical decisions and improve prognosis.

Research Methods

  • Goal: To predict survival outcomes for horses with colic by integrating advanced AI models with data preprocessing and interpretability techniques.
  • Machine Learning Models: Combined traditional algorithms and deep-learning architectures including:
    • XGBoost, Light Gradient Boosting Machine (LightGBM), Categorical Boosting (CatBoost) – traditional machine-learning algorithms.
    • TabNet, Feature Tokenizer Transformer (FT_Transformer), Neural Oblivious Decision Ensemble (NODE) – advanced deep-learning models specialized for tabular data.
  • Handling Missing Data: Missing clinical information was addressed with deep-learning imputation techniques:
    • Generative Adversarial Imputation Networks (GAIN) with both OneHot and embedding-based input encoding.
    • Missing Data Imputation via Denoising Autoencoder (MIDAS).
  • Addressing Class Imbalance: Since survival data may be imbalanced, synthetic data generation methods were applied:
    • Conditional Tabular Generative Adversarial Network (CTGAN).
    • Tabular Variational Autoencoder (TVAE).
  • Model Interpretability: SHapley Additive exPlanations (SHAP) framework was used as an Explainable AI (XAI) tool to:
    • Identify important clinical features driving the survival predictions.
    • Enhance trustworthiness and clinical interpretability of AI results.

Key Results

  • The best performing pipeline combined:
    • TVAE for synthetic data balancing.
    • GAIN with OneHot encoding for data imputation.
    • LightGBM as the predictive model.
  • This combination achieved an area under the curve (AUC) of 0.928, indicating very high classification accuracy in survival prediction.
  • The model outperformed both traditional statistical approaches and baseline machine-learning algorithms, verifying the benefit of deep learning-based imputation and synthetic data generation in the pipeline.
  • SHAP analysis spotlighted key clinical variables influencing model decisions:
    • Total protein levels.
    • Abdominal appearance.
    • Mucous membrane condition.
    • Packed cell volume.
    • Temperature of the extremities.

Conclusions and Implications

  • Ensuring high data quality through deep imputation and addressing class imbalance with synthetic data generation are critical for developing robust AI models in veterinary medicine.
  • Integration of explainable AI methods such as SHAP enables clinicians to understand the rationale behind predictions, which is vital for clinical acceptance and decision-making.
  • The study’s proposed AI framework offers a novel, accurate, and interpretable approach for survival prognosis in equine colic, with potential application across other clinical scenarios in veterinary care.
  • By demonstrating improved predictive performance and transparency, this work paves the way for wider adoption of AI systems in clinical veterinary practice, potentially leading to better health outcomes for horses.

Cite This Article

APA
Ozger ZB, Cihan P, Ozaydin I. (2025). A modern AI framework integrating deep imputation, synthetic data balancing, and explainable modeling for survival prediction in horse colic. Ann Anat, 264, 152767. https://doi.org/10.1016/j.aanat.2025.152767

Publication

ISSN: 1618-0402
NlmUniqueID: 100963897
Country: Germany
Language: English
Volume: 264
Pages: 152767
PII: S0940-9602(25)00394-2

Researcher Affiliations

Ozger, Zeynep Banu
  • Department of Computer Engineering, Faculty of Engineering and Architecture, Kahramanmaras Sutcu Imam University, Kahramanmaraş, Türkiye.
Cihan, Pınar
  • Department of Computer Engineering, Faculty of Corlu Engineering, Tekirdağ Namık Kemal University, Tekirdağ, Türkiye. Electronic address: pkaya@nku.edu.tr.
Ozaydin, Isa
  • Department of Surgery, Faculty of Veterinary Medicine, Kafkas University, Kars, Türkiye; Nakhchivan State University, Faculty of Natural Sciences and Agriculture, Department of Veterinary Medicine, Nakhchivan, Azerbaijan.

MeSH Terms

  • Animals
  • Horses
  • Colic / veterinary
  • Colic / mortality
  • Horse Diseases / mortality
  • Horse Diseases / diagnosis
  • Artificial Intelligence
  • Machine Learning
  • Algorithms
  • Deep Learning

Conflict of Interest Statement

Declaration of Competing Interest The authors declare that there is no conflict of interest.

Citations

This article has been cited 0 times.