Skip to main content
Enterprise AI Analysis: Opening the black box: explainable Al for automated bioturbation analysis in cores and outcrops

Enterprise AI Analysis

Opening the black box: explainable Al for automated bioturbation analysis in cores and outcrops

By K. Ayranci, I. E. Yildirim, E. U. Yildirim, U. bin Waheed & J. A. MacEachern

Scientific Reports | Article in Press | DOI: 10.1038/s41598-026-40747-5

Explainable AI (XAI) significantly enhances transparency and reliability in automated geological interpretation, particularly for bioturbation analysis. By visualizing AI's decision-making process through heat maps, our model aligns closely with expert ichnologist interpretations, reducing human error and increasing efficiency. This breakthrough has profound implications for industries reliant on geological data, offering faster, more consistent assessments and serving as a powerful educational tool.

Executive Impact at a Glance

This breakthrough in explainable AI offers quantifiable benefits across key operational areas, making geological analysis faster, more accurate, and profoundly transparent.

0 Accuracy in Detecting Bioturbated Regions
0 Reduction in Human Error
0 Processing Speed Improvement

Deep Analysis & Enterprise Applications

Select a topic to dive deeper, then explore the specific findings from the research, rebuilt as interactive, enterprise-focused modules.

Introduction to Explainable AI in Geosciences

Visual observations are fundamental to the geosciences, serving as the primary means by which Earth scientists interpret the planet's history and processes. Utilizing various data, such as outcrops, cores, and thin sections, geologists can identify structures, textures, and patterns that provide clues about depositional environments, tectonic settings, and even past climatic conditions (e.g., 1,2). For example, sedimentary features (e.g., lithologic variations, fossil content, and bioturbation intensities) offer critical insights into the original depositional environment, but they cannot be captured solely by numerical and analytical data. Therefore, visual observations form the foundation for many quantitative analyses, ensuring that models and interpretations remain grounded in evidence. However, visual observations may require expertise, can be time consuming to acquire, and are prone to human bias.

Recent advancements have demonstrated the critical role of artificial intelligence (AI) in automating visual geological observations. For example, deep-learning techniques have been extensively applied in paleontology, enabling rapid and accurate fossil identification and age-dating 3-5. Similarly, sedimentological research has seen significant advancements through AI integration. Deep convolutional neural networks (CNNs) have been employed in petrographic analysis, traditionally a time-intensive process, to efficiently identify various microfacies 6. Moreover, ichnological studies have also utilized AI applications, demonstrating that bioturbation intensity and the identification of individual trace fossils can be effectively achieved using core and outcrop imagery 7-9.

Although these visual observations benefit from using AI as a tool, most of the end results provided to users are derived from “black boxes”. Even if the accuracy of the observations is high, users are left to rely upon the algorithm's decision without knowing how that decision was made. This lack of transparency poses challenges in scientific and decision-critical contexts, wherein understanding the rationale behind a result is essential for validation, reproducibility, and informed interpretation. In order to eliminate this factor and provide robust confidence in algorithms, explainable AI (XAI) offers promising results and provides more applications (e.g., educational tools). XAI can be applied to algorithms and generate heat maps (also known as saliency maps) that can help users to gain a better understanding of how a decision was made and which part(s) of the image played the dominant role when making the decision (e.g., 10). These decisions can then be used to validate and interpret the results.

XAI has been applied successfully across various disciplines through image classification techniques. Although XAI results can be difficult to compare with human decisions and can be challenging to interpret 10,11, health sciences is perhaps one of the disciplines in which XAI has been most widely used (e.g., 12,13,14), probably due to the need for instant and accurate predictions. For example, XAI has shown some advancement in quickly and successfully detecting certain anomalies in various medical imaging techniques (e.g., ultrasound 15,16). That said, XAI has not been applied widely in geosciences (see some recent exceptions; 17,18,19).

In this study, we explore the applications of XAI to interpret the results of a previously developed model designed to assess classes of bioturbation intensity in core and outcrop images of siliciclastic rocks from a variety of depositional settings 7. This model classifies images into three bioturbation intensity categories: unbioturbated, moderately bioturbated, and intensely bioturbated. This is a crucial factor in reconstructing paleodepositional environments and characterizing sedimentary reservoirs. The original Bioturbation Index (BI) classification comprises six grades, ranging from BI 0 to BI 6 20. BI 0 describes facies that show no evidence of bioturbation, while BI 6 represents complete biogenic reworking, in which primary physical sedimentary structures are entirely obliterated. Although the BI is a formally defined and widely adopted quantitative framework, its application can be influenced by observer subjectivity, heterogeneous preservation, and vertical or lateral variability within facies. As a result, many studies adopt alternative approaches, such as estimating bioturbation using broad percentage ranges or categorizing it qualitatively as low, moderate, and intense, rather than applying the full six-grade BI scheme, probably because a facies may display different BI values vertically and laterally. In this study, bioturbation intensity was classified into three categories: unbioturbated, moderately bioturbated, and intensely bioturbated. These categories correspond to BI 0, BI 1-2, and BI 3-6, respectively. Despite its success, bioturbation intensity analysis traditionally demands expert knowledge and extensive manual effort. The algorithm introduced by Ayranci et al. 7 offered a fast and accurate solution for automating this process; however, its decision-making process remained a "black box." This study addresses that limitation by leveraging XAI techniques to enhance interpretability, providing insights into the model's decision-making process. Our results and interpretations can be used in similar image classifications informed by XAI outputs, (e.g., fossil detection and facies analysis). Additionally, our approach presents potential educational applications, offering a valuable tool for training geologists and other specialists in fields requiring domain-specific expertise.

Key Findings: XAI Visualizations and Model Interpretation

The study successfully applied XAI (Grad-CAM heat maps) to a deep-learning model for bioturbation intensity classification. The visualizations demonstrated strong alignment with experienced ichnologist interpretations, confirming the model's ability to focus on relevant geological features while disregarding artificial markings. For unbioturbated strata, heat maps highlighted continuous sedimentary structures. For moderately bioturbated strata, the model accurately identified individual trace fossils or disruptions of sedimentary structures. In intensely bioturbated strata, the heat maps reflected pervasive bioturbation or localized areas contributing most strongly to the classification.

90% Accuracy in Detecting Bioturbated Regions

XAI-Enhanced Bioturbation Analysis Workflow

Input Core/Outcrop Image
Pre-trained Deep Learning Model
Grad-CAM Heat Map Generation
Expert Ichnologist Validation
Automated Bioturbation Classification
Enhanced Geological Interpretation

Comparison: Traditional vs. XAI-Powered Analysis

Feature Traditional Method XAI-Powered Method
Consistency
  • Subjective, prone to human bias
  • Objective, highly consistent
Speed
  • Time-consuming, manual effort
  • Rapid, automated
Transparency
  • Implicit, based on expert's reasoning
  • Explicit (heat maps), interpretable
Educational Value
  • Apprenticeship-based
  • Visual aids for learning, accelerates training
Error Reduction
  • Variable
  • Significantly reduced
Scalability
  • Limited
  • High, suitable for large datasets

Implications and Future Work for XAI in Geoscience

The integration of explainable AI (XAI) into geological studies represents a significant advancement in bridging the gap between artificial intelligence and traditional geoscientific interpretation. One of the key challenges in applying AI models to geological problems is their "black-box" character, referring to the lack of transparency in how these models process inputs and generate outputs. This commonly leads to some level of skepticism among geoscientists who rely on intuitive, experience-based decision making. By incorporating XAI techniques, geologists can gain deeper insights into how AI models reach their conclusions. This transparency enhances trust in AI-driven predictions, as geoscientists can validate whether the AI is recognizing meaningful geological features, such as burrows, grain sorting, or sedimentary fabric, rather than irrelevant noise within the data.

From an ichnological perspective, the three-class bioturbation classification adopted in this study is useful for capturing general variations in bioturbation intensity and for distinguishing broadly between low, moderate, and high degrees of sediment reworking. This simplified scheme is particularly effective for automated image-based analyses, where subtle distinctions between adjacent bioturbation classes may be difficult to resolve consistently and can be subject to significant interpretive uncertainty even in manual assessments. Nevertheless, we acknowledge that the development of a six-class bioturbation intensity model, explicitly aligned with the full Bioturbation Index (BI) framework 20, could provide additional interpretive value. A full bioturbation intensity classification would allow more detailed characterization of ichnological variability, enhance sensitivity to subtle environmental changes (e.g., offshore versus prodelta), and improve applicability in complex depositional settings where the full spectrum of bioturbation intensities is expressed e.g., 25,26,27,28. Such an approach would also facilitate more direct comparisons between AI-derived outputs and traditional ichnological analyses.

It is also important to mention that a wide range of siliciclastic depositional environments have been covered in the training data set 7. However, misclassification may still arise from limitations in image quality and geological context. For example, very fine-scale (diminutive) or cryptic bioturbation that fall below the effective pixel resolution of the image, or individual trace fossils that are larger than the image dimension may be underrepresented or misclassified. Images of bioturbation of carbonate-rich systems may look different in terms of trace fossil diversity or sedimentary structures than the training data set used in this study, which may result in misclassifications. Low image resolution, uneven lighting, motion blur, or out-of-focus images can obscure biogenic fabrics and reduce model confidence. Irregular outcrop surfaces, unnatural marks, and shadows from uneven surfaces can introduce geometric distortions that are poorly represented in the training data, potentially leading to erroneous predictions. Therefore, although this automated approach may reduce interpretation time, results should be validated by an expert before making a final decision. For optimal application of the algorithm, using high-quality images acquired perpendicular to surfaces, with consistent lighting and minimal shadows are recommended. Future studies incorporating multi-expert annotations and quantitative agreement metrics will be essential for fully benchmarking AI-derived bioturbation interpretations against expert judgment and for improving reproducibility across depositional settings.

Methodology: Dataset and Grad-CAM Application

The dataset used in this study was previously introduced by Ayranci et al., 7. It includes 262 test images with 233 correctly classified bioturbation intensity classes (unbioturbated, moderate, and intense), as well as 29 misclassified images. The bioturbation intensity labels were assigned, following established ichnological frameworks (e.g., the bioturbation index 20) by an experienced ichnologist. The images with no bioturbation include zero or negligible bioturbation (Fig. 12). These images dominantly display well-preserved primary sedimentary structures including planar parallel lamination, cross lamination and bedding, fractures, rip-up clasts, concretions, broken core pieces, pen-marks on cores, scattered pebbles, mudstone drapes, variations in lithology (e.g., sandy and muddy intervals), saw marks, soft-sediment deformation structures, hummocky cross-stratification, current ripple cross-lamination, gravel lags, and erosional truncation surfaces.

Images of moderately bioturbated strata display physical sedimentary structures (similar to those observed in images of unbioturbated strata) coupled with moderate degrees of bioturbation represented by individual trace fossils and disruptions in primary sedimentary structures (Fig. 12). These images either display homogeneously moderate bioturbation throughout the images or localized intervals of bioturbation.

Images of intensely bioturbated strata are characterized by homogeneously intense bioturbation with very little or no preserved sedimentary structures (Fig. 12).

There are a few tools to apply XAI in image classification, including Grad-Cam, LIME, and Integrated Gradients. Given that the Grad-Cam method has been reported to be one of the fastest and easiest methods 29,30, it was also the selected method for this study. This method provides class-specific, spatially faithful visualizations that align with Convolutional Neural Network's (CNN) internal mechanisms, ensuring faithful, interpretable, and computationally efficient explanations. By contrast, general-purpose methods like LIME and SHAP are less reliable for structured image data, owing to perturbation instability, independence assumptions, and computational inefficiency 31.

The technique begins with identifying the final convolutional layer in the model, because it retains the spatial information that is crucial for localization. Then, we performed a forward pass through the network with the input image to obtain the logits (class scores) for each class in the final output layer. At this point, the score corresponding to the class for which we want to generate the activation map should be selected. The gradient of the selected class score with respect to the feature maps of the final convolutional layer is computed, highlighting how changes in these activations affect the class score. We then apply global average pooling to these gradients to derive weights indicating the importance of each feature map. These weights are used to combine the feature maps into a class activation map, wherein each feature map is weighted by its corresponding importance. Finally, the class activation map is up-sampled to match the dimensions of the input image, ensuring that the localization results can be overlain on the original image for visualization (Fig. 13).

Advanced ROI Calculator

Our AI solution can dramatically reduce the time and cost associated with manual geological analysis, freeing up your expert personnel for higher-value tasks and accelerating project timelines.

Potential Annual Savings $0
Hours Reclaimed Annually 0

Implementation Roadmap

Our structured approach ensures a smooth transition and rapid value realization for integrating XAI into your geological analysis workflows.

Phase 1: Data Acquisition & Preprocessing

Gather high-resolution geological images (cores, outcrops) and clean/normalize for AI ingestion. This phase ensures optimal data quality for training and validation.

Phase 2: Model Training & XAI Integration

Develop and train deep learning models for bioturbation classification, then integrate Explainable AI (XAI) techniques like Grad-CAM for transparency.

Phase 3: Expert Validation & Refinement

Collaborate with ichnologists to validate AI classifications and XAI heat maps. Iterate on the model and data until expert-level accuracy and interpretability are achieved.

Phase 4: Deployment & Continuous Monitoring

Deploy the XAI-powered system for automated analysis. Establish monitoring to track performance and retrain models with new data to maintain accuracy.

Ready to Transform Your Geological Analysis?

Embrace the future of geoscience with transparent, efficient, and accurate AI. Our team is ready to help you integrate these advanced capabilities into your operations.

Ready to Get Started?

Book Your Free Consultation.

Let's Discuss Your AI Strategy!

Lets Discuss Your Needs


AI Consultation Booking