Skip to main content
Enterprise AI Analysis: Research on Stock Price Prediction Based on Multi Model Ensemble Methods

Enterprise AI Analysis

Research on Stock Price Prediction Based on Multi Model Ensemble Methods

Our analysis reveals significant advancements in predictive accuracy and stability through ensemble methods, offering tangible benefits for quantitative investment strategies. The proposed multi-model ensemble framework, integrating deep learning and traditional machine learning, achieves superior prediction performance, particularly 100% directional accuracy on the TSAL dataset. Dynamic weight allocation further enhances adaptability to market fluctuations, providing robust decision support for quantitative investment.

Executive Impact: Key Performance Indicators

Our analysis reveals significant advancements in predictive accuracy and stability through ensemble methods, offering tangible benefits for quantitative investment strategies. The proposed multi-model ensemble framework, integrating deep learning and traditional machine learning, achieves superior prediction performance, particularly 100% directional accuracy on the TSAL dataset. Dynamic weight allocation further enhances adaptability to market fluctuations, providing robust decision support for quantitative investment.

0% Directional Accuracy
0% Enhanced Prediction Stability
0% Improved Model Generalization

Deep Analysis & Enterprise Applications

Select a topic to dive deeper, then explore the specific findings from the research, rebuilt as interactive, enterprise-focused modules.

Abstract Summary

Stock price forecasting is crucial in quantitative investment. While both traditional and deep learning models have their strengths, single models often lack generalization. This paper proposes a multi-model ensemble framework combining deep learning and traditional machine learning techniques, overcoming the limitations of individual models to improve prediction accuracy and stability. We employ three ensemble strategies: weighted averaging, voting, and dynamic weight allocation, with the latter adjusting model weights in real-time based on historical errors, enhancing adaptability. Experimental results on the TSAL dataset show that the Bayesian Ridge model performs well in specific data scenarios, while the ensemble model achieves 100% directional accuracy. This study highlights the potential of linear models in certain contexts and demonstrates the strength of multi-model ensembles in enhancing prediction robustness, providing reliable decision support for quantitative investment.

Introduction to Stock Price Prediction Challenges

Stock price forecasting has long been a critical research topic in the financial domain, providing decision support for investors and assisting financial institutions in market analysis [1]. However, due to the high volatility, complexity, and non-linear characteristics of the stock market, traditional prediction methods face numerous challenges. Early stock price prediction primarily relied on statistical methods, such as the ARIMA model [2]. These methods forecast by analyzing the linear characteristics of historical data and are suitable for stable and linear data environments. Nevertheless, given the highly volatile and non-linear nature of the stock market, the predictive accuracy and stability of traditional methods are inherently limited. With the advancement of computer science and deep learning technologies, deep learning models such as LSTM and CNN have gradually become mainstream methods for stock price prediction. LSTM, through its unique gating mechanisms, effectively captures long-term dependencies, making it particularly suitable for forecasting short-term volatility and long-term trends in financial markets. For instance, Fischer et al. [3] showed that LSTM captures market non-linearity, while CNN improves accuracy through spatial feature extraction. However, these methods face challenges such as the "black box" problem, dependence on large data, and limited generalization, which are critical in the financial sector, where interpretability is essential. Existing research also faces issues like poor model interpretability. Many studies prioritize accuracy at the expense of transparency, which is crucial in financial markets where investors and regulators must understand model decisions. Additionally, limited generalization and failure to integrate multi-source data constrain model performance. While some models perform well in stable conditions, they often struggle during market fluctuations. Furthermore, many models cannot effectively combine data from diverse sources-such as prices, technical indicators, and sentiment-leading to unstable predictions.[4].

Multi-Model Ensemble Framework Overview

Raw Stock Data Input
Deep Learning & Traditional ML Models
Individual Model Predictions
Ensemble Integration (Weighted Avg, Voting)
Final Stock Price Prediction

Deep Learning Models Utilized

Deep learning models are highly effective at handling complex nonlinear and time-dependent problems. Compared to traditional machine learning, deep learning can automatically extract high-level features through multi-layer networks, improving prediction accuracy. The main architectures used in this study include: (1) Transformer [10], which uses a self-attention mechanism to capture long-term dependencies and overcomes the sequential bottleneck of RNNs, making it ideal for long-term series data and stock market trends; (2) BiLSTM-Attention [11], which combines bidirectional LSTM with multi-head attention to capture both forward and backward dependencies, enhancing response to sudden market fluctuations; (3) GRU [12], which efficiently models sequences through update and reset gates, with attention to focus on key time points for improved prediction; (4) TCN [13], which uses causal dilated convolution to capture long-term dependencies, making it suitable for large-scale data and both short-term and long-term trends; (5) TabNet [14], designed for tabular data, using sparse activation for feature selection and offering strong interpretability and automatic key feature extraction; (6) LSTM-CNN [15], which combines LSTM for long-term trends and CNN for short-term fluctuations, further improving prediction accuracy and robustness.

Traditional Machine Learning Models Utilized

Traditional machine learning models perform well on low-dimensional and small-sample stock data. They can model nonlinear relationships, remain efficient with limited features, and reduce overfitting through feature selection and regularization. The main models used in this study include:(1) XGBoost [5], a GBDT-based model that integrates decision trees with regularization, improving performance on complex and noisy stock market data;(2) Extra-Trees, which uses strong randomness in feature and threshold selection to reduce overfitting and enhance efficiency, making it suitable for volatile financial data;(3) Lasso Regression [6], which applies L1 regularization for embedded feature selection, ideal for high-dimensional stock features; (4) ElasticNet Regression [7], combining L1 and L2 regularization to handle multicollinearity and improve stability;(5) Bayesian Ridge Regression, which adopts Bayesian inference to mitigate noise effects, providing robust predictions in financial environments; (6) Support Vector Regression (SVR), which searches for the optimal hyperplane to model nonlinear relationships and offers strong generalization; (7) K-Nearest Neighbors Regression (KNN) [8], which predicts using nearby sample patterns, suitable for capturing local market behaviors;(8) Multilayer Perceptron (MLP) [9], which learns complex nonlinear structures through multi-layer networks but may overfit on high-dimensional financial data.

Ensemble Learning Strategies

When dealing with stock price data, a single prediction model often faces significant challenges due to the high noise and nonlinear time series characteristics of the market data. This typically leads to poor generalization and high prediction bias. Ensemble learning effectively overcomes these issues by combining the predictions of multiple base models, improving prediction accuracy and robustness. Specifically, ensemble learning constructs a fusion function F(ŷt,1,..., Ŷt,k), which combines the predictions of multiple base models through weighted averaging or other methods to minimize common prediction error metrics like Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE), thus enhancing prediction performance. Below, we introduce three common ensemble strategies and explore their applications in stock price prediction.

Weighted Averaging: Assigns different weights to base models based on their historical performance, with better-performing models receiving higher weights. This includes Basic Weighted Average and Inverse Error Weighting.

Voting Mechanism: Extends to regression tasks by statistically evaluating central tendency to filter extreme values. Includes Median Voting (robust to noise) and Truncated Mean Voting (removes extreme values).

Dynamic Weight Allocation: Introduces a time window mechanism that dynamically adjusts model weights based on recent errors, improving long-term prediction accuracy and adaptability to changing market conditions.

Key Result: Ensemble Directional Accuracy

100% Achieved Directional Accuracy for Ensemble Model

Performance Evaluation Metrics

To assess the models' effectiveness, several key performance metrics were used:

  • Mean Absolute Error (MAE): Measures the average absolute difference between predicted and actual values, reflecting overall error magnitude.
  • Root Mean Square Error (RMSE): The square root of MSE, giving greater weight to large errors while keeping the same units as the target variable.
  • Mean Absolute Percentage Error (MAPE): Expresses prediction error as a percentage, removing scale effects.
  • Coefficient of Determination (R2): Measures the model's capacity to explain the variance of the data, ranging from -∞ to 1.
  • Directional Accuracy: Measures correctness in predicting price movement (rise/fall), crucial in quantitative trading.

Ablation Experiment Results (Table 1)

ModelMAERMSEMAPE(%)R2Dir(%)
Transformer [10]51.5856.8015.51-2.587446.6
GRU [12]60.7565.3518.38-3.749048.6
BILSTM [11]77.2981.2423.56-6.339947.4
TCN [13]38.0541.8511.48-0.948149.1
TabNet [14]35.8239.1010.84-0.700446.9
LSTM_CNN [15]67.8171.6920.66-4.716346.9
XGBoost [5]42.0349.8212.43-1.759865.7
ExtraTrees41.3348.9212.22-1.661684.0
Lasso [6]2.363.300.730.987981.7
ElasticNet [7]3.384.501.050.977587.0
BayesianRidge0.000.000.001.000099.7
SVR_RBF100.76109.2330.7712.267757.1
SVR_Poly59.8389.6218.72-7.933053.4
KNN_5 [8]70.3474.4221.49-5.159167.9
KNN_15 [8]74.3177.9122.75-5.750772.7
MLP_Large [9]44.6048.3913.72-1.604368.9
Ensemble0.000.000.001.0000100.0

Discussion of Ablation Experiment Results

This experiment evaluated 16 prediction models, including deep learning, traditional machine learning, and ensemble methods. The results show that ensemble learning methods outperformed others in stock price prediction, while deep learning and traditional models offered distinct advantages. The BayesianRidge and Ensemble models achieved the best performance across all metrics, with perfect predictions (MAE=0.00, RMSE=0.00, MAPE=0.00%, R2=1.0000). Notably, the Ensemble model also reached 100% in directional prediction accuracy, confirming the strength of ensemble learning for complex problems. Among traditional machine learning models, Lasso (MAE=2.36, MAPE=0.73%) and ElasticNet (MAE=3.38, MAPE=1.05%) performed well, particularly in directional accuracy, achieving 81.7% and 87.0%, respectively. While XGBoost (MAE=42.03) and ExtraTrees (MAE=41.33) were less accurate in numerical prediction, they excelled in directional accuracy, with 65.7% and 84.0%. In contrast, SVR_RBF and SVR_Poly performed poorly in both numerical and directional predictions, with SVR_RBF (MAE=100.76, MAPE=30.77%) failing to produce valuable forecasts. Among deep learning models, TabNet (MAE=35.82, MAPE=10.84%) and TCN (MAE=38.05, MAPE=11.48%) showed strong performance in accuracy, particularly with complex time-series data. However, other models like Transformer (MAE=51.58), GRU (MAE=60.75), and BiLSTM (MAE=77.29) underperformed, with low directional prediction accuracy (46.6% for Transformer and 48.6% for GRU). This may be due to stock market data noise and the dependency of deep learning models on large datasets. Ensemble learning methods demonstrate strong predictive performance, especially in directional forecasting, as shown in Figure 3. By combining multiple base models, the ensemble approach effectively reduces bias and variance, thereby improving overall accuracy and stability. Among the top-performing models, Bayesian-Ridge achieves near-perfect fitting, indicating a clear linear relationship between extracted features and stock price movements. Lasso and ElasticNet perform slightly weaker but still maintain high accuracy and strong regularization, making them suitable for high-dimensional financial data and effective in mitigating overfitting. In contrast, ExtraTrees and TabNet show weaker fitting ability and higher errors, performing even worse than a simple mean baseline, though they still contribute to directional prediction. To further highlight model differences, this study includes a set of visual evaluations. The prediction visualization shows high consistency between the ensemble model and actual stock trends. The model comparison chart illustrates the distinct fitting behaviors of the top five models. The error metric chart presents MAPE rankings, emphasizing the advantages of BayesianRidge and the ensemble approach. The explanatory-power chart compares R2 scores, with higher-performing models showing stronger interpretability. The directional-accuracy chart confirms the ensemble model's 100% success rate in predicting price movements. Finally, the residual analysis reveals stable error fluctuations, demonstrating the robustness of the ensemble model. Overall, these visual results show that ensemble learning provides substantial advantages over single models in fitting accuracy, directional prediction, and stability.

Conclusion

This paper proposes a multi-model ensemble framework that integrates deep learning with traditional machine learning to enhance stock price prediction accuracy and stability. In quantitative investment, accurate stock price forecasting is essential, but single models often suffer from insufficient generalization. To address this, three ensemble strategies are introduced: weighted averaging, voting, and dynamic weight allocation, where model weights are adjusted in real-time based on historical errors, improving adaptability to market fluctuations. Experimental results show that the Bayesian Ridge model excels in certain data scenarios, capturing linear relationships and achieving near-perfect predictions. The ensemble model further optimizes performance, particularly in directional prediction, achieving 100% accuracy and enhancing stability and precision. This study demonstrates the power of ensemble learning in stock price prediction, offering an efficient decision support solution for quantitative investment and suggesting further optimization of ensemble strategies for better prediction accuracy across various data scenarios.

Quantify Your AI Impact

Estimate the potential efficiency gains and cost savings for your enterprise by implementing intelligent automation and predictive analytics.

Estimated Annual Savings $0
Hours Reclaimed Annually 0

Your Path to AI Implementation

A structured approach to integrate multi-model ensemble AI into your financial operations, ensuring maximum impact and minimal disruption.

Phase 1: Discovery & Strategy

Conduct a deep dive into your existing data infrastructure, trading strategies, and specific forecasting needs. Define clear KPIs for AI integration and formulate a tailored implementation roadmap.

Phase 2: Data Engineering & Model Selection

Prepare and normalize financial datasets, including historical prices, technical indicators, and alternative data. Select and fine-tune initial deep learning and traditional machine learning models based on data characteristics.

Phase 3: Ensemble Framework Development

Implement the multi-model ensemble framework, incorporating weighted averaging, voting, and dynamic weight allocation strategies. Develop robust backtesting and validation protocols.

Phase 4: Deployment & Continuous Optimization

Integrate the ensemble model into your live trading or decision-making systems. Establish continuous monitoring for performance, data drift, and model decay, with regular re-calibration and optimization cycles.

Ready to Enhance Your Predictive Power?

Schedule a complimentary strategy session with our AI experts to explore how multi-model ensemble methods can revolutionize your quantitative investment decisions.

Ready to Get Started?

Book Your Free Consultation.

Let's Discuss Your AI Strategy!

Lets Discuss Your Needs


AI Consultation Booking