1Department of Computer Science & Engineering, RCC Institute of Information Technology monika.singh@rcciit.org.in, harinandan.tunga@gmail.com Department of Mathematics, National Institute of Technology Durgapur skar.maths@nitdgp.ac.in |
Author(s) | Title | Methodology Used | Techniques Used | Results | Limitations |
|---|---|---|---|---|---|
Chauhan et al. [1] | A novel deep learning model for stock market prediction using a sentiment analysis system… | Proposed SenT-In, a sentiment-aware deep learning model combining market data with news sentiment | CNN and GRU networks compute a daily sentiment index; an attention mechanism fuses the sentiment with price time series. | Outperformed baseline models (GRU, LSTM, CNN, SVM): ~+9% accuracy, + 7% F1, + 13% AUC-ROC on average | Tested on select indices; adds model complexity. |
Wang et al. [2] | A hybrid attention-based LSTM-CNN model for stock price forecasting… | Developed a hybrid DL model using heterogeneous data (e.g. prices, news) with attention. | Stack of LSTM and CNN layers with an attention mechanism to highlight important features from multi-source inputs. | Reported improved accuracy over single-model approaches. | Complex architecture; requires careful tuning. |
Sezer et al. [3] | Financial time series forecasting with deep learning: A systematic literature review (2005–2019) | Conducted a systematic literature review of DL methods in finance | Categorized studies by forecasting target (indices, forex, etc.) and DL model type (CNN, DBN, LSTM, etc.) | Found deep learning often outperforms traditional ML in financial forecasting; summarized trends and research gaps | Review covers up to 2019; no new experiments. |
Hiransha et al. [4] | NSE Stock Market Prediction Using Deep-Learning Models | Empirical comparison of four neural network models on stock data (NSE/NYSE). | Trained MLP, vanilla RNN, LSTM and CNN on historical prices. | CNN model gave the lowest error, outperforming RNN/LSTM and linear models | Short-term prediction only; no external sentiment or news data. |
Araci [5] | FinBERT: Financial Sentiment Analysis with Pre-trained Language Models | Developed FinBERT, a BERT-based model fine-tuned on financial text for sentiment analysis. | Transformer (BERT) pre-trained on generic data, then fine-tuned on financial domain corpora. | Outperforms prior state-of-art on financial sentiment benchmarks; higher precision/F1 even with limited training data. | Focus is sentiment labeling, not stock forecasting. |
Hutto & Gilbert [6] | VADER: A Parsimonious Rule-based Model for Sentiment Analysis of social media Text | Introduced VADER, a rule-based sentiment analyzer optimized for social media text | Lexicon of words with sentiment scores + grammar rules (e.g. handling negation, emphasis, capitalization) | Fast, effective for social media content; beats many lexicon methods. | No learning; fixed lexicon may miss context-specific language. |
Loughran & McDonald [7] | When is a liability NOT a liability? Textual analysis, dictionaries, and 10-Ks | Textual analysis of 10-K reports; built finance-specific sentiment dictionaries. | Analyzed word usage in financial filings; created new negative/positive word lists tailored to finance. | Showed common English “negative” words often are neutral in finance; developed word lists that better correlate with returns, volatility, etc. | Dictionaries fixed to US 10-K language; may need updates over time. |
Huang et al. [8] | Using Social Network Sentiment Analysis and Genetic Algorithm to Improve the Stock Prediction Accuracy of the Deep Learning-Based Approach | Combined social media sentiment and genetic algorithm (GA) with DL to predict stock changes. | GA selected relevant technical indicators; grey relational analysis identified key sentiment features; LSTM for forecasting. | Using GA-screened technical variables plus Twitter sentiment features significantly improved LSTM accuracy over baseline | Case study on one stock (TSMC); may not generalize to all markets. |
Patra & Mohanty [9] | An LSTM–GRU based hybrid framework for secured stock price prediction | Proposed a hybrid RNN model combining LSTM and GRU layers. | Input of 25 features (20 technical indicators) fed into an LSTM–GRU network | Achieved high accuracy on S&P 500 price forecasts (details in paper). | Complex model; potential overfitting with many indicators. |
Shi et al. [10] | Attention-based CNN-LSTM and XGBoost hybrid model for stock prediction | Built a multi-stage hybrid: ARIMA preprocessing → attention CNN-LSTM → XGBoost fine-tuning | ARIMA for trend removal; CNN extracts features; LSTM captures time-dependence; attention highlights key inputs; XGBoost regresses final output. | Reported that the hybrid model is more effective than individual methods, yielding relatively high forecast accuracy | Very complex pipeline; reliant on correct ARIMA fit and hyperparameter tuning. |
Thakkar & Chaudhari [11] | Applicability of genetic algorithms for stock market prediction: A systematic survey of the last decade | Survey of GA-based techniques in stock forecasting. | Reviewed GA applications: feature selection, parameter optimization, rule extraction in prior studies. | Summarized trends in GA usage; identified common practices (e.g. GA-optimized ANN, trading rules). | Literature-only; no empirical validation of their own. |
Barua et al. [12] | Comparative analysis of deep learning models for stock price prediction in the Indian market | Empirical comparison of 7 DL models on Indian stock indices. | Compared RNN, LSTM, CNN, GRU, Attention-LSTM, etc., using metrics (MAE, MSE, R²). | CNN and GRU models generally outperformed others on stable stocks; Attention-LSTM excelled on volatile stocks (e.g. Reliance); RNN/LSTM were less accurate. | Only Indian market data; fixed model hyperparameters may bias results. |
Awad et al. [13] | Forecasting stock market indices using the RNN-based hybrid models: CNN–LSTM, GRU–CNN, and ensemble models | Proposed hybrid RNN models and ensembles for index forecasting. | Built CNN–LSTM and GRU–CNN architectures; also constructed ensembles combining RNN variants. | Found that hybrid and ensemble models achieved higher accuracy on index predictions than single models. | Added model complexity; requires large data and training. |
Jiang et al. [14] | Financial sentiment analysis using FinBERT with application in predicting stock movement | Integrated FinBERT sentiment analysis with an LSTM forecasting model. | FinBERT to extract text sentiment scores; LSTM for price series prediction; compared FinBERT + LSTM with BERT, LSTM, ARIMA. | Incorporating FinBERT sentiment significantly improved market trend prediction accuracy compared to baselines. | Dependence on quality of news data; may overfit sentiment signals. |
Dhabe et al. [15] | Stock Market Trend Prediction Along with Twitter Sentiment Analysis | LSTM model for price prediction, enhanced with Twitter sentiment input. | Used Yahoo Finance API for real-time stock data and LSTM forecasting; computed Twitter sentiment (positive/negative %). | The mean squared error dropped by about 3% compared to a model using only price data(obtained least error in prediction, for Asian Paints data for the split of 80:20) | Tested on limited stocks (India); simple sentiment measure may miss nuance. |
Padmanayana et al. [16] | Stock Market Prediction Using Twitter Sentiment Analysis | Collected live tweets and news headlines using APIs; performed sentiment analysis; combined this with historical stock data for prediction | Naïve Bayes for sentiment classification And XGBoost for stock price prediction | Achieved 89.8% accuracy in predicting stock prices using combined sentiment and stock data | Relied on noisy Twitter sentiment data; Only daily-level predictions, not finer time granularity |
Shankar et al. [18] | Stock-Price-Prediction-Using LSTM ARIMA | Hybrid model using LSTM neural network combined with ARIMA (and Unobserved Component Model). | Time-series ARIMA model + LSTM network for prices; likely model ensemble. | Found that LSTM achieved the lowest error among compared methods. | Used non-standard data (“livestock market”); limited scope beyond case data. |
Varghese &. Mohan [19] | Study on the Sentimental Influence on Indian Stock Price | Built a four-stage framework: scraped financial news → applied modified VADER for sentiment scoring → used Granger causality and transfer entropy tests. | Modified VADER for sentiment analysis Granger causality (parametric) - Shannon and Renyi’stransfer entropy(non-parametric) | Sentiment has a causal influence on stock prices and Transfer entropy better captured non-linear relationships than Granger (Pharma stocks showed strongest sentiment-price link during COVID) | No Twitter or social media sentiment No sliding window used reduce overlap and dependency between samples |
Singh & Srivastava [20] | A Genetic Algorithm optimized hybrid model for stock price prediction | Built a GA-optimized hybrid forecasting model. | Genetic Algorithm tuned model parameters (possibly combining ARIMA/ANN layers). | Reported enhanced prediction accuracy (details not provided). | GA adds complexity; convergence and generalization Sentiment has a causal influence on stock prices (not vice versa) - Transfer entropy better captured non-linear relationships than Grangercan be issues. |
Patel et al. [21] | Sentiment analysis integrated LSTM model for stock price prediction | LSTM model augmented with sentiment analysis of news/social data. | Extracted sentiment features (e.g. from Twitter/news) and input them into LSTM time-series predictor. | Including sentiment signals improved LSTM predictive performance over price-only models. | Depends on sentiment data quality; lexicon or model choice may limit gains. |
Model Name | Description | Prediction Output ( |
|---|---|---|
Linear Regression (LR) | Fits a linear model using features like Open, High, Low, and Volume to predict Close price. | =3475.25 |
LSTM | Trained on sequential Close prices to capture short- and long-term dependencies. | =3492.10 |
GRU | A simplified RNN model for efficient sequence learning | =3481.60 |
Bi-LSTM | Learns bidirectional temporal dependencies | =3488.75 |
ARIMA | Classical time series model for stationary signals | =3469.80 |
Headlines | Positive | Neutral | Negative |
|---|---|---|---|
0.70 | 0.25 | 0.05 | |
0.50 | 0.35 | 0.15 |
Headlines | Compound Score |
|---|---|
+ 0.65 | |
+ 0.15 |
Sentiment Tools | Score ( ) |
|---|---|
FinBERT | |
VADER | |
Loughran-McDonald |
Candidate | Model Weights ( − ) | Sentiment Weights ( − ) | Constraints Check |
|---|---|---|---|
[0.10, 0.25, 0.20, 0.25, 0.20] | [0.50, 0.30, 0.20] | Correct | |
[0.15, 0.15, 0.30, 0.20, 0.20] | [0.40, 0.40, 0.20] | Correct | |
[0.20, 0.20, 0.20, 0.20, 0.20] | [0.33, 0.33, 0.34] | Correct |
ALGORITHM 4.1: HYBRID STOCK PREDICTION USING GA | |||||||
|---|---|---|---|---|---|---|---|
Input: | S ← Selected stock | ||||||
H ← Historical OHLCV data for S | |||||||
K ← List of IT-sector keywords | |||||||
W_prev ← Previously optimized model and sentiment weights | |||||||
Output: Final_Prediction ← Next-day Predicted closing price for stock S | |||||||
1 | Begin | ||||||
2 | Step 1: | Data Acquisition | |||||
3 | H ← Fetch_Historical_Data (S) | ||||||
4 | Headlines ← Scrape_Financial_News (K) | ||||||
5 | V ← Calculate_Volatility (H) | ||||||
6 | Step 2: | Data Preprocessing | |||||
7 | Scaled_H ← Normalize (H) | ||||||
8 | Sequences ← Generate_Time_Step_Sequences (Scaled_H) | ||||||
9 | Step 3: | Forecasting Model Predictions | |||||
10 | For each model | ||||||
11 | Train on Sequences or Scaled_H | ||||||
12 | M_Output[ ]←Predict_Next ( ) | ||||||
13 | End For | ||||||
14 | Step 4: | Sentiment Analysis | |||||
15 | For each tool | ||||||
16 | S_Score[ ] ←Analyze_Sentiment ( , Headlines) | ||||||
17 | End For | ||||||
18 | Step 5: | Weight Optimization Using GA | |||||
19 | If Current_Date is the first of the month Then | ||||||
20 | Population← Initialize_Weight_Population() | ||||||
21 | While not converged Do | ||||||
22 | For each weight vector | ||||||
23 | P_pred ← Weighted_Prediction(W, M_Output, S_Score, V) | ||||||
24 | RMSE ← Compute_RMSE( Ppred, Pactual) | ||||||
25 | Tol_Acc ← Compute_Tolerance(Ppred, Pactual, 1%) | ||||||
26 | Fitness[W] ← Aggregate_Fitness(RMSE, Var, Tol_Acc) | ||||||
27 | End For | ||||||
28 | Parents ← Select_Parents(Fitness) | ||||||
29 | Offspring ← Crossover_And_Mutate(Parents) | ||||||
30 | Population ← Update (Population, Offspring) | ||||||
31 | End While | ||||||
32 | W_optimal ← Select_Best_Weights(Population) | ||||||
33 | Else: | ||||||
34 | W_optimal ← Load(W_prev) | ||||||
35 | End If | ||||||
36 | Step 6: | Final Prediction Generation | |||||
37 | Final_Prediction ← Weighted_Prediction(W_optimal, M_Output, S_Score, V) | ||||||
38 | Step 7: | Output Results | |||||
39 | Display Final_Prediction, W_optimal, RMSE, sentiment weights, and volatility factors | ||||||
40 | End | ||||||
Models / Tools | Pros | Cons |
|---|---|---|
Linear Regression | ⎫ Simple and interpretable ⎫ Fast to train and deploy | o Assumes linearity o Poor performance with non-linear data |
LSTM | ⎫ Captures long-term dependencies ⎫ Handles non-linear time series | o Computationally intensive o Prone to overfitting on small datasets |
GRU | ⎫ Similar to LSTM but with fewer parameters ⎫ Faster training | o Slightly less expressive than LSTM o May underperform in long sequences |
Bi-LSTM | ⎫ Learns from both past and future contexts ⎫ Strong sequence modeling | o Higher computational cost o Requires more memory |
ARIMA | ⎫ Effective for stationary time series ⎫ Easy to interpret | o Assumes linearity o Not suitable for multi-feature or non-stationary data |
FinBERT | ⎫ Context-aware sentiment analysis ⎫ Fine-tuned for financial domain | o Requires high computation o Needs GPU for efficient processing |
VADER | ⎫ Fast and efficient ⎫ Good for short texts (e.g., headlines, tweets) | o Not domain-specific o Limited handling of complex expressions |
Loughran-McDonald | ⎫ Finance-specific lexicon ⎫ Transparent and interpretable | o Rule-based and static o Ignores contextual meaning and sentence structure |
SentiVol-GA Hybrid Model (Proposed) | ⎫ Combines strengths of all models ⎫ Dynamically adjusts weights using GA ⎫ Accounts for sentiment and volatility ⎫ Robust and adaptive across market conditions | o Complex implementation o Requires significant computation for monthly GA optimization o Dependent on reliable input from all submodule |
Metrics | LR | LSTM | GRU | ARIMA | Bi-LSTM | Hybrid Model |
|---|---|---|---|---|---|---|
For TCS Stock Prediction | ||||||
R² | 0.9055 | 0.8665 | 0.8138 | 0.6752 | 0.7673 | 0.9149 |
MSE | 1654.66 | 2336.23 | 3259.07 | 5684.17 | 63.82 | 1489.64 |
RMSE | 40.68 | 48.33 | 57.09 | 75.39 | 4073.02 | 38.60 |
MAE | 36.90 | 31.79 | 45.26 | 47.46 | 43.54 | 29.61 |
Tolerance ± 1.0% | 53.57% | 64.29% | 46.43% | 57.14% | 53.57% | 67.86% |
Tolerance ± 1.5% | 78.57% | 82.14% | 64.29% | 67.86% | 67.86% | 92.86% |
Tolerance ± 2.0% | 92.86% | 92.86% | 82.14% | 78.57% | 78.57% | 92.86% |
For HCL Technologies Stock Prediction | ||||||
R² | 0.9403 | 0.7457 | 0.7136 | 0.5876 | 0.6840 | 0.9332 |
MSE | 365.56 | 1557.35 | 1753.56 | 2525.56 | 1935.29 | 409.28 |
RMSE | 19.12 | 39.46 | 41.88 | 50.25 | 43.99 | 20.23 |
MAE | 15.91 | 26.73 | 34.73 | 33.45 | 28.48 | 14.60 |
Tolerance ± 1.0% | 46.43% | 50.00% | 21.43% | 42.86% | 50.00% | 60.71% |
Tolerance ± 1.5% | 71.43% | 57.14% | 39.29% | 50.00% | 64.29% | 75.00% |
Tolerance ± 2.0% | 96.43% | 71.43% | 46.43% | 57.14% | 71.43% | 82.14% |
For Infosys Stock Prediction | ||||||
R² | 0.9109 | 0.8561 | 0.7137 | 0.7165 | 0.8160 | 0.9239 |
MSE | 567.14 | 915.91 | 1822.45 | 1804.96 | 1171.41 | 484.23 |
RMSE | 23.81 | 30.26 | 42.69 | 42.48 | 34.23 | 22.00 |
MAE | 21.59 | 23.13 | 34.05 | 26.66 | 23.84 | 18.08 |
Tolerance ± 1.0% | 21.43% | 42.86% | 32.14% | 53.57% | 53.57% | 46.43% |
Tolerance ± 1.5% | 64.29% | 57.14% | 39.29% | 67.86% | 64.29% | 64.29% |
Tolerance ± 2.0% | 82.14% | 71.43% | 50.00% | 71.43% | 67.86% | 85.71% |
For Wipro Stock Prediction | ||||||
R² | 0.9728 | 0.8620 | 0.7644 | 0.7612 | 0.8549 | 0.9769 |
MSE | 4.71 | 23.93 | 40.86 | 41.41 | 25.17 | 4.01 |
RMSE | 2.17 | 4.89 | 6.39 | 6.44 | 5.02 | 2.00 |
MAE | 1.46 | 3.63 | 4.85 | 4.76 | 3.38 | 1.50 |
Tolerance ± 1.0% | 85.71% | 46.43% | 32.14% | 50.00% | 57.14% | 75.00% |
Tolerance ± 1.5% | 92.86% | 64.29% | 39.29% | 53.57% | 67.86% | 92.86% |
Tolerance ± 2.0% | 92.86% | 78.57% | 64.29% | 53.67% | 78.57% | 100.00% |
Metrics | LR | LSTM | GRU | ARIMA | Bi-LSTM | Hybrid Model |
|---|---|---|---|---|---|---|
For Persistent Systems | ||||||
R² Score | 0.9649 | 0.8826 | 0.8112 | 0.5952 | 0.8220 | 0.9730 |
MSE | 4005.52 | 13415.20 | 21570.30 | 46256.90 | 20335.20 | 3045.79 |
RMSE | 63.29 | 115.82 | 146.87 | 215.07 | 142.60 | 55.19 |
MAE | 52.04 | 77.02 | 111.82 | 141.33 | 104.91 | 48.27 |
Tolerance ± 1.0% | 48.15% | 51.85% | 33.33% | 37.04% | 40.74% | 62.96% |
Tolerance ± 1.5% | 74.07% | 70.37% | 48.15% | 40.74% | 48.15% | 77.78% |
Tolerance ± 2.0% | 88.89% | 70.37% | 55.56% | 51.85% | 59.26% | 92.59% |
For Sasken Technologies | ||||||
R² Score | 0.8581 | 0.5845 | 0.6493 | 0.5000 | 0.5473 | 0.8828 |
MSE | 1281.43 | 3751.82 | 3166.88 | 6334.19 | 4087.95 | 1058.47 |
RMSE | 35.80 | 61.25 | 56.28 | 79.59 | 63.94 | 32.53 |
MAE | 25.86 | 49.19 | 47.68 | 58.97 | 48.53 | 28.13 |
Tolerance ± 1.0% | 51.85% | 18.52% | 14.81% | 25.93% | 25.93% | 18.52% |
Tolerance ± 1.5% | 62.96% | 29.63% | 25.93% | 25.93% | 40.74% | 40.74% |
Tolerance ± 2.0% | 62.96% | 37.04% | 25.93% | 29.63% | 44.44% | 51.85% |
Metrics | LR | LSTM | GRU | ARIMA | Bi-LSTM | Hybrid Model |
|---|---|---|---|---|---|---|
For Policy Bazar Stock Prediction | ||||||
R² Score | 0.9720 | 0.7432 | 0.5524 | 0.5217 | 0.7632 | 0.9724 |
MSE | 292.17 | 2680.46 | 4672.47 | 4993.13 | 2471.59 | 288.14 |
RMSE | 17.09 | 51.77 | 68.36 | 70.66 | 49.72 | 16.97 |
MAE | 12.69 | 36.85 | 261.05 | 46.82 | 39.62 | 13.72 |
Tolerance ± 1.0% | 66.67% | 33.33% | 14.81% | 37.04% | 25.93% | 62.96% |
Tolerance ± 1.5% | 81.48% | 40.74% | 25.93% | 44.44% | 40.74% | 81.48% |
Tolerance ± 2.0% | 88.89% | 59.26% | 33.33% | 51.85% | 44.44% | 88.89% |
For Quick Heal Stock Prediction | ||||||
R² Score | 0.8465 | –3.1443 | 0.0636 | –23.4400 | 0.2940 | 0.8406 |
MSE | 17.71 | 478.11 | 108.03 | 2819.51 | 81.45 | 18.39 |
RMSE | 4.21 | 21.87 | 10.39 | 53.10 | 9.03 | 4.29 |
MAE | 3.15 | 10.54 | 8.44 | 17.41 | 5.77 | 3.49 |
Tolerance ± 1.0% | 59.26% | 29.63% | 18.52% | 48.15% | 55.56% | 51.85% |
Tolerance ± 1.5% | 70.37% | 37.04% | 25.93% | 51.85% | 55.56% | 59.26% |
Tolerance ± 2.0% | 85.19% | 51.85% | 37.04% | 55.56% | 59.26% | 81.48% |
Model | Strengths | Limitations | Best Performing On | Metric(s) Where Outperformed |
|---|---|---|---|---|
Linear Regression (LR) | Fast and simple baseline; competitive for stable stocks. | Poor in handling non-linearity and complex patterns. | Large Caps (HCL Tech, Infosys) | R² on HCL Tech (0.9403), MAE on Quick Heal (3.15) |
LSTM | Captures sequential patterns well; stable for trending data. | Sensitive to training parameters; underperforms on highly volatile stocks. | Mid Caps (Persistent Systems) | RMSE (115.82), MAE (77.02) on Persistent |
GRU | Efficient training; good generalization; balanced speed vs. accuracy. | Slightly less accurate on volatile stocks than Bi-LSTM or Hybrid. | Stable stocks (TCS, Sasken) | R² on TCS (0.8138), Tolerance ± 1.5% on Infosys |
Bi-LSTM | Bidirectional learning improves context; useful in volatility. | Computationally heavier; prone to overfitting on small datasets. | Small Caps (Quick Heal, Policy Bazar) | Best RMSE (9.03) and MAE (5.77) on Quick Heal |
ARIMA | Effective for linear historical trends. | Cannot model non-linear or sentiment effects; static behavior. | Moderate performance on stable stocks | Competitive MAE on Infosys (26.66); moderate Tolerance ± 2% |
Hybrid Model | Integrates technical models, sentiment, and GA optimization; highly adaptive. | Computationally intensive; needs tuning for real-time deployment. | All: Large, Mid, and Small Caps | Highest R², lowest RMSE, top tolerance in most cases |
Parameter | Description | Sensitivity |
|---|---|---|
Model Weights (w₁–w₅) | Relative contribution of LR, LSTM, GRU, ARIMA, Bi-LSTM predictions. | Observation: • Increasing weight of LSTM and GRU improves accuracy slightly under trending markets. • Excessive emphasis on ARIMA (linear model) degrades performance under high volatility. Example: ⎫ When increases by ~ 12% ⎫ When observed (≈ 0.92) |
Sentiment Weights (sw₁–sw₃) | Weights assigned to sentiment scores from FinBERT, VADER, LM Dictionary. | Observation: • FinBERT consistently yields the most predictive sentiment score. • Optimal configuration: . Deviation Impact: ⎫ Reducing sw_FinBERT to 0.3 increases MAE by ~ 8%. ⎫ Assigning equal weights decreases ± 1.5% tolerance accuracy by ~ 5–7%. |
Volatility Factor (V) | Scaling term for sentiment influence, derived from historical price volatility. | Observation: • The volatility factor V plays a regulatory role: • Low volatility (V ≈ 0.1): less influence from sentiment • High volatility (V ≈ 0.7): sentiment scores strongly impact prediction • Insights: ⎫ Fixing V at 0.5 leads to average performance. ⎫ Adaptive V (calculated from rolling std dev) improves tolerance accuracy by ~ 10%. |
GA Population Size | Number of individuals in each Genetic Algorithm generation. | ⎫ Value Range Tested: 50 to 200 ⎫ Larger p → better convergence but slower. Best: 100–150 |
GA Mutation Rate | Controls diversity and exploration of weight combinations. | Value Range Tested: ⎫ Low (< 0.05) → stuck in local minima; High (> 0.15) → erratic predictions |
Crossover Rate | Probability of combining parent solutions in GA | ⎫ Value Range Tested: 0.7 to 0.9 ⎫ Minor influence. 0.8 is stable. ⎫ Optimal range: 0.75–0.85 ensures convergence and diversity |
Parameter | Highly Sensitive? | Impact on Performance | Recommendation |
|---|---|---|---|
Model Weights (w₁–w₅) | Yes | R² and RMSE change significantly | Use GA-based dynamic optimization |
Sentiment Weights | Yes | Affects tolerance and directional accuracy | Prioritize |
Volatility Scaling (V) | Yes | Improves adaptability under market shifts | Use rolling std-dev for calculation |
GA Mutation Rate | Medium | Affects optimization convergence | Set between and |
GA Population Size | Medium | Impacts computation time and fitness | Optimal between individuals |
Dataset | Best Model | R² Score | RMSE | Tolerance ± 2% |
|---|---|---|---|---|
TCS | Hybrid | 0.9149 | 38.59 | 92.86% |
Infosys | Hybrid | 0.9239 | 22.00 | 85.71% |
HCL Tech | Hybrid | 0.9332 | 20.23 | 82.14% |
Wipro | Hybrid | 0.9769 | 2.00 | 100% |
Persistent Systems | Hybrid | 0.973 | 55.19 | 92.59% |
Sasken Technologies | Hybrid | 0.8828 | 32.53 | 51.85% |
Policy Bazar | Hybrid | 0.9724 | 16.97 | 88.89% |
Quick Heal | Hybrid | 0.8406 | 4.29 | 81.48% |
Test Name | Purpose | Result |
|---|---|---|
Friedman Test | Tests ranking differences among all models across datasets | - Significant difference |
Wilcoxon Signed-Rank Test | Pairwise test: Hybrid vs. each baseline model on RMSE | - Hybrid significantly better |