This entry provides a comprehensive stateoftheart investigation of the recent advances in data science in emerging economic applications. The analysis is performed on the novel data science methods in four individual classes of deep learning models, hybrid deep learning models, hybrid machine learning, and ensemble models. Application domains include a broad and diverse range of economics research from the stock market, marketing, and ecommerce to corporate banking and cryptocurrency. Prisma method, a systematic literature review methodology, is used to ensure the quality of the survey. The findings reveal that the trends follow the advancement of hybrid models, which outperform other learning algorithms. It is further expected that the trends will converge toward the evolution of sophisticated hybrid deep learning models.
Due to the rapid advancement of databases and information technologies, and the remarkable progress in data analysis methods, the use of data science (DS) in various disciplines, including economics, has been increasing exponentially ^{[1]}. Advancements in data science technologies for economics applications have been progressive with promising results ^{[2]}^{[3]}. Several studies suggest that data science applications in economics can be categorized and studied in various popular technologies, such as deep learning, hybrid learning models, and ensemble algorithms ^{[4]}. Machine learning (ML) algorithms provide the ability to learn from data and deliver indepth insight into problems ^{[5]}. Researchers use machine learning models to solve various problems associated with economics. Notable applications of data science in economics are presented in Table 1. Deep learning (DL), as an emerging field of machine learning, is currently applied in many aspects of today’s society, from selfdriving cars to image recognition, hazard prediction, health informatics, and bioinformatics ^{[5]}^{[6]}. Several comparative studies have evaluated the performance of DL models with standard ML models, e.g., support vector machine (SVM), Knearest neighbors (KNN), and generalized regression neural networks (GRNN) in economic applications. The evolution of DS methods has progressed at a fast pace, and every day, many new sectors and disciplines are added to the number of users and beneficiaries of DS algorithms. On the other hand, hybrid machine learning models consist of two or more single algorithms and are used to increase the accuracy of the other models ^{[7]}. Hybrid models can be formed by combining two predictive machine learning algorithms or a machine learning algorithm and an optimization method to maximize the prediction function ^{[8]}. It has been demonstrated that the hybrid machine learning models outperform the single algorithms, and such an approach has improved the prediction accuracy ^{[9]}. Ensemble machine learning algorithms are one of the supervised learning algorithms that use multiple learning algorithms to improve learning processes and increase predictive accuracy ^{[10]}. Ensemble models apply different training algorithms to enhance training and learning from data ^{[11]}.
Table 1. Examples of notable classic machine learning methods applied in economicsrelated fields.
Sources 
Machine Learning Models 
Objectives 
Lee et al. ^{[12]} 
Support Vector Regression (SVR) 
Anomaly Detection 
Husejinović ^{[13]} 
Naive Bayesian And C4.5 Decision Tree Classifiers 
Credit Card Fraud Detection 
Zhang ^{[14]} 
Improved BP Neural Network 
Aquatic Product Export Volume Prediction 
Sundar and Satyanarayana ^{[15]} 
Multilayer Feed Forward Neural Network 
Stock Price Prediction 
Hew et al. ^{[16]} 
Artificial Neural Network (ANN) 
Mobile Social Commerce 
Abdillah and Suharjito ^{[17]} 
Adaptive NeuroFuzzy Inference System (ANFIS) 
EBanking Failure 
Sabaitytė et al. ^{[18]} 
Decision Tree (DT) 
Customer Behavior 
Zatevakhina, Dedyukhina, and Klioutchnikov ^{[19]} 
Deep Neural Network (ANN) 
Recommender Systems 
Benlahbib and Nfaoui ^{[20]} 
Naïve Bayes and Linear Support Vector Machine (LSVM) 
Sentiment Analysis 
Various works exist on the stateoftheart of DS methods in different disciplines, such as image recognition ^{[21]}, animal behavior ^{[22]}, renewable energy forecasting ^{[23]}. Hybrid methods have also been investigated in various fields, including financial time series ^{[24]}, solar radiation forecasting ^{[25]}, and FOREX rate prediction ^{[26]}, while ensemble methods have been mostly in the fields, ranging from breast cancer ^{[27]}, image categorization^{[28]}, electric vehicle user behavior prediction ^{[29]}, and solar power generation forecasting ^{[30]}. Exploring the scientific databases such as Thomson Reuters WebofScience (WoS) shows an exponential rise in using both DL and ML in economics. The results of an inquiry of essential ML and DL in the emerging applications to economics over the past decade is illustrated in Figure 1. Even though many researchers have applied DS methods to address different problems in the field of economics, these studies are scattered. At the same time, no single study provides a comprehensive overview of the contributions of DS in economicrelated fields. Therefore, the current study is conducted to bridge this literature gap. In other words, the main objective of this study is to investigate the advancement of DS in three parts: deep learning methods, hybrid deep learning methods, and ensemble machine learning techniques in economicsrelated fields. The present work aims to answer the following research questions. (1) what are the emerging economics domains with the involvement of data science technologies? (2) what are the popular data science models and applications in these domains?
Figure 1. Rapid rise in the applications of data science in economics.
Figure 2 shows that this study's database consists of 57 articles that were analyzed and categorized according to two criteria: (1) research/application area, and (2) the method type. Based on the review of articles by application, it was found that these articles were designed to address the issues of five different applications, namely the Stock Market (37 articles), Marketing (6 articles), Ecommerce (8 articles), Corporate Bankruptcy (3 articles), and Cryptocurrency (3 articles) (Tables 2–7). In addition, the articles were analyzed by the type of method, revealing that 42 unique algorithms were employed among the 57 reviewed articles (see Figure 3). It was further found that 9 articles used 9 single DL models (Table 8), 18 hybrid deep learning (HDL) models (Table 9), 7 hybrid machine learning models (Table 10), and 8 ensemble models (Table 11). In the following section, the identified applications and each of these methods are described in detail.
Figure 3. Notable methods of deep learning and hybrid deep learning models applied in economicsrelated fields; the size of the rectangle is proportional to the number of publications (source: WoS).
Applying deep learning in the stock market has become more common than in other economics areas, considering that most of the research articles reviewed in the present study are classified in this category (37 out of 57). Table 2 summarizes the articles that employed predictive models in stock market studies, including research objectives, data sources, and applied models of each article. Investment in the stock market is profitable, while the higher the profit, the higher the risk. Therefore, investors always try to determine and estimate the stock value before any action. The stock value is often influenced by uncontrollable economical and political factors that make it notoriously difficult to identify the future stock market trends. Not only is the nature of the stock market so volatile and complex, but the financial time series data are also noisy and nonstationary. Thus, the traditional forecasting models are not reliable enough to predict the stock value. Researchers are continuously seeking new methodologies based on DS algorithms to enhance the accuracy of such predictions. Forecasting stock price was found to be the objective of 29 out of 37 articles. Other studies aimed at applying DS in sentiment analysis, or the analysis of the context of texts to extracts subjective information, to identify future trends in the stock market. In addition, portfolio management, algorithmic trading (i.e., using a preprogrammed automated system for trading), automated stock trading, socially responsible investment portfolios, the S&P 500 index trend prediction, and exchangetradefund (EFT) options prices prediction were the objectives of other articles that projected to employ DS methods. Financial time series served as the data source of all these studies, except for the studies aimed at sentiment analysis, which used different data sources, such as social media and financial news.
LSTM
Long shortterm memory (LSTM) networks are a special kind of recurrent neural network (RNN) that can overcome the main issue of RNN, i.e., vanishing gradients using the gates to retain relevant information and discard unrelated details selectively. The structure of an LSTM neural network is shown in Figure 4, which is composed of a memory unit , a hidden state and three types of gates, where indexes the time step. Specifically, for each step , LSTM receives an input and the previous hidden state then calculates the activation of the gates. Finally, the memory unit and the hidden state are updated. The computations involved are described below:
where , , and denote the weights of inputs; and and indicate weights of recurrent output and biases, respectively. The subscripts , , and represent the forget, input, and output gate vectors, respectively; , , and denote the biases; and is the elementwise multiplication.
Figure 4. The structure of the long shortterm memory (LSTM) network
Many researchers have strived to forecast the stock value relying on the LSTM algorithm, either a single long shortterm memory (LSTM) or a hybrid model of LSTM. Adapting the LSTM algorithm, Moon and Kim ^{[31]}^{[32]} proposed an algorithm to predict the stock market index and volatility. Fischer and Krauss ^{[33]} expand the LSTM networks to forecast outofsample directional movements in the stock market. The comparison between the performance of their model with the random forest (RF), deep neural network (DNN), and logistic regression classifier (LOG) illustrates the remarkable outperformance of the LSTM model. Tamura et al. ^{[34]} introduced a twodimensional approach to predict the stock values in which the technical financial indices of the Japanese stock market were entered as input data to the LSTM for the prediction, then the data on financial statements of the related companies were retrieved and added to the database. Wang et al. ^{[35]} tried to find the best model to predict portfolio management's financial timeseries to optimize portfolio formation. They compared the results of LSTM against SVM, RF, DNN, and the autoregressive integrated moving average model (ARIMA) and realized that LSTM is more suitable for financial timeseries forecasting. Using LSTM, Fister et al. ^{[36]} designed a model for automated stock trading. They argue that the performance of LSTM is remarkably higher than the traditional trading strategies, such as passive and rulebased trading strategies. In their case studies, the German bluechip stock and BMW in the period between 2010 and 2018 formed the data sources.
In addition, there is much evidence in the literature that hybrid LSTM methods outperform the other single DL methods ^{[37]}. In the application of the stock market, LSTM has been combined with different methods to develop a hybrid model. For instance, Tamura et al. ^{[34]} used LSTM to predict stock price and reported that the accuracytest results outperform other literature models. Employing optimal long shortterm memory (OLSTM), Agrawal et al. ^{[38]} proposed a model for the stock price prediction using correlationtensor, which is formed by stock technical indicators (STIs) to optimize the deep learning function. As a result, two predictive models were developed, namely one for price trend prediction and the other for making the buysell decision at the end of the day.
Integrating wavelet transforms (WT), stacked autoencoders (SAEs), and LSTM, Bao, Yue, and Rao [39] established a new method to predict the stock price. In the first stage, WT first eliminates noises to decompose the stock price time series, then predictive features for the stock price are created by SAEs in the next stage. Finally, the LSTM is applied to predict the next day's closing price based on the previous stage's features. Bao et al. ^{[39]} claim that their model outperforms stateoftheart literature models in terms of predictive accuracy and profitability performance. To cope with nonlinearity and nonstationary characteristics of financial time series, Yan and Ouyang ^{[40]} integrate wavelet analysisLSTM (WALSTM) to forecast the daily closing price of the Shanghai Composite Index. Results show that their proposed model outperformed multiple layer perceptron (MLP), SVM, and KNN in finding the patterns in the financial time series. Vo et al. ^{[41]} used a multivariate bidirectionalLSTM (MBLSTM) to develop a deep responsible investment portfolio (DRIP) model for the prediction of stock returns for socially responsible investment portfolios. They applied the deep reinforcement learning (DRL) model to retrain neural networks. Fang, Chen, and Xue ^{[42]} developed a methodology to predict the exchangetradefund (EFT) options prices. Through integrating LSTM and support vector regression (SVR), they produce two models of LSTMSVR for modeling the final transaction price, buy price, highest price, lowest price, volume, historical volatility, and the implied volatility of the time segment. They predicted the price with promising results. In the second generation of LSTMSVR, the hidden state vectors of LSTM and the seven factors affecting the option price are considered the SVR’s inputs. They also compare the results with the LSTM and RF models, where the proposed model outperforms other methods.
Table 2. Notable machine learning and deep learning methods in stock market.
Source 
Modeling Methods 
Data Source 
Research Objective 
Wang et al. ^{[35]} 
LSTM Comparing with SVM, RF, DNN, and ARIMAs 
Financial Time Series 
Portfolio management 
Lei et al. ^{[43]} 
timedriven featureaware and DRL 
Financial Time Series 
Algorithmic trading 
Vo et al. ^{[41]} 
Multivariate Bidirectional LSTM Comparing with DRL 
Financial Time Series 
Socially Responsible Investment Portfolios 
Sabeena and Venkata Subba Reddy ^{[44]} 
GRU–CNN 
Financial Time Series 
Stock Price Prediction 
Das and Mishra ^{[45]} 
Adam optimizerMDNN 
Financial Time Series 
Stock Price Prediction 
Go and Hong ^{[46]} 
DNN 
Financial Time Series 
Stock Price Prediction 
Agrawal et al. ^{[38]} 
OLSTMSTI 
Financial Time Series 
Stock Price Prediction 
Gonçalves et al. ^{[47]} 
CNN comparing with DNNC and LSTM 
Financial Time Series 
Stock Price Prediction 
Moews et al. ^{[48]} 
DNNSLR 
Financial Time Series 
Stock Price Prediction 
Song et al. ^{[49]} 
DNN 
Financial Time Series 
Stock Price Prediction 
Fang et al. ^{[42]} 
LSTMSVR comparing with RF and LSTM 
Financial Time Series 
Exchangetradefund (EFT) Options Prices Prediction 
Long et al. ^{[50]} 
MFNN (CNN and RNN) 
Financial Time Series 
Stock Price Prediction 
Fister et al. ^{[36]} 
LSTM 
Financial Time Series 
Automated Stock Trading 
Rajesh ^{[51]} 
RF, SVM, and KNN 
Financial Time Series 
Stock Price Prediction 
Moon and Kim ^{[32]} 
LSTM 
Financial Time Series 
Stock Price Prediction 
Sim Kim, and Ahn ^{[52]} 
CNN comparing with ANN and SVM 
Financial Time Series 
Stock Price Prediction 
Agrawal et al. ^{[53]} 
LSTMSTIs 
Financial Time Series 
Stock Price Prediction 
Tashiro et al. ^{[54]} 
CNN 
Financial Time Series 
Stock Price Prediction 
Sirignano and Cont ^{[55]} 
LSDL 
Financial Time Series 
Stock Price Prediction 
Weng et al. ^{[56]} 
BRT comparing with NNRE, SVRE, and RFR 
Financial Time Series 
Stock Price Prediction 
Preeti et al. ^{[57]} 
ELMAE comparing with GARCH, GRNN, MLP, RF, and GRDH 
Financial Time Series 
Stock Price Prediction 
Sohangir et al. ^{[58]} 
CNN comparing with doc2vec and LSTM 
Social media 
Sentiment Analysis 
Fischer and Krauss ^{[33]} 
LSTM comparing with RF, DNN, and LOG 
Financial Time Series 
Stock Price Prediction 
Lien Minh et al. ^{[59]} 
twostream GRU 
Financial news 
Sentiment Analysis 
Das et al. ^{[60]} 
DNN 
Financial Time Series 
The S&P 500 Index Trend Prediction 
Yan and Ouyang ^{[40]} 
wavelet analysis with LSTM, comparing with SVM, KNN, and MLP 
Financial Time Series 
Stock Price Prediction 
Kim et al. ^{[61]} 
MACN 
Financial Time Series 
Stock Price Prediction 
FaghihiNezhad and MinaeiBidgoli ^{[62]} 
ELANN 
Financial Time Series 
Stock Price Prediction 
Tamura et al. ^{[34]} 
LSTM 
Financial Time Series 
Stock Price Prediction 
Chong et al. ^{[63]} 
DNN comparing with PCA, Autoencoder, and RBM 
Financial Time Series 
Stock Price Prediction 
Dingli and Fournier ^{[64]} 
CNN 
Financial Time Series 
Stock Price Prediction 
Singh and Srivastava ^{[65]} 
(2D)2PCA–DNN comparing with RBFNN 
Financial Time Series 
Stock Price Prediction 
Bao et al. ^{[39]} 
WTSAEsLSTM 
Financial Time Series 
Stock Price Prediction 
Shekhar and Varshney ^{[66]} 
GASVM 
Financial Time Series 
Stock Price Prediction 
Ahmadi et al. ^{[67]} 
ICA SVM 
Financial Time Series 
Stock Price Prediction 
Ebadati and Mortazavi ^{[68]} 
GAANN 
Financial Time Series 
Stock Price Prediction 
Johari et al. ^{[69]} 
GARCHSVM 
Financial Time Series 
Stock Price Prediction 
DNN
Deep neural network (DNN), which is composed of multiple nonlinear operations levels, and each layer only receives the connections from its previous training layer as shown in Figure 5, adapted from [49]. Suppose be the input data, and be a filter bank, where is the number of layers. The multilayer features of the DNN can be represented as

Currently, DNN has been widely applied in the stock market to identify the trends and patterns among the financial time series data. Go and Hong [46] used the DNN method to predict the stock value. They first trained the method by the time series data and then tested and confirmed their model's predictability. Song et al. [49] developed DNN using 715 novel inputfeatures to forecast the stock price fluctuation. They also compared the performance of their model with the other models that include simple pricebased inputfeatures. For predicting the stock market behavior, Chong, Han, and Park [63] examined the performance of DNN. They consider highfrequency intraday stock returns as the input in their model. They analyzed the predictability of principal component analysis (PCA), autoencoder, and RBM. According to their results, DNN has good predictability with the information they receive from the autoregressive mode residuals. Although applying the autoregressive model to the network's residuals may not contribute to the model's predictability. , Chong et al. [63] found out applying covariancebased market structure analysis to the predictive network remarkably increases the covariance estimation. Das et al. [60] used DNN to predict the future trends of the S&P 500 Index. Their results show that their model can poorly forecast the underlying stocks' behavior in the S&P 500 index. They believe that randomness and nonstationarity are the reasons that make hard the predictability of this index.
Figure 5. Structure of the deep neural network.
In addition, hybrid methods that are constructed based on DNN have been reported to be very accurate in the financial time series data. For example, Das and Mishra proposed an advanced model to plan, analyze, and predict the stock value, using a multilayer deep neural network (MDNN) optimized by Adam optimizer (AO) to find the patterns among the stock values. Moews et al. proposed a method to predict the stock market's behavior, as a complex system with a massive number of noisy time series. Their model integrates DNN and stepwise linear regressions (SLR). Moews et al. considered regression slopes as trend strength indicators for a given time interval. To predict the Google stock price, Singh and Srivastava compared two integrated models, i.e., 2directional 2dimensional principal component analysisDNN ((2D)2PCADNN) and (2D)2PCAradial basis function neural network (RBFNN). According to their results, the (2D) 2PCADNN model has higher accuracy in predicting the stock price. They also compared their results with the RNN model and reported that the predictability of (2D)2PCADNN outperforms RNN as well.
CNN
Convolutional neural network (CNN) is one of the most popular methods in deep learning and is widely applied in various fields ^{[70]}^{[71]}^{[72]}, such as classification, language processing, and object detection. A classical CNN structure is presented in Figure 6, adapted from ^{[73]}, which mainly consists of three components, i.e., convolution layers, pooling layers, and fully connected layers. Different layers have different roles in the training process and are discussed in more detail below:
Convolutional layer: This layer is composed of a set of trainable filters, which are used to perform feature extraction. Suppose is the input data and there are filters in convolutional layers, then the output of the convolutional layer can be determined as follows:
where and are the weight and bias, respectively; and denotes an activation function. represents the convolution operation.
Figure 6. Structure of the convolutional neural network (CNN).
Pooling layer: In general, this layer is used to decrease the obtained feature data and network parameters' dimensions. Currently, max pooling and average pooling are the most widely used methods. Let be a window size, then the average pooling operation can be expressed as follows:

where indicates the activation value at ; and is the total number of elements in .
Fully connected layer: Following the last pooling layer, the fully connected layer is utilized to reshape the feature maps into a 1D feature vector, which can be expressed as

where and denote the output vector and input features; and and represent the weight and bias of the fully connected layer, respectively.
Recently, many researchers have extensively applied CNN for predicting stock values using financial time series data. Sim et al. developed a model to predict the stock price by adapting CNN. Their results reveal that CNN's predictive performance demonstrated a better performance in forecasting stock price than ANN and SVM. Tashiro et al. first criticized the current models for the price prediction in the stock markets in that the properties of market orders are ignored. Therefore, they constructed a CNN architecture integrating orderbased features to predict the stock markets' midprice trends. Their results prove that adding the features of orders to the model increased its accuracy. Dingli and Fournier applied CNN to predict the future movement of stock prices and found that the predictive accuracy of their model was 65% in predicting the following month’s price and was 60% for the following week’s price. Gonçalves et al. compared the results of the prediction of CNN, LSTM, and deep neural network classifier (DNNC) for finding the best model to predict the price trends in the exchange markets. Their findings reveal that CNN, on average, has the best predictive power. Sohangir et al. compared the performance of several neural network models, including CNN, LSTM, and doc2vec, for sentiment analysis among the experts' posts and opinions in StockTwits to predict the movements in the stock markets. Their results disclose that CNN has the highest accuracy in predicting the sentiment of experts.
To increase CNN's accuracy, some researchers integrated CNN with other models and proposed new hybrid models. For example, by integrating gated recurrent unit (GRU) and CNN, Sabeena et al. introduced a hybrid DL model to predict financial fluctuations in the realtime stock market that is able to process the realtime data from online financial sites. GRUs are developed based on the RNN architectures. They represent a simpler implementation for LSTMs to address gradient functions and learn longrange dependencies. To predict the price movement from financial time series samples, Long et al. introduced an endtoend model called multifilters neural network (MFNN) that incorporates CNN and recurrent neural network (RNN).
Other Algorithms
In addition to LSTM, DNN, and CNN, other DS methods have been employed for predicting stock value using time series data. For example, Sirignano and Cont developed an LSDL model to study the USA market quotes and transactions. Their results disclose that there is a universal and stationary relationship between order flow history and price trends. Kim et al. proposed a multiagent collaborated network (MACN) model to optimize financial time series data, claiming that their model can share and generalize agents' experience in stock trading.
In addition, various other hybrid methods have been applied by the researchers for financial time series data. For instance, to predict the stock price trends, Lien Minh et al. developed and trained a twostream GRU (TSGRU) network and Stock2Vec model to analyze the sentiments in the financial news and their relationship with the financial prices, based on their belief that financial news and sentiment dictionaries affect the stock prices. Their findings support the outperformance of their model in comparison with the current models. Lien Minh et al. also claim that Stock2Vec is highly efficient in financial datasets. Lei et al. combined deep learning models and reinforcement learning models to develop a timedriven featureaware (TDFA) jointly deep reinforcement learning model (TFJDRL) for financial timeseries forecasting in algorithmic trading. Preeti et al. introduced an extreme learning machine (ELM)autoencoder (AE) model to identify patterns in the financial time series. They tested the accuracy of their model in the time series data of gold price and crude oil price and compared their results with those of generalized autoregressive conditional heteroskedasticity (GARCH), GRNN, MLP, RF, and group method of data handling (GRDH). The result of the mean square error (MSE) test proved that the performance of their model was higher than the existing methods.
In addition to the hybrid deep learning models, four articles applied hybrid machine learning models to financial time series data. Shekhar and Varshney integrated a hybrid model of genetic algorithmSVM (GVSVM) with sentiment analysis to predict the future of the stock market. Using quantitative empirical analysis, they proved that the combination of sentiment analysis with GVSVM increased the model's accuracy by 18.6% and reported the final model’s accuracy of about 89.93%. Ahmadi et al. compared the performance of two hybrid machine learning models in predicting the timing of the stock markets, using imperialist competition algorithmSVM (ICA SVM), and SVMGA. Their results exposed that SVMICA has a higher performance compared with SVMGA in the prediction of stock market trends for periods of 1–6 days. To predict stock prices using financial time series data, Ebadati and Mortazavi applied a hybrid model by integrating GAANN, where GA was employed to select ANN features and optimize parameters. Their study suggests that this hybrid machine learning model has an improved sum square error (SSE) (i.e., performance accuracy) by 99.99% and improved time (i.e., speed accuracy) by 90.66%. Johari et al. compared the accuracy and performance of GARCHSVM and GARCHANN models in the financial time series data for stock price forecasting. They found that GARCHSVM outperformed GARCHANN, SVM, ANN, and GARCH based on MSE and RMSE accuracy metrics.
Rajesh et al. used ensemble learning to predict future stock trends by applying heat maps and ensemble learning to the top 500 companies' financial data in the S&P stock exchange. Their results show that the combination of RF, SVM, and Kneighbors classifiers had the most accurate results, and the accuracy of the proposed model was 23% higher than a single classifier labeling prediction model. Weng et al. aimed to design a financial expert system to forecast shortterm stock prices. For data analysis and predicting stock prices, they employed four machine learning ensemble methods, namely neural network regression ensemble (NNRE), support vector regression ensemble (SVRE), boosted regression tree, and random forest regression (RFR). Using Citi Group stock ($C) data, they were able to forecast the oneday ahead price of 19 stocks from different industries. Weng et al. claim that the boosted regression tree (BRT) outperformed other ensemble models with a considerably better mean absolute percent error (MAPE) than those reported in the literature. FaghihiNezhad and MinaeiBidgoli employed ensemble learning and ANN to develop a twostage model to predict the stock price. They first predicted the next price movement's direction and then created a new training dataset to forecast the stock price. They used a genetic algorithm (GA) optimization and particle swarm optimization (PSO) to optimize the results of each stage. The results ultimately reveal that the accuracy of their model in prediction of stock price outperformed other models in the literature.
Reviewing the articles categorized in the stock market category reveals that, although the articles' research objectives are different, most utilized financial timeseries data (35 out of 37 articles). Only two articles used financial news and social media as the data source to determine future trends in the stock market (see Table 3).
Table 3. Classification of articles using data science by research purpose and data source in the stock market section.
Research Objective 
Data Source 
Number of Documents 
Stock Price Prediction 
Financial Time Series 
29 
Sentiment Analysis 
Financial News, Social Media 
2 
Portfolio management 
Financial Time Series 
1 
Algorithmic trading 
Financial Time Series 
1 
Socially Responsible Investment Portfolios 
Financial Time Series 
1 
Automated Stock Trading 
Financial Time Series 
1 
The S&P 500 Index Trend Prediction 
Financial Time Series 
1 
Exchangetradefund (EFT) Options Prices Prediction 
Financial Time Series 
1 
Studying the purpose of the articles discloses that DS algorithms were mostly used for the purpose of studying customer behavior and promotional activities, which is why these articles are classified in group labeled marketing. As seen in Table 4, two studies applied a single DL method, and three used hybrid DL methods. In addition, these studies used various data sources, such as customer time series data, case studies, and social media. For example, Paolanti et al. ^{[74]} employed deep convolutional neural network (DCNN) to develop a mobile robot, socalled ROCKy, to analyze realtime store heat maps of retail store shelves for detection of the shelfoutofstock (SOOS) and promotional activities during working hours. Dingli, Marmara, and Fournier ^{[75]} investigated solutions to identify the patterns and features among transactional data to predict customer churn within the retail industry. To do so, they compared the performance of CNN and restricted Boltzmann machine (RBM), realizing the RBM outperformed in customer churn prediction.
Table 4. Notable machine learning and deep learning methods in Marketing.
Source 
Modeling Methods 
Data Source 
Research Objective 
Ładyżyński et al. ^{[76]} 
RF–DNN 
Time Series data of Customers 
Customer Behavior 
Ullah et al. ^{[77]} 
RF 
Time Series data of Customers 
Customer Behavior 
Paolanti et al. ^{[74]} 
DCNN 
Primary Data 
Detection of Shelf Out of Stock (SOOS) and Promotional Activities 
Agarwal ^{[78]} 
RNNsCNNs 
Social media 
Sentiment Analysis 
Shamshirband et al. ^{[79]} 
SNCFM 
Social media 
Customer behavior 
Dingli et al. ^{[75]} 
RBM 
Primary Data 
Customer behavior 
On the other hand, RFDNN, RNNCNN, and similarity, neighborhoodbased collaborative filtering model (SNCFM) are hybrid models that researchers used to study customer behavior. Ładyżyński et al. ^{[76]}, for instance, employed Random Forest (RF) and DNN methods and customers’ historical transactional data to propose a hybrid model that can predict the customers’ willingness to purchase credit products from the banks. Ullah et al. ^{[77]} used the RF algorithm to predict churn customers and to formulate strategies for customer relationship management to prevent churners. Ullah et al. ^{[77]} explained that a combination of churn classification, utilizing the RF algorithm and customer profiling using kmeans clustering, increased their model’s performance. Agarwal ^{[78]} integrated RNN and CNN to develop a model for sentiment analysis. According to Agarwal ^{[78]}, sentiment analysis is the best approach to receive the customers’ feedback. He tested his proposed model using social media data and believes that the result of the sentiment analysis provides guidance to the business to improve the quality of their service and presents evidence for the startups to improve customer experience. Shamshirband et al. ^{[79]} proposed SNCFM to predict consumer preferences according to the similarity of features of users and products that are acquired from the Internet of Things and social networks.
Another category that emerged after reviewing the articles is labeled ecommerce, where the authors employed data science models to address problems in the ecommerce sector. A summary of these studies is presented in Table 5. Based on the GRU model, Lei ^{[80]} designed a neural network investment quality evaluation model to support the decisionmaking related to investment in ecommerce. Their proposed model is able to evaluate different index data that provides a better picture to investors. Leung et al. ^{[81]} argue that the ability to handle orders and logistics management is one of the major challenges in ecommerce. Therefore, using a hybrid autoregressiveadaptive neurofuzzy inference system (ARANFIS), they developed a prediction model for B2C ecommerce order arrival. According to their results, their proposed hybrid model can successfully forecast eorder arrivals. Cai et al. ^{[82]} used deep reinforcement learning to develop an algorithm to address allocating impression problems on ecommerce websites, such as www.taobao.com, www.ebay.com, and www.amazon.com. In this algorithm, the buyers are allocated to the sellers based on the buyer’s impressions and the seller’s strategies in a way that maximizes the income of the platform. To do so, they applied GRU, and their findings show that GRU outperforms the deep deterministic policy gradient (DDPG). Ha, Pyo, and Kim ^{[83]} applied RNN to develop a deep categorization network (Deep CN) for item categorization in ecommerce, which refers to classifying the leaf category of items from their metadata. They used RNN to generate features from text metadata and categorize the items accordingly. Xu et al. ^{[84]} designed an advanced credit risk evaluation system for ecommerce platforms to minimize the transaction risk associated with buyers and sellers. To this end, they employed a hybrid machine learning model combined with decision treeANN (DTANN) and found that this model has high accuracy and outperforms other hybrid machine learning models, such as DTLogistic Regression and DRdynamic Bayesian network.
Selling products online has unique challenges, for which data science has been able to provide solutions. To increase the buyer's trust in the quality of the products and to buy online, Saravanan and Charanya ^{[85]} designed an algorithm where the products are categorized according to several criteria, including reviews and ratings of other users. By integrating a hybrid feature extraction method principle component analysis (PCA) and tdistributed stochastic neighbor embedding (tSNE) with SVM using lexiconbased method, Saravanan and Charanya ^{[85]} also proposed a model that separates the products from the large collection of different products based on characteristics, best product ratings, and positive reviews. Wang, Mo, and Tseng ^{[86]} used RNN to develop a personalized product recommendation system on ecommerce websites. The result of their model disclosed the outperformance of RNN to Knearest neighbors (KNN). Wu and Yan ^{[87]} claim that the main assumption of the current production recommender models for ecommerce websites is that all historical data of users are recorded, while in practice, many platforms fail to record such dada. Therefore, they devised a listwise DNN (LWDNN) to model the temporal online behaviors of users and provide recommendations for anonymous users.
Table 5. Notable machine learning and deep learning methods in ecommerce.
Source 
Modeling Methods 
Data Source 
Research Objective 
Lei ^{[80]} 
GRU 
Financial Time Series 
Investment Quality Evaluation Model 
Leung et al. ^{[81]} 
ARANFIS 
Primary Data 
Order Arrival Prediction 
Cai et al. ^{[82]} 
GRU 
Customers Time Series 
Impression Allocation Problem 
Ha et al. ^{[83]} 
RNN 
Primary Data 
Item Categorization 
Xu et al. ^{[84]} 
DT—ANN 
Credit Data 
Dynamic Credit Risk Evaluation 
Saravanan and Charanya ^{[85]} 
PCA tSNESVM 
Primary Data 
Product Recommendation 
Wang et al. ^{[86]} 
RNN 
Primary Data 
Product Recommendation 
Wu and Yan ^{[87]} 
LWDNN 
Customers Time Series 
Product Recommendation 
The decisionmaking process related to investing in the cryptocurrencies is similar to investing in the stock market, where the prediction of future value is very determinant and effective on the investment decisions. Applying machine learning and DL models to predict the trends of cryptocurrency prices is an attractive research problem that is emerging in the literature (see Table 6). For example, Lahmiri and Bekiros ^{[88]}applied deep learning methods for the prediction of the price of cryptocurrencies, including Bitcoin, Digital Cash, and Ripple, and compared the predictive performance of LSTM and GRNN. Their findings disclose that the LSTM model has a better performance in comparison with GRNN. Altan, Karasu, and Bekiros ^{[89]} claim that integrating LSTM and empirical wavelet transform (EWT) improves the performance of LSTM in forecasting the digital currency price by testing their proposed model using Bitcoin, Ripple, Digital Cash, and Litecoin time series data. Jiang and Liang ^{[90]} developed a CNN model to predict the price of Bitcoin as a cryptocurrency example. They train their proposed model using historical data of financial assets prices and used portfolio weights of the set as the output of their model.
Table 6. Notable machine learning and deep learning methods in Cryptocurrency.
Source 
Modeling Methods 
Data Source 
Research Objective 
Lahmiri and Bekiros ^{[88]} 
LSTM comparing with GRNN 
Financial Time Series 
Cryptocurrencies Price prediction 
Altana et al. ^{[89]} 
LSTMEWT 
Financial Time Series 
Cryptocurrencies Price prediction 
Jiang and Liang ^{[90]} 
CNN 
Financial Time Series 
Cryptocurrencies Price prediction 
Corporate bankruptcy prediction has become an important tool to evaluate the future financial situation of a company. Utilizing machine learningbased methods is widely recommended to address bankruptcy prediction problems. Such as done by Chen, Chen, and Shi ^{[91]} , who utilized bagging and boosting ensemble strategies and develop two models of Baggedproportion support vector machines (pSVM) and boostedpSVM. Using datasets of UCI and LibSVM, they test their models and explain that ensemble learning strategies increased the performance of the models in bankruptcy prediction. Lin, Lu, and Tsai ^{[92]} believe that finding the best match of feature selection and classification techniques improves the prediction performance of bankruptcy prediction models. Their results reveal that the genetic algorithm as the wrapper‐based feature selection method and the combination of the genetic algorithm with the naïve Bayes and support vector machine classifiers had a remarkable predictive performance. Lahmiri et al. ^{[93]}, to develop an accurate model for forecasting corporate bankruptcy, compare the performance of different ensemble systems of AdaBoost, LogitBoost, RUSBoost, subspace, and bagging. Their finding reveals that the AdaBoost model has been effective in terms of shorttime data processing and low classification error, and limited complexity. Faris et al. ^{[94]} investigate the combination of resampling (oversampling) techniques and multiple features of election methods to improve the accuracy of bankruptcy prediction methods. According to their results, employing SMOTE oversampling technique and AdaBoost ensemble method using reduced error pruning (REP) tree provides reliable promising results to bankruptcy prediction. A summary of these research articles is presented in Table 7.
Table 7. Notable machine learning and deep learning methods in corporate bankruptcy prediction.
Source 
Modeling Methods 
Data Source 
Research Objective 
Chen et al. ^{[91]} 
BaggedpSVM and BoostedpSVM 
UCI and LibSVM datasets 
bankruptcy prediction 
Lin et al. ^{[92]} 
Genetic Algorithm with the Naïve Bayes and SVM classifiers 
Australian credit, German credit, and Taiwan bankruptcy datasets 
bankruptcy prediction 
Lahmiri et al. ^{[93]} 
AdaBoost 
University of California Irvine (UCI) Machine Learning Repository 
bankruptcy prediction 
Faris et al. ^{[94]} 
SMOTEAdaBoostREP Tree 
Infotel database 
bankruptcy prediction 