A comparative study on artificial neural networks and random forests for stock market prediction

Size: px
Start display at page:

Download "A comparative study on artificial neural networks and random forests for stock market prediction"

Transcription

1 DEGREE PROJECT IN TECHNOLOGY, FIRST CYCLE, 15 CREDITS STOCKHOLM, SWEDEN 2016 A comparative study on artificial neural networks and random forests for stock market prediction VICTOR ERIKSSON THUJEEPAN VARATHARAJAH KTH ROYAL INSTITUTE OF TECHNOLOGY SCHOOL OF COMPUTER SCIENCE AND COMMUNICATION

2 En komparativ studie mellan artificial neural networks och random forests med avseende på aktieprognoser VICTOR ERIKSSON THUJEEPAN VARATHARAJAH Degree Project in Computer Science, DD143X Supervisor: Atsuto Maki Examiner: Örjan Ekeberg May 11, 2016 CSC, KTH

3 Abstract This study investigates the predictive performance of two different machine learning (ML) models on the stock market and compare the results. The chosen models are based on artificial neural networks (ANN) and random forests (RF). The models are trained on two separate data sets and the predictions are made on the next day closing price. The input vectors of the models consist of 6 different financial indicators which are based on the closing prices of the past 5, 10 and 20 days. The performance evaluation are done by analyzing and comparing such values as the root mean squared error (RMSE) and mean average percentage error (MAPE) for the test period. Specific behavior in subsets of the test period is also analyzed to evaluate consistency of the models. The results showed that the ANN model performed better than the RF model as it throughout the test period had lower errors compared to the actual prices and thus overall made more accurate predictions.

4 Sammanfattning Denna studie undersöker hur väl två olika modeller inom maskininlärning (ML) kan förutspå aktiemarknaden och jämför sedan resultaten av dessa. De valda modellerna baseras på artificiella neurala nätverk (ANN) samt random forests (RF). Modellerna tränas upp med två separata datamängder och prognoserna sker på nästföljande dags stängningskurs. Indatan för modellerna består av 6 olika finansiella nyckeltal som är baserade på stängningskursen för de senaste 5, 10 och 20 dagarna. Prestandan utvärderas genom att analysera och jämföra värden som root mean squared error (RMSE) samt mean average percentage error (MAPE) för testperioden. Även specifika trender i delmängder av testperioden undersöks för att utvärdera följdriktigheten av modellerna. Resultaten visade att ANN-modellen presterade bättre än RF-modellen då den sett över hela testperioden visade mindre fel jämfört med de faktiska värdena och gjorde därmed mer träffsäkra prognoser.

5 Contents 1 Introduction Objective and problem statement Scope Outline Background Stock prediction Machine learning Artificial neural networks Random forests Decision tree Regression tree Random forest Method Data selection Technical analysis indicators used in the input vector ANN model RF model Performance metrics Root mean squared error Mean absolute percentage error Results Performance on WMT stock Performance on IXIC index Discussion and conclusion Discussion of results Discussion of method Amount of data/observations Partitioning of the data The choice of technical indicators

6 5.3 Comments on the reviewed sources Final conclusion Appendix A Biological neural networks 25 Bibliography 27 List of Figures 30 List of Tables 31 Acronyms & abbreviations 32

7 Chapter 1 Introduction Stock market prediction is an important field of interest among investors. Accurate predictions mean that one potentially can foresee events or trends thus making investments more profitable. Making such predictions is difficult due to the complex nature of the market which is influenced by a wide variety of factors. Several methods exist and technological methods are one main branch that has become increasingly popular in recent years. These kinds of methods utilize algorithmic models in order to make the predictions. The problem can be seen as a time series prediction problem which is the task of estimating a value based solely on previous values in a time series. Two examples of technological methods from machine learning (ML) that have been used are artificial neural networks (ANN) and random forests (RF) [1, 2, 3, 4, 5]. An ANN is a model that is inspired by the human nervous system and a RF is a collection of decision trees. Both these models are suitable for applications like stock prediction as they have the ability to model complex structures such as nonlinear patterns. These methods make predictions by analyzing existing data, building a model to reflect the underlying nature of the data and then uses this model to generalize on previously unseen data. There exist quite a lot of research on ANNs within the field of stock prediction which has yielded some interesting results [1]. RFs have also been researched within the area, but have generally not attracted the same amount of attention as ANNs. However, in Caruana and Niculescu-Mizil s [6] study where they did large scale empirical comparisons on a set of supervised learning algorithms, the overall performance of the RF model was ranked higher than the ANN model. Thus an interesting question emerges regarding how ANN and RF models compare when applied to stock prediction. The goal of this study is to compare the performance between optimized implementations of ANNs and RFs, respectively. 1.1 Objective and problem statement The objective of this thesis is to investigate and compare the predictive properties that artificial neural networks and random forests have on the stock market. The question to be answered is: 1

8 CHAPTER 1. INTRODUCTION How does artificial neural networks and random forests compare to one another when applied as one day prediction models on the stock market and which of the two has the best predictive performance? 1.2 Scope The ANNs and RFs will be implemented for the Walmart stock and the Nasdaq Composite stock index respectively. Due to the complexity of the models together with constraints in time and resources, the models will be implemented and optimized using built-in tools and classes in Matlab. To some extent, this also limits the level of optimization and alteration that can be made on the models. Performance will be measured with the chosen metrics, and the conclusion will thus be based on these as well. It is important to note that this study will strictly be of a comparative nature. There will be no analysis as to why one model performs better than the other. Instead, the study will draw its conclusion from comparing the performance results of both models on the chosen stocks. There will not be any guidelines in how one can earn money from the stock market using these models as this is not the aim of this study. 1.3 Outline The remainder of this report is outlined in the following fashion. Chapter 2 presents relevant background information on the topics of stock prediction, machine learning, ANNs and RFs. Chapter 3, presents the methodology of the conducted study in terms of data selection, how the input vectors were constructed and how the models were implemented and optimized. Chapter 3 also presents the metrics that are used to evaluate the performance of both models. Chapter 4 presents the results of the conducted study in terms of the performance metrics as well as graphs showing the overview of the results. Chapter 5 discusses the results, methodology, offers source criticism and presents the final conclusion to answer the proposed problem statement of section 1.1. Appendix A gives a brief description of biological neural networks. 2

9 Chapter 2 Background In this chapter, some of the key concepts regarding stock prediction, ML, ANNs and RFs are explained. 2.1 Stock prediction Stocks relate to a stake in certain kinds of companies in which one may invest money and be liable only in proportion to the amount invested. These are called limited companies and have a finite number of stocks that is partitioned into shares. When buying shares in such a company, one becomes a shareholder and thus an owner of some fraction of the company. The type of stock regulates the terms of ownership and therefore shareholders may have different influence on the business operation. [7] A limited company can be listed on a stock exchange which is a marketplace for trading shares but these can also be traded privately and the combination of all possible trading places are in the context of this thesis referred to as the stock market. The stock price at a moment in time is determined by the most recent price that a buyer and seller agrees upon. Consequently, one can say that the stock price is regulated by supply and demand [8]. The total amount of shares available in a company corresponds to the supply while many different factors affect the demand which may in turn have varying predictability. There are often several different prices to consider when analyzing a specific stock but in this thesis, the closing price is the only one used which is the price of the last transaction just before the stock exchange closes for the day. There are generally two branches of stock prediction which are fundamental analysis and technical analysis. Fundamental analysis focuses on financial factors such as a company s balance sheet, market positions, credit value etc. and uses these to make an estimate of the company s intrinsic value. Technical analysis on the other hand bases its estimates purely on historical data such as the stock price and different key indicators. [8] 3

10 CHAPTER 2. BACKGROUND 2.2 Machine learning Machine learning can be seen as a subfield of Artificial intelligence (AI). Where AI generally involves any computational model based on intelligence, ML concerns itself with developing algorithms with the ability to learn and improve themselves by being exposed to new data. In this regard ML is a natural part of AI as AI involves learning as well as other cognitive attributes often related to intelligence such as reasoning and perception. [9] A model of ML can be described as a function representing the nature of the input data. Moreover can ML be divided into two main branches which is supervised learning and unsupervised learning. In supervised learning, the training set of the algorithm contains data points of inputs as well as its desired outputs. In unsupervised learning, the desired outputs are unknown to the learning algorithm [10]. The branch used throughout this thesis is supervised learning because the implemented models have historical data points available in the form of dates in combination with stock prices. ML can be further divided into regression or classification models. Classification models output a class which is a value of a discrete set while regression models instead output a continuous value [10]. This makes regression the valid choice for the application of this thesis as the implementations must be able to output real valued stock prices. Ultimately, the goal of a ML model is to analyze a training set, find some underlying nature of the data and then be able to generalize this model to previously unseen data as well [11]. In supervised learning, this is done through bias-variance tradeoff which is the problem of simultaneously minimizing both the bias and the variance of a model. The bias is the difference between the predicted value and the actual value while the variance indicates to what degree the values are spread out [12]. Bias: B = 1 n n f k ˆf k (2.1) where f k are the true values, ˆf k are the predictions and n is the number of points in the data set. Sample variance: s 2 = 1 n f k n 1 f k (2.2) where f k are a value, f k are the mean of all values and n is the number of points in the data set. High bias may cause underfitting which means that relevant properties of the data are missed in the learning process. High variance may cause overfitting which is the reverse of underfitting and means that the model is overly sensitive to certain properties of the training data. Both underfitting and overfitting result in decreased performance regarding the ability to generalize well to unseen data. [10] The general learning approach of ML can roughly be described in three steps. The first step is to partition the total data set into three parts which is the training data, validation data and test data. The second and third steps is the hyperparameter optimization k=1 k=1 4

11 CHAPTER 2. BACKGROUND and model parameter optimization, respectively. Basically the hyperparameters are the parameters that are adjusted before the actual learning while the model parameters are adjusted during the learning. In this respect the hyperparameter optimization is a so called meta-optimization task meaning that it optimizes another optimizer which in this case is the learning algorithm itself. There are a wide variety of methods used for optimizing the hyperparameters and this study takes the approach of using crossvalidation together with loss function optimization. This means that the hyperparameters of a ML model are iteratively changed until the minimum of the chosen loss function, evaluated over the validation set, are found. The third step of the model parameter optimization can also be done in different ways which are the case for the two models in this study, see sections 3.3 and 3.4. [10] 2.3 Artificial neural networks ANNs are a collection of models inspired by the biological neural network (BNN) model of the human nervous system (see appendix A for a basic explanation of BNNs). Like BNNs, ANNs generally prove to be well suited for typical cognitive tasks such as pattern recognition and learning [13, 14]. Even though simplified compared to a BNN, the ANN models implement the basic structure of their biological equivalent and can be described as finite directed weighted graphs. Edges that are directed into vertices are inputs while edges directed out are outputs and the collection of the m inputs or n outputs can be seen as n- or m-dimensional vectors, respectively. ANN models have at least an input and an output layer and the number of vertices in these is determined by the dimensionality of the problem at hand. An ANN can be viewed as a function as described in section 2.2 and for the particular regression problem of this thesis, more specifically a function from R 6 to R. One of the simplest ANN models is the feed forward neural network (FFNN) which can be modeled as acyclic graphs. There are many other kinds of ANN models such as the FFNNs cyclic counterpart which is called Recurrent Neural Networks (RNN) but the FFNNs are more commonly used compared to its alternatives [15]. A more mathematically stringent formulation of a vertex in an ANN is that the jth vertex can be seen as a composite function F ( x, w) = (A T )( x, w) where T is the transfer function, A is the activation function, x = x 1,..., x n is the input vector and w = w 1,..., w n is a vector containing the edge weights for the corresponding inputs in x, see figure 2.1. [13, 16] The transfer function T for the jth vertex is most often a summation function defined as the dot product of the input and weight vectors, or equivalently the weighted sum of all the inputs, T = x w. The output of T is passed on to the activation function A that determines whether the artificial neuron is activated or not. The activation function may or may not use the threshold value (theta in figure 2.1) to perform the evaluation [13]. Different activation functions are used depending on the model but generally a distinction can be made between linear and nonlinear activation functions. The training of an ANN is done by the general principles of ML discussed in section 2.2, through 5

12 CHAPTER 2. BACKGROUND Figure 2.1: A basic outline of an artificial neuron. Source: org/wiki/file:artificialneuronmodel_english.png exposing the network iteratively to known data and adjusting model parameters like the edge weights correspondingly. One of the early ANNs, modeled in the 1960 s by Frank Rosenblatt, is called a single layer perceptron (SLP) and was a FFNN built on the concept of the perceptron algorithm [13, 17]. This algorithm can be defined as an artificial neuron that uses the Heaviside step function (2.3) as the activation function which means that the output is a binary value dependent on the value of the transfer function in comparison with the threshold value theta [13], see figure 2.1. { 0, θ < 0 H(θ) = (2.3) 1, θ 0 The learning of a SLP is usually done according to the delta rule which is an iterative gradient descent algorithm. This means that steps proportional to the negative gradient of the function are taken in order to find a minimum after a finite number of steps [17]. In the case of SLPs, the edge weights are first chosen randomly, then the desired output data is compared to the actual output data in a loss function and the weights are adjusted thereafter. A first limitation of this model is that the input data must be linearly separable, otherwise the training algorithm may not converge towards a solution in a finite number of steps which means that there is a large number of fundamental examples that SLPs are unable to learn [13]. A second limitation is that the activation function only can perform binary classification and thus regression applications cannot be modeled using SLPs. To overcome the limitations of the SLP model, the multilayer perceptron (MLP) was developed in the 1970 s which is a FFNN model that differs from SLPs in several respects. The MLP model contains except for the input and output layers an arbitrarily large number of hidden layers between these. MLPs also use more sophisticated learning methods based mainly on backpropagation (BP) that is a generalization of the delta rule used in SLPs. BP algorithms may aside 6

13 CHAPTER 2. BACKGROUND Figure 2.2: A Multilayer perceptron (MLP) with one hidden layer. Source: MultiLayerNeuralNetworkBigger_english.png from adjusting the edge weights also shift the activation function by using a so called bias vertex to allow it to trigger in a different manner [13]. The added layers and the BP algorithm provide the network with an internal structure that gives MLPs the ability to handle problems such as non-linearly separable ones that SLPs are unable to model appropriately [17, 13]. MLPs may also use other activation functions than the Heaviside step function which means that MLP models can be used effectively both for classification and regression problems. Common activation functions used in MLPs are different linear functions, hyperbolic tangent functions as well as the sigmoid function [16]. The sigmoid function is the primary function in the implementations of this thesis and can be seen below in (2.4). These kinds of non-constant activation functions are all continuous in the range y [0, 1] and this is vital for the functionality of the BP algorithm. Several different activation functions may be used in one single MLP model. S(t) = e t (2.4) There are a wide range of learning algorithms that implement BP with some additional optimization technique. The Levenberg-Marquardt algorithm (LMA) that is used in this thesis, utilizes the Gauss-Newton algorithm (GNA) as the optimization and MSE as the loss function. LMA is generally considered to be a fast and efficient choice among learning algorithms [18]. For ANNs, the training data is used by the learning algorithm to adjust the edge weights and the bias vertex of the network. The validation data are used during training to evaluate the loss function and determine when satisfactory performance of the network has been reached. It is also used to optimize the hyperparameters as described in sections 2.2 and 3.3. The test data is purely used for performance evaluation. 7

14 CHAPTER 2. BACKGROUND 2.4 Random forests Decision tree A decision tree is a model of supervised learning that can be represented by a tree graph. A decision tree is thus a set of questions organized in a hierarchical manner where given an input vector the decision tree estimates an unknown property by asking successive questions about its known properties [19]. Starting at the root, the interior nodes of the tree represent each question and the edges between them are the corresponding answers. As such, the question that is asked next depend on the previous answer. In following such a path, a decision is made once a terminal node, or leaf, has been reached. Figure 2.3: An illustrative example of a decision tree that is used to figure out whether a photo (the input vector) represents an indoor or an outdoor scene. Source: Criminisi et al. [19] (p.88) Regression tree A regression tree is as the name implies, a decision tree that is used for regression problems. The general idea is to use recursive partitioning on data that act in complicated nonlinear ways to get smaller partitions with lower variance allowing for better predictions [19]. The process involves partitioning the data space into smaller disjoint regions, and again partition those sub-regions recursively so that we get chunks of space that have simple predictive models contained in them (to predict unknown values). This recursive partitioning is what is represented in the regression tree. As such, each interior node represents a split as well as a partition of space following from earlier splits. The interior nodes guide a path to the terminal nodes which represents a cell of a partition, and which 8

15 CHAPTER 2. BACKGROUND has a predictive model attached to it. It is the predictive model contained in the terminal node that decides the actual form of the prediction and common models include using the constant mean or linear/polynomial regression [19]. Figure 2.4: Regression tree (a) An illustrative example of a 2-dimensional regression tree with binary splits. The predictive model in the terminal node takes the mean of all data points contained in that partition. (b) An illustrative example of the partitioned space and the containing data points. Source: regression-tree-using-gini%e2%80%99s The performance of a regression tree is mainly down to how well the tree is set up and partitioned. A common way to split a regression tree is to use a greedy approach where each split maximizes the variance reduction [19, 20]. This is done by initially choosing the data point and the split that maximizes the variance reduction, and set this as the root node. As there are finitely many data points in the data set, there are only finitely many splits that needs to be considered. The same process is then done recursively for each interior node. For instance, in a data space with a regression tree that does binary splits, the root node would represent the split on the data point that minimizes the sum of the variances on both partitions of the split. The interior nodes, which represent the subregions, would then go through the same process recursively. The recursive partitioning keeps going until a stopping criterion has been reached which is when the growing of the tree stops. A typical stopping criterion is when a partition has some minimal amount of observations (data points) contained in it. Thus the leaf with the predictive model represents the partition that has met the stopping criterion [19, 21]. A tree that has been overfitted may be a tree that has grown too much. This may occur due to having a stopping criterion where the amount of minimal observations are too low, leading the tree to have bad generalization. The process of pruning, meaning removing unreliable branches after growing the tree, may improve its ability to generalize [20]. 9

16 CHAPTER 2. BACKGROUND Random forest A RF is an ensemble learning technique that has shown to have better performance in generalization and overfitting avoidance in comparison to a single decision tree [19]. The introduction of RFs in classification and regression trees (CART) was first proposed in a paper by Leo Breiman [22]. As this study involves only using regression trees, the usage of the term random forest will throughout this thesis correspond to a random regression forest. A RF is a collection of randomly trained regression trees that are collectively used to output a prediction. In a RF, there is no need to prune each individual tree after growing them. The main reason being its use of the random subspace method in a collection of trees [22]. The random subspace method involves the following process when growing each individual tree: 1. Randomly choose a subset of the training data for training the individual regression tree. 2. At each node, choose a random subset of features, and only consider splitting on these features. The prediction of the RF model, is the mean of the predictions of all individual trees in the RF. The use of RF instead of a single regression tree have shown to considerably decrease the risk of overfitting. As the individual trees are all randomly different from one another, this leads to decorrelation between their individual predictions thus making the prediction of the RF generalize better [19, 22]. Hyperparameters that can affect the performance of a RF includes variables such as the number of trees, how many features that are considered at each node, and the stopping criterion [23]. 10

17 Chapter 3 Method To answer the proposed problem statement, this study was initially conducted by doing a literature study on the topics of stock market prediction, ANNs, RFs and ML in general. The literature study also involved reading available material on the subject of ML techniques for stock market prediction and in particular research material that used ANN and RF models as the models for prediction. Suitable input vectors for one-day stock prediction were also researched and it was decided that the use of technical analysis indicators would be a relevant choice. This decision was made after reading similar studies of ANN and RF based stock prediction using technical analysis indicators as well [2, 4, 5]. Both predictive models and the necessary operations to gain results were implemented in Matlab R2015a ( ). The steps taken in Matlab involved processing data, initializing ANN and RF models, training and testing the models for prediction as well as calculating performance metrics. 3.1 Data selection The input data set comes from the Walmart stock price (WMT) and the Nasdaq Composite stock index (IXIC). These stocks were chosen due to differences regarding some properties such as the closing prices and the volatility which gives a good basis for a comparison between different types of data. To gain fair and comparable results, the same amount of datasets needed to be gathered and used for both models. When researching available studies of stock prediction using ANNs and RFs, the literature study showed that the amount of gathered datasets has been of varying amounts. Paluch and Jackowska-Strumillo [2] used one year of data in their ANN based stock prediction model. In contrast to that Ticknor [1] used 734 days to gain optimal results in his study of an ANN based prediction model. Manojlovic and Stajduhar [4] used 1512 days of data in their paper on RF based stock forecasting. In this study, the training, validation and test data consisted of 254 days for both models as can be seen in figures 3.1 and 3.2. The data was gathered by initially fetching 274 days of closing price data from Yahoo Finance for WMT and IXIC in the timespan 11

18 CHAPTER 3. METHOD Figure 3.1: The actual prices of the WMT stock in the interval between Feb Feb Source: Matlab. January 5th February 4th They were then processed to 6-dimensional input vectors with 254 observations representing the timespan February 3rd February 4th The input vector consisted of technical indicators that were based on the closing prices of 5, 10 and 20 days prior to the day that is to be predicted (see section 3.2). The last 15 % of the gathered data was chosen as a subset to test performance. Thus the initial 85 % of the data was to be separated into training and validation sets. This was done by studying the properties of the gathered dataset for both stocks. The proposed methods aim was to include unusual spikes and dips in the training data. For instance, in the WMT stock when considering the initial 85 % of the 254 observations, there is a sudden dip in the closing price on day 177 and it reaches it s lowest price at day 199 before starting to recover. With the proposed method, it was reasonable to have these days in the training data. In such a manner, the partitioning for each stock was done as detailed below. The WMT stock was partitioned by setting 80 % as training data, 5 % was set as validation data and the remaining 15 % was set as test data to evaluate performance. The IXIC index was partitioned by setting 70 % as training data, 15 % was set as validation data and the remaining 15 % was set as test data to evaluate performance. 12

19 CHAPTER 3. METHOD Figure 3.2: The actual prices of the IXIC index in the interval between Feb Feb Source: Matlab. This partitioning also means that both the ANN and RF models for the respective stocks had the same subset of data for building their structure and the same subset to test their performance. Table 3.1 presents both partitioning schemes that was used. Training days Validation days Test days WMT IXIC Table 3.1: Table presenting the partitioning of the data sets for each stock. The respective stock had the same partitioning to build and evaluate both of its respective ANN and RF models. 3.2 Technical analysis indicators used in the input vector The most commonly used technical indicators are moving averages and oscillators [2, 4]. As such, the approach was to use 5, 10 and 20 day exponential moving averages (EMA) and rates of change (ROC) to construct the input vector. 13

20 CHAPTER 3. METHOD Exponential moving average (N = 5, 10, 20 days): EMA N (k) = C(k) + ac(k 1) + a2 C(k 2) a N 1 C(k N + 1) 1 + a + a a N 1 (3.1) where C(k) indicates the closing price on day k, and a = 1 (2/(N + 1)). Rate of change (N = 5, 10, 20 days): ROC N (k) = C(k) C(k N) C(k N) (3.2) where C(k) indicates the closing price on day k. The ROC determines the rate of price change in a given period. As such, a positive rate indicates an upgoing trend in the closing price while a negative rate indicates a downgoing trend. 3.3 ANN model As mentioned in the beginning of this chapter, the ANN model implementation as well as the pre- and post processing of data was done in Matlab where the implementation utilized the built in software neural network toolbox. The model is a feedforward three-layered MLP. This network configuration is common as one hidden layer is enough to model nonlinear functions while additional hidden layers pose an increased risk of overfitting depending on the nature of the data [3]. Furthermore, models with multiple hidden layers are more likely to fall into a local minimum when training the network [24]. As suggested from the dimensionality of the data, the input layer consists of six nodes and the output layer of a single node. Determining the optimal hyperparameter of the number of neurons in the hidden layer is not entirely straightforward and a variety of recommendations exist regarding the subject. The following are some commonly used rule-of-thumb methods [25]: The size of the hidden layer is somewhere between the sizes of the input and output layers. The size of the hidden layer is 2/3 of the size of the input layer plus the size of the output layer. The size of the hidden layer is less than twice the size of the input layer. The hyperparameter was empirically determined with these three recommendations as a guideline to be optimal at five nodes. This was done by cross-validation and loss function optimization as described in section 2.2 with RMSE as the loss function (see section 3.5.1). In this particular situation, the estimated optimal size coincides exactly with that of the second recommendation above. 14

21 CHAPTER 3. METHOD The activation functions used were a sigmoid function for the hidden layer and a linear function for the output layer. During the training of the network, the inputs and targets were normalized in the range [-1, 1] with the mapminmax function and the trained network was configured to output the unnormalized values for better interpretability. The stopping criterion used during training was that when the RMSE failed to decrease for six successive times, validation checks halted and thus training terminated. Figure 3.3: A schematic image reflecting the structure of the ANN model. Source: Matlab. For backpropagation, the Levenberg-Marquardt algorithm was used. It is recommended in the API as a first choice due to its speed compared to some other alternatives [18]. The Bayesian regularization algorithm was tested without any noticeable improvement in performance so Levenberg-Marquardt was deemed satisfactory. A schematic representation of the model is shown in figure RF model The RF models were implemented using the TreeBagger class in Matlab and the implementation follows Breiman s RF model. It uses the mean of observations in its terminal nodes to predict unknown values. The models for WMT and IXIC were individually tuned with the validation data to gain optimal performance. The hyperparameters tuned were NumTrees, NVarToSample, and MinLeafSize and are explained below. NumTrees - Amount of regression trees to grow in the RF model. MinLeafSize - Minimum number of observations at each leaf, also known as the stopping criteria. NVarToSample - The amount of features to be chosen at random for each split in the trees. NumTrees was set to be 100 trees for both WMT and IXIC. In general, more trees in a RF could lead to better performance but this performance gain eventually converges to a limiting value [22]. In his original paper, Breiman used 100 trees for his RF model. In this study, testing was done to see if more trees than 100 had any positive effects on 15

22 CHAPTER 3. METHOD the RMSE (see section 3.5.1) of the validation data. As there was little to no gain, 100 trees was used as the hyperparameter value for both RF models. MinLeafSize was set to be 5 observations for both WMT and IXIC. This is the default value of the TreeBagger class. In order to test if it was a suitable value, the performance was compared between models that used 5, 10 or 20 observations on the training and validation data. As 5 observations gave the lowest RMSE, it was chosen as the hyperparameter value. NVarToSample was set to 2 features for the WMT stock and 4 features for the IXIC index. The values were chosen by iterating through possible values on each stock and choosing the value that gave the lowest RMSE on the validation set. NumTrees MinLeafSIze NVarToSample WMT IXIC Table 3.2: Table presenting the hyperparameter values set for each RF model for the respective stocks. 3.5 Performance metrics The chosen performance metrics are the root mean square error (RMSE) and the mean absolute percentage error (MAPE). These are relevant measures of performance as they involve calculating the bias error and as such these metrics calculate the errors based on the actual price compared to the predicted price. The equations together with brief descriptions are presented in the subsections below Root mean squared error RMSE = 1 n n ( ˆf k f k ) 2 (3.3) where ˆf k are the predicted values, f k are the true values and n is the number of data points. The RMSE is the square root of the average of the sum of squared bias errors. k= Mean absolute percentage error MAPE = 1 n f k ˆf k n f k k=1 (3.4) where ˆf k are the predicted values, f k are the true values and n is the number of data points. 16

23 CHAPTER 3. METHOD The MAPE measures the size of the sum of bias errors in percentage terms. It is calculated as the average of the unsigned percentage error. 17

24 Chapter 4 Results In this section, the results of the ANN and RF models, with respect to both stocks are presented. This involves graphs illustrating the difference between the actual price and the one-day predicted price of the test set as well as tables containing the performance metrics. 4.1 Performance on WMT stock In figure 4.1, the graph presents the differences between the actual and the one-day predicted closing prices on the WMT stock using the respective models on the test set. Figure 4.1: A graph presenting the predictions of the two models on the WMT stock against the actual prices during the 38 day test period ranging between days

25 CHAPTER 4. RESULTS In table 4.1, the performance metrics for the models on the WMT stock are presented. RMSE MAPE ANN RF Table 4.1: Table presenting the RMSE and MAPE metrics for the WMT stock. The figures are represented with 4 decimal points. Both the RMSE and MAPE were better for the ANN model, in comparison to the RF model. This indicates that the ANN had better performance on the WMT stock. 4.2 Performance on IXIC index In figure 4.2, the graph represents the differences between the actual and the one-day predicted closing prices on the IXIC stock index using the respective models on the test set. Figure 4.2: A graph presenting the predictions of the two models on the IXIC stock index against the actual prices during the 38 day test period ranging between days

26 CHAPTER 4. RESULTS In table 4.2, the performance metrics for the models on the IXIC stock index are presented. RMSE MAPE ANN RF Table 4.2: Table presenting the RMSE and MAPE metrics for the IXIC stock index. The figures are represented with 4 decimal points. Both the RMSE and MAPE were better for the ANN model, in comparison to the RF model. This indicates that the ANN had better performance on the IXIC stock index. 20

27 Chapter 5 Discussion and conclusion In this section, the results and the methodology of the performed study are discussed. The reviewed sources that are used in this study are also commented upon in the manner of source criticism. Finally, the proposed problem statement described in section 1.1 is answered in the final conclusion. 5.1 Discussion of results The results for both WMT and IXIC consistently show better performance for the ANN model compared to the RF model. The MAPE metric shows that the errors in the predictive models all are within a couple of percents with 2.88 % being the highest for the RF model prediction on the WMT stock. An observation of the WMT stock prediction seen in figure 4.1 is that there is a drastic price drop from around 67 USD to 57 USD between the days This is incidentally exactly the price range as represented in the test data except that the price development looks completely different there. As there are no other representations of this particular price range in the training data, this could potentially impact the models abilities to accurately generalize these prices in the test period. The ANN model has better generalization performance in this case compared to the RF model. When analyzing the results for the IXIC index in figure 4.2, a similar pattern is seen as for the WMT stock. The sudden divergence from day 235 and onwards is consistent with the hypothesis of bad generalization mentioned above, i.e after day 235 the value drops unusually low when compared to the training data in figure 3.2. Just as for the WMT stock, here the ANN model handles the lack of reliable training data better than the RF model as it consistently shows lower biases for every data point throughout this period. Further observations of the results might indicate that the quality of the validation data has different impacts on the two models for their respective hyperparameter optimization process. For the WMT stock prediction in figure 4.1, the RF model has a rather large bias consistently throughout the test period in comparison to the ANN model. In the IXIC index prediction in figure 4.2 prior to day 235, the difference in bias between the 21

28 CHAPTER 5. DISCUSSION AND CONCLUSION ANN and RF model are not that extreme. This might suggest that the hyperparameters potentially have been better optimized for the IXIC models in comparison to the WMT models. However, as the ANN model still generalize better for both stocks it seems to suggest that it was less sensitive to overfitting, in comparison to the RF model. The chosen partitioning of the data is further discussed in section Overall the ANN model had some noticeable spikes during the test period for both stocks while the RF model was smooth to a larger extent. The spikes generally synced well with the changes of the actual price even if it occasionally, in connection with very sudden changes, resulted in a bigger misprediction than usual. The cause of this may be that, when sudden changes with high variance in the price occur, the ANN model tries too follow the recent change but is too slow in order to estimate these rather extreme day-to-day changes. These kinds of scenarios may be harder to predict than those over a longer time period because of the increased difficulty of generalizing such unique events from a limited set of training data. 5.2 Discussion of method The complexity of ANNs and RFs means that there are several different hyperparameters that can be tuned to regulate their performance. When implementing the ANN and RF models, some of the hyperparameters were chosen by general literature study as well as performing own tests to choose what could be regarded as optimal. In contrast, some other hyperparameters were set rather arbitrarily, generally due to the time constraint. Studying these factors further could have improved the results Amount of data/observations One of the factors that may affect performance is the total amount of data to collect for each stock. The literature study showed different reports using very different amounts of data. Choosing 255 trading days in total (about one year of trading days) was an amount that could have been adjusted by measuring performance using different amounts of datasets. It is possible that in the instances where the predictive models had difficulties to generalize (see section 5.1), using more historical data for training and validating the models, could have aided in increasing the performance of the models. As such, when considering the size of the datasets, it might be viable to also study the behaviour of the stocks. For instance, considering a factor such as the stock s volatility over a certain period of time could aid in deciding whether to use more or less data for that specific stock. Further studies of the stock and it s behaviour could thus aid in selecting an optimal amount of days/observations. However, studying different amounts of datasets and assessing the optimal amount would have required much time and resources, especially considering the need to choose the same amount of data for both the ANN and RF models. It should also be noted that the aim of this study is to compare the performances of ANN and RF models. Using the amount that was chosen in this study can have aided in noticing possible strengths and 22

29 CHAPTER 5. DISCUSSION AND CONCLUSION weaknesses of the respective models Partitioning of the data The partitioning of the data into training, validation and test sets is a factor that may affect performance as well (as briefly discussed in section 5.1). The chosen amount of test days was set rather arbitrarily, and further large-scale studies to determine optimal amount of test days for both models could improve the results. The quality of the validation data is also of importance when optimizing the models. As mentioned in section 3.1, the training data was slightly arbitrarily chosen to include sudden spikes and dips in the closing price, and the validation data was the remaining set of days that were prior to the test days. In the WMT stock in figure 3.1, it is noticed that the stock price declines until around day 199 when it is starting to recover. The partitioning presented in section 3.1 shows that the training data (days 1-204) capture prices of all possible levels while the validation data (days ) reflect a more local behaviour in this case as these prices do not quite include the volatility compared to all 254 days and the test data in particular. An implication of this may be that the validation data do not give an entirely fair reflection of the true nature of the stock and thus the hyperparameters may be further optimizable to some extent. In contrast, the IXIC index in figure 3.2 shows that in the training data (days 1-178), the range of prices is approximately in the interval from 4500 USD to 5200 USD. The validation data (days ) consist of prices from 4800 USD to 5150 USD and matches the typical price development of the training data well, and to a large extent the test data as well. As discussed in section 5.1, this might suggest that the hyperparameter optimization process for the IXIC models was more effective than for the WMT models. Distributing the training and validation set is a complicated task. Increasing the validation data and decreasing the training data might create a superior model in some cases, whilst in some other cases it might be the direct opposite. The complexity increases further due to the need for both the ANN and RF models to use the same partitioning scheme. Overall, as both the ANN and RF models used the same partitioning scheme for the same stock, it would be viable to state that the chosen methodology was fair for the comparative analysis. The fact that ANNs are able to perform better for both stocks, regardless of the possible shortcomings discussed above, is an interesting result to notice. It is something that can be used to strengthen the case for the use of ANN models within the field of stock prediction The choice of technical indicators The decision to use technical indicators to process the input vector was made after reading similar studies using the same approach. Thus, there was no greater knowledge on the subject of technical analysis apart from the general ideas and the technical indicators chosen were based on what had been read in similar reports. Further research on the 23

30 CHAPTER 5. DISCUSSION AND CONCLUSION subject of technical analysis and existing indicators and possibly using more indicators could have been something that could have improved the results and the methodology. 5.3 Comments on the reviewed sources The primary sources used in this study includes Bishop s [13] textbook on ANNs for pattern recognition, Breiman s [22] paper on RFs, as well as Criminisi, Shotton and Konukohglu s [19] review of RFs and its use cases. These were part of the main sources used to gain knowledge on ANNs and RFs and were thus key sources to be able to understand, implement and optimize the models. The legitimacy of these sources can be regarded as strong as all authors have extensive backgrounds within the subject. Additional sources that were used to gain knowledge within these fields can also be regarded as strong as they generally were printed within respectable institutes and/or were quite extensive in their material. Other key sources include various studies made on the subject of implementing ANN and RF based models for stock prediction. The reliability of these sources could be questioned as most reports were summarizing in how the models were implemented rather than having an exhaustive description. However, as these studies were made within educational institutes, it would be justifiable to regard them as viable sources. 5.4 Final conclusion The purpose of this study was to compare the predictive performances of ANNs and RFs for one-day predictions on the stock market. ANN and RF based models were implemented and their predictability was tested on two different stocks for an interval of 38 consecutive days. The results of the study indicate the following: Seen over the whole test period, the ANN model is less erroneous than the RF model for both stocks. As indicated in tables 4.1 and 4.2, the ANN model showed better RMSE and MAPE. The nature of the data has a definite impact on the performance of the models. The ANN model was able to generalize better than the RF model on unreliable training data even if both models deteriorated to some extent at these occasions. There are few isolated occasions in the test data where the RF model outperforms the ANN model. According to these points, the ANN model showed better performance over the test period as a whole, showed to be less sensitive to unreliable training data and proved to be the more consistent of the two. This makes the ANN model a better choice for one day stock prediction compared to the RF model under the circumstances. 24

Abstract Making good predictions for stock prices is an important task for the financial industry. The way these predictions are carried out is often

Abstract Making good predictions for stock prices is an important task for the financial industry. The way these predictions are carried out is often Abstract Making good predictions for stock prices is an important task for the financial industry. The way these predictions are carried out is often by using artificial intelligence that can learn from

More information

The Use of Artificial Neural Network for Forecasting of FTSE Bursa Malaysia KLCI Stock Price Index

The Use of Artificial Neural Network for Forecasting of FTSE Bursa Malaysia KLCI Stock Price Index The Use of Artificial Neural Network for Forecasting of FTSE Bursa Malaysia KLCI Stock Price Index Soleh Ardiansyah 1, Mazlina Abdul Majid 2, JasniMohamad Zain 2 Faculty of Computer System and Software

More information

Artificially Intelligent Forecasting of Stock Market Indexes

Artificially Intelligent Forecasting of Stock Market Indexes Artificially Intelligent Forecasting of Stock Market Indexes Loyola Marymount University Math 560 Final Paper 05-01 - 2018 Daniel McGrath Advisor: Dr. Benjamin Fitzpatrick Contents I. Introduction II.

More information

Two kinds of neural networks, a feed forward multi layer Perceptron (MLP)[1,3] and an Elman recurrent network[5], are used to predict a company's

Two kinds of neural networks, a feed forward multi layer Perceptron (MLP)[1,3] and an Elman recurrent network[5], are used to predict a company's LITERATURE REVIEW 2. LITERATURE REVIEW Detecting trends of stock data is a decision support process. Although the Random Walk Theory claims that price changes are serially independent, traders and certain

More information

Performance analysis of Neural Network Algorithms on Stock Market Forecasting

Performance analysis of Neural Network Algorithms on Stock Market Forecasting www.ijecs.in International Journal Of Engineering And Computer Science ISSN:2319-7242 Volume 3 Issue 9 September, 2014 Page No. 8347-8351 Performance analysis of Neural Network Algorithms on Stock Market

More information

International Journal of Computer Engineering and Applications, Volume XII, Issue II, Feb. 18, ISSN

International Journal of Computer Engineering and Applications, Volume XII, Issue II, Feb. 18,   ISSN Volume XII, Issue II, Feb. 18, www.ijcea.com ISSN 31-3469 AN INVESTIGATION OF FINANCIAL TIME SERIES PREDICTION USING BACK PROPAGATION NEURAL NETWORKS K. Jayanthi, Dr. K. Suresh 1 Department of Computer

More information

COMPARING NEURAL NETWORK AND REGRESSION MODELS IN ASSET PRICING MODEL WITH HETEROGENEOUS BELIEFS

COMPARING NEURAL NETWORK AND REGRESSION MODELS IN ASSET PRICING MODEL WITH HETEROGENEOUS BELIEFS Akademie ved Leske republiky Ustav teorie informace a automatizace Academy of Sciences of the Czech Republic Institute of Information Theory and Automation RESEARCH REPORT JIRI KRTEK COMPARING NEURAL NETWORK

More information

An enhanced artificial neural network for stock price predications

An enhanced artificial neural network for stock price predications An enhanced artificial neural network for stock price predications Jiaxin MA Silin HUANG School of Engineering, The Hong Kong University of Science and Technology, Hong Kong SAR S. H. KWOK HKUST Business

More information

Predicting Economic Recession using Data Mining Techniques

Predicting Economic Recession using Data Mining Techniques Predicting Economic Recession using Data Mining Techniques Authors Naveed Ahmed Kartheek Atluri Tapan Patwardhan Meghana Viswanath Predicting Economic Recession using Data Mining Techniques Page 1 Abstract

More information

STOCK PRICE PREDICTION: KOHONEN VERSUS BACKPROPAGATION

STOCK PRICE PREDICTION: KOHONEN VERSUS BACKPROPAGATION STOCK PRICE PREDICTION: KOHONEN VERSUS BACKPROPAGATION Alexey Zorin Technical University of Riga Decision Support Systems Group 1 Kalkyu Street, Riga LV-1658, phone: 371-7089530, LATVIA E-mail: alex@rulv

More information

Based on BP Neural Network Stock Prediction

Based on BP Neural Network Stock Prediction Based on BP Neural Network Stock Prediction Xiangwei Liu Foundation Department, PLA University of Foreign Languages Luoyang 471003, China Tel:86-158-2490-9625 E-mail: liuxwletter@163.com Xin Ma Foundation

More information

Pattern Recognition by Neural Network Ensemble

Pattern Recognition by Neural Network Ensemble IT691 2009 1 Pattern Recognition by Neural Network Ensemble Joseph Cestra, Babu Johnson, Nikolaos Kartalis, Rasul Mehrab, Robb Zucker Pace University Abstract This is an investigation of artificial neural

More information

International Journal of Computer Engineering and Applications, Volume XII, Issue II, Feb. 18, ISSN

International Journal of Computer Engineering and Applications, Volume XII, Issue II, Feb. 18,   ISSN International Journal of Computer Engineering and Applications, Volume XII, Issue II, Feb. 18, www.ijcea.com ISSN 31-3469 AN INVESTIGATION OF FINANCIAL TIME SERIES PREDICTION USING BACK PROPAGATION NEURAL

More information

Stock Trading Following Stock Price Index Movement Classification Using Machine Learning Techniques

Stock Trading Following Stock Price Index Movement Classification Using Machine Learning Techniques Stock Trading Following Stock Price Index Movement Classification Using Machine Learning Techniques 6.1 Introduction Trading in stock market is one of the most popular channels of financial investments.

More information

Backpropagation and Recurrent Neural Networks in Financial Analysis of Multiple Stock Market Returns

Backpropagation and Recurrent Neural Networks in Financial Analysis of Multiple Stock Market Returns Backpropagation and Recurrent Neural Networks in Financial Analysis of Multiple Stock Market Returns Jovina Roman and Akhtar Jameel Department of Computer Science Xavier University of Louisiana 7325 Palmetto

More information

Iran s Stock Market Prediction By Neural Networks and GA

Iran s Stock Market Prediction By Neural Networks and GA Iran s Stock Market Prediction By Neural Networks and GA Mahmood Khatibi MS. in Control Engineering mahmood.khatibi@gmail.com Habib Rajabi Mashhadi Associate Professor h_mashhadi@ferdowsi.um.ac.ir Electrical

More information

ECS171: Machine Learning

ECS171: Machine Learning ECS171: Machine Learning Lecture 15: Tree-based Algorithms Cho-Jui Hsieh UC Davis March 7, 2018 Outline Decision Tree Random Forest Gradient Boosted Decision Tree (GBDT) Decision Tree Each node checks

More information

Business Strategies in Credit Rating and the Control of Misclassification Costs in Neural Network Predictions

Business Strategies in Credit Rating and the Control of Misclassification Costs in Neural Network Predictions Association for Information Systems AIS Electronic Library (AISeL) AMCIS 2001 Proceedings Americas Conference on Information Systems (AMCIS) December 2001 Business Strategies in Credit Rating and the Control

More information

AN ARTIFICIAL NEURAL NETWORK MODELING APPROACH TO PREDICT CRUDE OIL FUTURE. By Dr. PRASANT SARANGI Director (Research) ICSI-CCGRT, Navi Mumbai

AN ARTIFICIAL NEURAL NETWORK MODELING APPROACH TO PREDICT CRUDE OIL FUTURE. By Dr. PRASANT SARANGI Director (Research) ICSI-CCGRT, Navi Mumbai AN ARTIFICIAL NEURAL NETWORK MODELING APPROACH TO PREDICT CRUDE OIL FUTURE By Dr. PRASANT SARANGI Director (Research) ICSI-CCGRT, Navi Mumbai AN ARTIFICIAL NEURAL NETWORK MODELING APPROACH TO PREDICT CRUDE

More information

STOCK MARKET PREDICTION AND ANALYSIS USING MACHINE LEARNING

STOCK MARKET PREDICTION AND ANALYSIS USING MACHINE LEARNING STOCK MARKET PREDICTION AND ANALYSIS USING MACHINE LEARNING Sumedh Kapse 1, Rajan Kelaskar 2, Manojkumar Sahu 3, Rahul Kamble 4 1 Student, PVPPCOE, Computer engineering, PVPPCOE, Maharashtra, India 2 Student,

More information

Predictive Model Learning of Stochastic Simulations. John Hegstrom, FSA, MAAA

Predictive Model Learning of Stochastic Simulations. John Hegstrom, FSA, MAAA Predictive Model Learning of Stochastic Simulations John Hegstrom, FSA, MAAA Table of Contents Executive Summary... 3 Choice of Predictive Modeling Techniques... 4 Neural Network Basics... 4 Financial

More information

An introduction to Machine learning methods and forecasting of time series in financial markets

An introduction to Machine learning methods and forecasting of time series in financial markets An introduction to Machine learning methods and forecasting of time series in financial markets Mark Wong markwong@kth.se December 10, 2016 Abstract The goal of this paper is to give the reader an introduction

More information

Predicting the stock price companies using artificial neural networks (ANN) method (Case Study: National Iranian Copper Industries Company)

Predicting the stock price companies using artificial neural networks (ANN) method (Case Study: National Iranian Copper Industries Company) ORIGINAL ARTICLE Received 2 February. 2016 Accepted 6 March. 2016 Vol. 5, Issue 2, 55-61, 2016 Academic Journal of Accounting and Economic Researches ISSN: 2333-0783 (Online) ISSN: 2375-7493 (Print) ajaer.worldofresearches.com

More information

$tock Forecasting using Machine Learning

$tock Forecasting using Machine Learning $tock Forecasting using Machine Learning Greg Colvin, Garrett Hemann, and Simon Kalouche Abstract We present an implementation of 3 different machine learning algorithms gradient descent, support vector

More information

Barapatre Omprakash et.al; International Journal of Advance Research, Ideas and Innovations in Technology

Barapatre Omprakash et.al; International Journal of Advance Research, Ideas and Innovations in Technology ISSN: 2454-132X Impact factor: 4.295 (Volume 4, Issue 2) Available online at: www.ijariit.com Stock Price Prediction using Artificial Neural Network Omprakash Barapatre omprakashbarapatre@bitraipur.ac.in

More information

Introducing GEMS a Novel Technique for Ensemble Creation

Introducing GEMS a Novel Technique for Ensemble Creation Introducing GEMS a Novel Technique for Ensemble Creation Ulf Johansson 1, Tuve Löfström 1, Rikard König 1, Lars Niklasson 2 1 School of Business and Informatics, University of Borås, Sweden 2 School of

More information

Chapter 6 Forecasting Volatility using Stochastic Volatility Model

Chapter 6 Forecasting Volatility using Stochastic Volatility Model Chapter 6 Forecasting Volatility using Stochastic Volatility Model Chapter 6 Forecasting Volatility using SV Model In this chapter, the empirical performance of GARCH(1,1), GARCH-KF and SV models from

More information

Stock market price index return forecasting using ANN. Gunter Senyurt, Abdulhamit Subasi

Stock market price index return forecasting using ANN. Gunter Senyurt, Abdulhamit Subasi Stock market price index return forecasting using ANN Gunter Senyurt, Abdulhamit Subasi E-mail : gsenyurt@ibu.edu.ba, asubasi@ibu.edu.ba Abstract Even though many new data mining techniques have been introduced

More information

Foreign Exchange Rate Forecasting using Levenberg- Marquardt Learning Algorithm

Foreign Exchange Rate Forecasting using Levenberg- Marquardt Learning Algorithm Indian Journal of Science and Technology, Vol 9(8), DOI: 10.17485/ijst/2016/v9i8/87904, February 2016 ISSN (Print) : 0974-6846 ISSN (Online) : 0974-5645 Foreign Exchange Rate Forecasting using Levenberg-

More information

Statistical and Machine Learning Approach in Forex Prediction Based on Empirical Data

Statistical and Machine Learning Approach in Forex Prediction Based on Empirical Data Statistical and Machine Learning Approach in Forex Prediction Based on Empirical Data Sitti Wetenriajeng Sidehabi Department of Electrical Engineering Politeknik ATI Makassar Makassar, Indonesia tenri616@gmail.com

More information

Investing through Economic Cycles with Ensemble Machine Learning Algorithms

Investing through Economic Cycles with Ensemble Machine Learning Algorithms Investing through Economic Cycles with Ensemble Machine Learning Algorithms Thomas Raffinot Silex Investment Partners Big Data in Finance Conference Thomas Raffinot (Silex-IP) Economic Cycles-Machine Learning

More information

Chapter IV. Forecasting Daily and Weekly Stock Returns

Chapter IV. Forecasting Daily and Weekly Stock Returns Forecasting Daily and Weekly Stock Returns An unsophisticated forecaster uses statistics as a drunken man uses lamp-posts -for support rather than for illumination.0 Introduction In the previous chapter,

More information

LITERATURE REVIEW. can mimic the brain. A neural network consists of an interconnected nnected group of

LITERATURE REVIEW. can mimic the brain. A neural network consists of an interconnected nnected group of 10 CHAPTER 2 LITERATURE REVIEW 2.1 Artificial Neural Network Artificial neural network (ANN), usually ly called led Neural Network (NN), is an algorithm that was originally motivated ted by the goal of

More information

Decision Trees An Early Classifier

Decision Trees An Early Classifier An Early Classifier Jason Corso SUNY at Buffalo January 19, 2012 J. Corso (SUNY at Buffalo) Trees January 19, 2012 1 / 33 Introduction to Non-Metric Methods Introduction to Non-Metric Methods We cover

More information

Session 5. Predictive Modeling in Life Insurance

Session 5. Predictive Modeling in Life Insurance SOA Predictive Analytics Seminar Hong Kong 29 Aug. 2018 Hong Kong Session 5 Predictive Modeling in Life Insurance Jingyi Zhang, Ph.D Predictive Modeling in Life Insurance JINGYI ZHANG PhD Scientist Global

More information

ALGORITHMIC TRADING STRATEGIES IN PYTHON

ALGORITHMIC TRADING STRATEGIES IN PYTHON 7-Course Bundle In ALGORITHMIC TRADING STRATEGIES IN PYTHON Learn to use 15+ trading strategies including Statistical Arbitrage, Machine Learning, Quantitative techniques, Forex valuation methods, Options

More information

STOCK MARKET FORECASTING USING NEURAL NETWORKS

STOCK MARKET FORECASTING USING NEURAL NETWORKS STOCK MARKET FORECASTING USING NEURAL NETWORKS Lakshmi Annabathuni University of Central Arkansas 400S Donaghey Ave, Apt#7 Conway, AR 72034 (845) 636-3443 lakshmiannabathuni@gmail.com Mark E. McMurtrey,

More information

Stock Market Forecasting Using Artificial Neural Networks

Stock Market Forecasting Using Artificial Neural Networks Stock Market Forecasting Using Artificial Neural Networks Burak Gündoğdu Abstract Many papers on forecasting the stock market have been written by the academia. In addition to that, stock market prediction

More information

Stock Market Prediction using Artificial Neural Networks IME611 - Financial Engineering Indian Institute of Technology, Kanpur (208016), India

Stock Market Prediction using Artificial Neural Networks IME611 - Financial Engineering Indian Institute of Technology, Kanpur (208016), India Stock Market Prediction using Artificial Neural Networks IME611 - Financial Engineering Indian Institute of Technology, Kanpur (208016), India Name Pallav Ranka (13457) Abstract Investors in stock market

More information

Design and implementation of artificial neural network system for stock market prediction (A case study of first bank of Nigeria PLC Shares)

Design and implementation of artificial neural network system for stock market prediction (A case study of first bank of Nigeria PLC Shares) International Journal of Advanced Engineering and Technology ISSN: 2456-7655 www.newengineeringjournal.com Volume 1; Issue 1; March 2017; Page No. 46-51 Design and implementation of artificial neural network

More information

Academic Research Review. Algorithmic Trading using Neural Networks

Academic Research Review. Algorithmic Trading using Neural Networks Academic Research Review Algorithmic Trading using Neural Networks EXECUTIVE SUMMARY In this paper, we attempt to use a neural network to predict opening prices of a set of equities which is then fed into

More information

Improving Stock Price Prediction with SVM by Simple Transformation: The Sample of Stock Exchange of Thailand (SET)

Improving Stock Price Prediction with SVM by Simple Transformation: The Sample of Stock Exchange of Thailand (SET) Thai Journal of Mathematics Volume 14 (2016) Number 3 : 553 563 http://thaijmath.in.cmu.ac.th ISSN 1686-0209 Improving Stock Price Prediction with SVM by Simple Transformation: The Sample of Stock Exchange

More information

Evaluate the Ability of Artificial Neural Network to Predict the Stock Price of Non-Metallic Mineral Products Industry in Tehran's Stock Exchange

Evaluate the Ability of Artificial Neural Network to Predict the Stock Price of Non-Metallic Mineral Products Industry in Tehran's Stock Exchange Evaluate the Ability of Artificial Neural Network to Predict the Stock Price of Non-Metallic Mineral Products Industry in Tehran's Stock Exchange Mohammad Sarchami, Department of Accounting, College Of

More information

Forecasting stock market prices

Forecasting stock market prices ICT Innovations 2010 Web Proceedings ISSN 1857-7288 107 Forecasting stock market prices Miroslav Janeski, Slobodan Kalajdziski Faculty of Electrical Engineering and Information Technologies, Skopje, Macedonia

More information

Predicting stock prices for large-cap technology companies

Predicting stock prices for large-cap technology companies Predicting stock prices for large-cap technology companies 15 th December 2017 Ang Li (al171@stanford.edu) Abstract The goal of the project is to predict price changes in the future for a given stock.

More information

Quantile Regression. By Luyang Fu, Ph. D., FCAS, State Auto Insurance Company Cheng-sheng Peter Wu, FCAS, ASA, MAAA, Deloitte Consulting

Quantile Regression. By Luyang Fu, Ph. D., FCAS, State Auto Insurance Company Cheng-sheng Peter Wu, FCAS, ASA, MAAA, Deloitte Consulting Quantile Regression By Luyang Fu, Ph. D., FCAS, State Auto Insurance Company Cheng-sheng Peter Wu, FCAS, ASA, MAAA, Deloitte Consulting Agenda Overview of Predictive Modeling for P&C Applications Quantile

More information

Predicting Foreign Exchange Arbitrage

Predicting Foreign Exchange Arbitrage Predicting Foreign Exchange Arbitrage Stefan Huber & Amy Wang 1 Introduction and Related Work The Covered Interest Parity condition ( CIP ) should dictate prices on the trillion-dollar foreign exchange

More information

Keywords: artificial neural network, backpropagtion algorithm, derived parameter.

Keywords: artificial neural network, backpropagtion algorithm, derived parameter. Volume 5, Issue 2, February 2015 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Stock Price

More information

PREDICTION OF CLOSING PRICES ON THE STOCK EXCHANGE WITH THE USE OF ARTIFICIAL NEURAL NETWORKS

PREDICTION OF CLOSING PRICES ON THE STOCK EXCHANGE WITH THE USE OF ARTIFICIAL NEURAL NETWORKS Image Processing & Communication, vol. 17, no. 4, pp. 275-282 DOI: 10.2478/v10248-012-0056-5 275 PREDICTION OF CLOSING PRICES ON THE STOCK EXCHANGE WITH THE USE OF ARTIFICIAL NEURAL NETWORKS MICHAŁ PALUCH,

More information

Accepted Manuscript AIRMS: A RISK MANAGEMENT TOOL USING MACHINE LEARNING. Spyros K. Chandrinos, Georgios Sakkas, Nikos D. Lagaros

Accepted Manuscript AIRMS: A RISK MANAGEMENT TOOL USING MACHINE LEARNING. Spyros K. Chandrinos, Georgios Sakkas, Nikos D. Lagaros Accepted Manuscript AIRMS: A RISK MANAGEMENT TOOL USING MACHINE LEARNING Spyros K. Chandrinos, Georgios Sakkas, Nikos D. Lagaros PII: DOI: Reference: S0957-4174(18)30190-8 10.1016/j.eswa.2018.03.044 ESWA

More information

An Improved Approach for Business & Market Intelligence using Artificial Neural Network

An Improved Approach for Business & Market Intelligence using Artificial Neural Network Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology ISSN 2320 088X IMPACT FACTOR: 5.258 IJCSMC,

More information

Lecture 9: Classification and Regression Trees

Lecture 9: Classification and Regression Trees Lecture 9: Classification and Regression Trees Advanced Applied Multivariate Analysis STAT 2221, Spring 2015 Sungkyu Jung Department of Statistics, University of Pittsburgh Xingye Qiao Department of Mathematical

More information

SURVEY OF MACHINE LEARNING TECHNIQUES FOR STOCK MARKET ANALYSIS

SURVEY OF MACHINE LEARNING TECHNIQUES FOR STOCK MARKET ANALYSIS International Journal of Computer Engineering and Applications, Volume XI, Special Issue, May 17, www.ijcea.com ISSN 2321-3469 SURVEY OF MACHINE LEARNING TECHNIQUES FOR STOCK MARKET ANALYSIS Sumeet Ghegade

More information

APPLICATION OF ARTIFICIAL NEURAL NETWORK SUPPORTING THE PROCESS OF PORTFOLIO MANAGEMENT IN TERMS OF TIME INVESTMENT ON THE WARSAW STOCK EXCHANGE

APPLICATION OF ARTIFICIAL NEURAL NETWORK SUPPORTING THE PROCESS OF PORTFOLIO MANAGEMENT IN TERMS OF TIME INVESTMENT ON THE WARSAW STOCK EXCHANGE QUANTITATIVE METHODS IN ECONOMICS Vol. XV, No. 2, 2014, pp. 307 316 APPLICATION OF ARTIFICIAL NEURAL NETWORK SUPPORTING THE PROCESS OF PORTFOLIO MANAGEMENT IN TERMS OF TIME INVESTMENT ON THE WARSAW STOCK

More information

Application of selected methods of statistical analysis and machine learning. learning in predictions of EURUSD, DAX and Ether prices

Application of selected methods of statistical analysis and machine learning. learning in predictions of EURUSD, DAX and Ether prices Application of selected methods of statistical analysis and machine learning in predictions of EURUSD, DAX and Ether prices Mateusz M.@mini.pw.edu.pl Faculty of Mathematics and Information Science Warsaw

More information

A Comparative Study of Ensemble-based Forecasting Models for Stock Index Prediction

A Comparative Study of Ensemble-based Forecasting Models for Stock Index Prediction Association for Information Systems AIS Electronic Library (AISeL) MWAIS 206 Proceedings Midwest (MWAIS) Spring 5-9-206 A Comparative Study of Ensemble-based Forecasting Models for Stock Index Prediction

More information

Prediction of securities behavior using a multi-level artificial neural network with extra inputs between layers

Prediction of securities behavior using a multi-level artificial neural network with extra inputs between layers EXAMENSARBETE INOM TEKNIK, GRUNDNIVÅ, 15 HP STOCKHOLM, SVERIGE 2017 Prediction of securities behavior using a multi-level artificial neural network with extra inputs between layers ERIC TÖRNQVIST XING

More information

Applications of Neural Networks in Stock Market Prediction

Applications of Neural Networks in Stock Market Prediction Applications of Neural Networks in Stock Market Prediction -An Approach Based Analysis Shiv Kumar Goel 1, Bindu Poovathingal 2, Neha Kumari 3 1Asst. Professor, Vivekanand Education Society Institute of

More information

Valencia. Keywords: Conditional volatility, backpropagation neural network, GARCH in Mean MSC 2000: 91G10, 91G70

Valencia. Keywords: Conditional volatility, backpropagation neural network, GARCH in Mean MSC 2000: 91G10, 91G70 Int. J. Complex Systems in Science vol. 2(1) (2012), pp. 21 26 Estimating returns and conditional volatility: a comparison between the ARMA-GARCH-M Models and the Backpropagation Neural Network Fernando

More information

Lending Club Loan Portfolio Optimization Fred Robson (frobson), Chris Lucas (cflucas)

Lending Club Loan Portfolio Optimization Fred Robson (frobson), Chris Lucas (cflucas) CS22 Artificial Intelligence Stanford University Autumn 26-27 Lending Club Loan Portfolio Optimization Fred Robson (frobson), Chris Lucas (cflucas) Overview Lending Club is an online peer-to-peer lending

More information

SELECTION BIAS REDUCTION IN CREDIT SCORING MODELS

SELECTION BIAS REDUCTION IN CREDIT SCORING MODELS SELECTION BIAS REDUCTION IN CREDIT SCORING MODELS Josef Ditrich Abstract Credit risk refers to the potential of the borrower to not be able to pay back to investors the amount of money that was loaned.

More information

A Novel Iron Loss Reduction Technique for Distribution Transformers Based on a Combined Genetic Algorithm Neural Network Approach

A Novel Iron Loss Reduction Technique for Distribution Transformers Based on a Combined Genetic Algorithm Neural Network Approach 16 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS PART C: APPLICATIONS AND REVIEWS, VOL. 31, NO. 1, FEBRUARY 2001 A Novel Iron Loss Reduction Technique for Distribution Transformers Based on a Combined

More information

Classifying Mutual Funds based on Relative Performance using Artificial Neural Networks

Classifying Mutual Funds based on Relative Performance using Artificial Neural Networks DEGREE PROJECT IN TECHNOLOGY, FIRST CYCLE, 15 CREDITS STOCKHOLM, SWEDEN 2017 Classifying Mutual Funds based on Relative Performance using Artificial Neural Networks ADAM ORRE AXEL PERS KTH ROYAL INSTITUTE

More information

Alternate Models for Forecasting Hedge Fund Returns

Alternate Models for Forecasting Hedge Fund Returns University of Rhode Island DigitalCommons@URI Senior Honors Projects Honors Program at the University of Rhode Island 2011 Alternate Models for Forecasting Hedge Fund Returns Michael A. Holden Michael

More information

Evolution of Strategies with Different Representation Schemes. in a Spatial Iterated Prisoner s Dilemma Game

Evolution of Strategies with Different Representation Schemes. in a Spatial Iterated Prisoner s Dilemma Game Submitted to IEEE Transactions on Computational Intelligence and AI in Games (Final) Evolution of Strategies with Different Representation Schemes in a Spatial Iterated Prisoner s Dilemma Game Hisao Ishibuchi,

More information

How Can YOU Use it? Artificial Intelligence for Actuaries. SOA Annual Meeting, Gaurav Gupta. Session 058PD

How Can YOU Use it? Artificial Intelligence for Actuaries. SOA Annual Meeting, Gaurav Gupta. Session 058PD Artificial Intelligence for Actuaries How Can YOU Use it? SOA Annual Meeting, 2018 Session 058PD Gaurav Gupta Founder & CEO ggupta@quaerainsights.com Audience Poll What is my level of AI understanding?

More information

Accepted Manuscript. Enterprise Credit Risk Evaluation Based on Neural Network Algorithm. Xiaobing Huang, Xiaolian Liu, Yuanqian Ren

Accepted Manuscript. Enterprise Credit Risk Evaluation Based on Neural Network Algorithm. Xiaobing Huang, Xiaolian Liu, Yuanqian Ren Accepted Manuscript Enterprise Credit Risk Evaluation Based on Neural Network Algorithm Xiaobing Huang, Xiaolian Liu, Yuanqian Ren PII: S1389-0417(18)30213-4 DOI: https://doi.org/10.1016/j.cogsys.2018.07.023

More information

Essays on Some Combinatorial Optimization Problems with Interval Data

Essays on Some Combinatorial Optimization Problems with Interval Data Essays on Some Combinatorial Optimization Problems with Interval Data a thesis submitted to the department of industrial engineering and the institute of engineering and sciences of bilkent university

More information

Can Twitter predict the stock market?

Can Twitter predict the stock market? 1 Introduction Can Twitter predict the stock market? Volodymyr Kuleshov December 16, 2011 Last year, in a famous paper, Bollen et al. (2010) made the claim that Twitter mood is correlated with the Dow

More information

Providing a Model to Predict Future Cash Flow Using Neural Networks on the Pharmaceutical and Chemical Industries of Tehran Stock Market

Providing a Model to Predict Future Cash Flow Using Neural Networks on the Pharmaceutical and Chemical Industries of Tehran Stock Market Providing a Model to Predict Future Cash Flow Using Neural Networks on the Pharmaceutical and Chemical Industries of Tehran Stock Market Mohammad Khakrah Kahnamouei (Corresponding author) Dept. of Accounting,

More information

COGNITIVE LEARNING OF INTELLIGENCE SYSTEMS USING NEURAL NETWORKS: EVIDENCE FROM THE AUSTRALIAN CAPITAL MARKETS

COGNITIVE LEARNING OF INTELLIGENCE SYSTEMS USING NEURAL NETWORKS: EVIDENCE FROM THE AUSTRALIAN CAPITAL MARKETS Asian Academy of Management Journal, Vol. 7, No. 2, 17 25, July 2002 COGNITIVE LEARNING OF INTELLIGENCE SYSTEMS USING NEURAL NETWORKS: EVIDENCE FROM THE AUSTRALIAN CAPITAL MARKETS Joachim Tan Edward Sek

More information

A new look at tree based approaches

A new look at tree based approaches A new look at tree based approaches Xifeng Wang University of North Carolina Chapel Hill xifeng@live.unc.edu April 18, 2018 Xifeng Wang (UNC-Chapel Hill) Short title April 18, 2018 1 / 27 Outline of this

More information

Academic Research Review. Classifying Market Conditions Using Hidden Markov Model

Academic Research Review. Classifying Market Conditions Using Hidden Markov Model Academic Research Review Classifying Market Conditions Using Hidden Markov Model INTRODUCTION Best known for their applications in speech recognition, Hidden Markov Models (HMMs) are able to discern and

More information

Dr. P. O. Asagba Computer Science Department, Faculty of Science, University of Port Harcourt, Port Harcourt, PMB 5323, Choba, Nigeria

Dr. P. O. Asagba Computer Science Department, Faculty of Science, University of Port Harcourt, Port Harcourt, PMB 5323, Choba, Nigeria PREDICTING THE NIGERIAN STOCK MARKET USING ARTIFICIAL NEURAL NETWORK S. Neenwi Computer Science Department, Rivers State Polytechnic, Bori, PMB 20, Rivers State, Nigeria. Dr. P. O. Asagba Computer Science

More information

Fast Convergence of Regress-later Series Estimators

Fast Convergence of Regress-later Series Estimators Fast Convergence of Regress-later Series Estimators New Thinking in Finance, London Eric Beutner, Antoon Pelsser, Janina Schweizer Maastricht University & Kleynen Consultants 12 February 2014 Beutner Pelsser

More information

PRE CONFERENCE WORKSHOP 3

PRE CONFERENCE WORKSHOP 3 PRE CONFERENCE WORKSHOP 3 Stress testing operational risk for capital planning and capital adequacy PART 2: Monday, March 18th, 2013, New York Presenter: Alexander Cavallo, NORTHERN TRUST 1 Disclaimer

More information

Development and Performance Evaluation of Three Novel Prediction Models for Mutual Fund NAV Prediction

Development and Performance Evaluation of Three Novel Prediction Models for Mutual Fund NAV Prediction Development and Performance Evaluation of Three Novel Prediction Models for Mutual Fund NAV Prediction Ananya Narula *, Chandra Bhanu Jha * and Ganapati Panda ** E-mail: an14@iitbbs.ac.in; cbj10@iitbbs.ac.in;

More information

Application of Deep Learning to Algorithmic Trading

Application of Deep Learning to Algorithmic Trading Application of Deep Learning to Algorithmic Trading Guanting Chen [guanting] 1, Yatong Chen [yatong] 2, and Takahiro Fushimi [tfushimi] 3 1 Institute of Computational and Mathematical Engineering, Stanford

More information

Stock Market Index Prediction Using Multilayer Perceptron and Long Short Term Memory Networks: A Case Study on BSE Sensex

Stock Market Index Prediction Using Multilayer Perceptron and Long Short Term Memory Networks: A Case Study on BSE Sensex Stock Market Index Prediction Using Multilayer Perceptron and Long Short Term Memory Networks: A Case Study on BSE Sensex R. Arjun Raj # # Research Scholar, APJ Abdul Kalam Technological University, College

More information

Enforcing monotonicity of decision models: algorithm and performance

Enforcing monotonicity of decision models: algorithm and performance Enforcing monotonicity of decision models: algorithm and performance Marina Velikova 1 and Hennie Daniels 1,2 A case study of hedonic price model 1 Tilburg University, CentER for Economic Research,Tilburg,

More information

Using neural network to approximate macroeconomic forecast models for BRICS nations

Using neural network to approximate macroeconomic forecast models for BRICS nations Using neural network to approximate macroeconomic forecast models for BRICS nations Three strategies are used to train neural network economic forecast models on temporal data: a standard feedforward neural

More information

Leverage Financial News to Predict Stock Price Movements Using Word Embeddings and Deep Neural Networks

Leverage Financial News to Predict Stock Price Movements Using Word Embeddings and Deep Neural Networks Leverage Financial News to Predict Stock Price Movements Using Word Embeddings and Deep Neural Networks Yangtuo Peng A THESIS SUBMITTED TO THE FACULTY OF GRADUATE STUDIES IN PARTIAL FULFILLMENT OF THE

More information

Bond Market Prediction using an Ensemble of Neural Networks

Bond Market Prediction using an Ensemble of Neural Networks Bond Market Prediction using an Ensemble of Neural Networks Bhagya Parekh Naineel Shah Rushabh Mehta Harshil Shah ABSTRACT The characteristics of a successful financial forecasting system are the exploitation

More information

International Journal of Research in Engineering Technology - Volume 2 Issue 5, July - August 2017

International Journal of Research in Engineering Technology - Volume 2 Issue 5, July - August 2017 RESEARCH ARTICLE OPEN ACCESS The technical indicator Z-core as a forecasting input for neural networks in the Dutch stock market Gerardo Alfonso Department of automation and systems engineering, University

More information

Prediction of Stock Price Movements Using Options Data

Prediction of Stock Price Movements Using Options Data Prediction of Stock Price Movements Using Options Data Charmaine Chia cchia@stanford.edu Abstract This study investigates the relationship between time series data of a daily stock returns and features

More information

Forecasting Foreign Exchange Rate during Crisis - A Neural Network Approach

Forecasting Foreign Exchange Rate during Crisis - A Neural Network Approach International Proceedings of Economics Development and Research IPEDR vol.86 (2016) (2016) IACSIT Press, Singapore Forecasting Foreign Exchange Rate during Crisis - A Neural Network Approach K. V. Bhanu

More information

Foreign Exchange Forecasting via Machine Learning

Foreign Exchange Forecasting via Machine Learning Foreign Exchange Forecasting via Machine Learning Christian González Rojas cgrojas@stanford.edu Molly Herman mrherman@stanford.edu I. INTRODUCTION The finance industry has been revolutionized by the increased

More information

UPDATED IAA EDUCATION SYLLABUS

UPDATED IAA EDUCATION SYLLABUS II. UPDATED IAA EDUCATION SYLLABUS A. Supporting Learning Areas 1. STATISTICS Aim: To enable students to apply core statistical techniques to actuarial applications in insurance, pensions and emerging

More information

MS&E 448 Final Presentation High Frequency Algorithmic Trading

MS&E 448 Final Presentation High Frequency Algorithmic Trading MS&E 448 Final Presentation High Frequency Algorithmic Trading Francis Choi George Preudhomme Nopphon Siranart Roger Song Daniel Wright Stanford University June 6, 2017 High-Frequency Trading MS&E448 June

More information

The Balance-Matching Heuristic *

The Balance-Matching Heuristic * How Do Americans Repay Their Debt? The Balance-Matching Heuristic * John Gathergood Neale Mahoney Neil Stewart Jörg Weber February 6, 2019 Abstract In Gathergood et al. (forthcoming), we studied credit

More information

Modelling the Sharpe ratio for investment strategies

Modelling the Sharpe ratio for investment strategies Modelling the Sharpe ratio for investment strategies Group 6 Sako Arts 0776148 Rik Coenders 0777004 Stefan Luijten 0783116 Ivo van Heck 0775551 Rik Hagelaars 0789883 Stephan van Driel 0858182 Ellen Cardinaels

More information

Neuro-Genetic System for DAX Index Prediction

Neuro-Genetic System for DAX Index Prediction Neuro-Genetic System for DAX Index Prediction Marcin Jaruszewicz and Jacek Mańdziuk Faculty of Mathematics and Information Science, Warsaw University of Technology, Plac Politechniki 1, 00-661 Warsaw,

More information

Machine Learning in Risk Forecasting and its Application in Low Volatility Strategies

Machine Learning in Risk Forecasting and its Application in Low Volatility Strategies NEW THINKING Machine Learning in Risk Forecasting and its Application in Strategies By Yuriy Bodjov Artificial intelligence and machine learning are two terms that have gained increased popularity within

More information

Journal of Internet Banking and Commerce

Journal of Internet Banking and Commerce Journal of Internet Banking and Commerce An open access Internet journal (http://www.icommercecentral.com) Journal of Internet Banking and Commerce, December 2017, vol. 22, no. 3 STOCK PRICE PREDICTION

More information

Deep Learning for Time Series Analysis

Deep Learning for Time Series Analysis CS898 Deep Learning and Application Deep Learning for Time Series Analysis Bo Wang Scientific Computation Lab 1 Department of Computer Science University of Waterloo Outline 1. Background Knowledge 2.

More information

Credit Card Default Predictive Modeling

Credit Card Default Predictive Modeling Credit Card Default Predictive Modeling Background: Predicting credit card payment default is critical for the successful business model of a credit card company. An accurate predictive model can help

More information

A Dynamic Hedging Strategy for Option Transaction Using Artificial Neural Networks

A Dynamic Hedging Strategy for Option Transaction Using Artificial Neural Networks A Dynamic Hedging Strategy for Option Transaction Using Artificial Neural Networks Hyun Joon Shin and Jaepil Ryu Dept. of Management Eng. Sangmyung University {hjshin, jpru}@smu.ac.kr Abstract In order

More information

Prior knowledge in economic applications of data mining

Prior knowledge in economic applications of data mining Prior knowledge in economic applications of data mining A.J. Feelders Tilburg University Faculty of Economics Department of Information Management PO Box 90153 5000 LE Tilburg, The Netherlands A.J.Feelders@kub.nl

More information

A Comparative Study of Various Forecasting Techniques in Predicting. BSE S&P Sensex

A Comparative Study of Various Forecasting Techniques in Predicting. BSE S&P Sensex NavaJyoti, International Journal of Multi-Disciplinary Research Volume 1, Issue 1, August 2016 A Comparative Study of Various Forecasting Techniques in Predicting BSE S&P Sensex Dr. Jahnavi M 1 Assistant

More information

The use of artificial neural network in predicting bankruptcy and its comparison with genetic algorithm in firms accepted in Tehran Stock Exchange

The use of artificial neural network in predicting bankruptcy and its comparison with genetic algorithm in firms accepted in Tehran Stock Exchange Journal of Novel Applied Sciences Available online at www.jnasci.org 2014 JNAS Journal-2014-3-2/151-160 ISSN 2322-5149 2014 JNAS The use of artificial neural network in predicting bankruptcy and its comparison

More information