SlideShare ist ein Scribd-Unternehmen logo
1 von 58
Portfolio Selection with
Artificial Neural Networks
Andrew J Ashwood
Supervisor: Dr Anup Basu
1
Outline
• Introduction
• Research Motivation
• Neural Network Primer
• Literature Review
• Research Questions
• Network specification
• Results
• Conclusions
• Further work and limitations
2
Research Motivation
• Trading in equities is big business in
Australia
– Value of All Ordinaries c. $1.2 trillion
– Stocks turned over in 2012 c. 244 billion
– Funds invested in superannuation c. $1.4
trillion
3
Research Motivation
• Outperforming the broader market is
difficult
– From 1969 to 2010, an index investor would
achieve returns 80 per cent greater than the
average investor in a managed fund.’
(Malkiel, 2011)
– ‘…if many managers have sufficient skill to
cover costs, they are hidden by the mass of
managers with insufficient skill. …true α in net
returns to investors is negative for most if not
all active funds’ (Fama & French, 2009)
4
Research Motivation
• The Australian experience is similar
– 70% of active retail funds underperformed the
benchmark over both the 1 and 3 year period
(Karaban & Maguire, 2012)
– Over a five year period 69% of active retail
funds failed to outperform the ASX200AI
(Karaban & Maguire, 2012)
5
Research Motivation
• ‘The attempt to predict accurately the future
course of stock prices and thus the
appropriate time to buy or sell a stock must
rank as one of investors’ most persistent
endeavours.’ (Malkiel, 2011)
• Advances in computing power in combination
with the widespread availability of historical
datasets has provided investors with
increased opportunity to test markets for
predictable returns
6
Research Motivation
• Artificial Neural Networks (‘ANNs’) have
several traits that suggest their suitability
for stock price prediction
– Ability to generalise and learn
– Provide an analytic alternative to other
techniques that rely on assumptions of
normality, linearity, and variable independence
7
Neural Network Primer
• An artificial neural network is a mathematical
model that mimics the structure and function
of biological brains
• A structure of highly interconnected
processing neurons
8
Neural Network Primer
• Typical neural processing element
X1
X2
Xn
Neuron
Inputs
X1 to Xn
W1
W2
Wn
Input
Weights
W1 to Wn
WX i
n
i
i
v
0
Neuron
Activation
Transform-
ation
)tanh(vy
Output
Output
Learning
Error Back
propagation
9
Neural Network Primer
• Neural networks comprise interconnected
structure of neurons
Input
Input
Input
Output
Output
Output
Input
Layer
Hidden
Layer
Output
Layer
10
Neural Network Primer
• Network typology
– 2 fundamental classes of network architecture
• Feed-forward network – information travels in the
forward direction only
– Single layer feed-forward network
– Multi layer feed-forward network
• Recurrent network – contains at least one
feedback loop
– Many types (Fully recurrent, Hopfield, Elman, Jordan, etc
etc)
– Several studies demonstrate that feed-
forward networks outperform recurrent
networks at predicting non-linear time series
data
11
Neural Network Primer
• Network learning – The Back Propagation
Algorithm
– Input weights are randomly applied to the network
– Network calculations occur and the network
produces an output signal
– Simulated output compared to target output to
provide system error
– Error is back-propagated through the network
– Process repeats until error reaches desired level
or maximum number of iterations is reached
12
Literature Review
• Neural networks are relatively new.
– No published research in the financial domain
prior to 1990. By 1993, over 10 studies per
year were being published. (Wong &
Yakup, 1998)
13
Literature Review
• Yoon & Swales (1991) used ANNs to
predict prices of 98 companies and
compare the ANN predictions to multiple
discriminant analysis
– The study concluded the ANN technique
significantly outperformed multiple
discriminant analysis
14
Literature Review
• Kryzanowski, Galler & Wright (1992) used
ANNs to pick stocks
– Fourteen fundamental ratios were used as
inputs – coded as trending
upwards/downwards or stable
– The ANN model correctly predicted price
directional movement over the next year in
66% of cases
– The ANN did not predict extreme movements
well
15
Literature Review
• Hall, 1994 used ANN to create style based
portfolio
– ANNs used to identify stocks with largest
recent price increased and apply this style to
the current market to select a portfolio
– Utilised 1,000 largest US stocks
– 37 inputs used to predict future stock prices
– The portfolio has exceeded the S&P500 since
inception after transaction costs, but, ‘Actual
performance results and specific details of the
system are proprietary’.
16
Literature Review
• Jai & Lang (1994) used ANNs to trade the
TSE
– Used 16 technical indicators to predict the
trend of the price movement over the
subsequent 6 trading days
Year 1990 1991
Type of ANN Fixed
structure
Dual
adaptive
structure
Fixed
structure
Dual
adaptive
structure
Total return % -30.66% 21.61% 25.76% 29.20%
Annual TSEWPI
return%
-39.52% 23.42%
Note the large
divergence in
performance
17
Literature Review
• Freitas et al. (2001) used neural network
to provide an estimate of future returns of
66 representative stocks listed on the
Brazilian BOVESPA stock exchange.
– Worst network prediction error 33%
– Best network prediction error 7%
– Average error almost 17%
– Neural network portfolios achieved better
performance in 19 out of 21 weeks and
outperformed the benchmark portfolio by
12.39%. 18
Literature Review
• Ellis & Wilson (2005) used ANNs to
classify Australian listed REITs as value
stocks.
– Value stocks were bought and portfolio
returns measured
– ANN portfolio outperformed the benchmark
index 6.97% per month on average
– Sharpe and Sortino ratios significantly greater
than the benchmark indices
19
Literature Review
• Ko & Lin (2008) used ANNs for portfolio
selection on Taiwan Stock Exchange
(‘TSE’).
– Focussed on 21 different equities
– Used price, variance, covariance as input
variables
– ANN used to predict equity allocation ratio
– 5yr data set including training & validation
sets
– ANN model outperformed TSE (no 20
Literature Review
• Vanstone et al. (2010) used neural networks
based on a stock trading filter rule. The filter
rule bought stocks when the following criteria
were met
• PE < 10
• Market Price < Book Value
• ROE < 12
• Dividend Payout Ratio < 25%
– System yielded low number of trading
opportunities
– The return achieved by the system exceeded the
return available using the original filter rule
– It is worth noting however that the higher return
achieved, was at least partially due to increased
risk.
21
Research Questions
• Core Research Question 1
– Can neural networks that utilise a combination
of fundamental and technical inputs predict
(a) future stock prices, (b) stock price
directional movement?
• Secondary Research Question 1
– Which network inputs contribute most to
network accuracy?
22
Research Questions
• Secondary Research Question 2
– What network parameters lead to optimal
predictive capability?
• Secondary Research Question 3
– Do neural networks predict any of the known
performance attribution factors?
• All research questions are to be answered
with reference to stocks listed on the ASX
23
Network Specification
• Kaastra & Boyd (1996) framework for
design of financial and econometric time
seriesStep 1: Variable selection
Step 2: Data collection
Step 3: Data preprocessing
Step 4: Training, testing, and
validation sets
Step 7: Neural network training
number of training iterations
learning rate and momentum
Step 6: Evaluation criteria
Step 5: Neural network paradigms
number of hidden layers
number of hidden neurons
number of output neurons
transfer functions
Step 8: Implementation
Iterative
Process
24
Variable Selection
FUNDAMENTAL INPUTS TECHNICAL INPUTS
Profit Margin DPS Open High Low Close
Dividend
Yield
CF per share 20d MA 50d MA 20d EMA 50d EMA
BV per share BV per share
growth
MACD RSI CLV ADI
Current ratio Quick Ratio 20d Momentum 50d
Momentum
6mth
Momentum
12mth
Momentum
Asset
Turnover
Debt to
Equity
Fast Stochastic
Oscillator
Fast
Stochastic
Indicator
Slow
Stochastic
Oscillator
Slow
Stochastic
Indicator
Interest
Cover
Price to CF Rate of Change
(Price 1wk)
Rate of Change
(Price 4wk)
Rate of Change
(Price 10wk)
Rate of Change
(MACD)
PE Price to Sales Rate of Change
(MACD Signal)
Rate of Change
(RSI)
Rate of Change
(ADI)
Rate of Change
(20d MA)
ROA ROE Rate of Change
(50d MA)
Rate of Change
(20d EMA)
Rate of Change
(50d EMA)
Rate of Change
(12mth Mom)
FCF per share Sales growth Rate of Change
(FSO)
Rate of Change
(FSI)
Rate of Change
(SSO)
Rate of Change
(SSI)
- - Volume - - - 25
Data Set
• Dataset obtained from Bloomberg
• Weekly data from Jan 1997 to Dec 2011
• Early data (1997-1999) used to calculate
technical indicators
• 20 widely held ASX50 stocks selected
AMC ANZ BHP CBA CCA
CSL DJS GPT LEI NAB
ORG QAN QBE RIO SUN
TLS WBC WDC WES WOW
26
Data Preprocessing
• 3 pre-processing algorithms utilised –
– Remove duplicate data
• Duplicate data provides no useful information to the
ANN
• Speeds calculation
– Scale data
• Ensures all inputs treated equally
• Avoids saturation of the tansig transformation function
– Mark missing data
• Ensures that missing data is not used in the learning
process
27
Frequency of Data
• Little theory to assist in determining
optimum frequency of data for a given
prediction horizon Deboeck (1994) –
– Long term prediction are unreliable
– Short term predictions are unreliable
– There is some period into the future for which
useful predictions can be made
28
Training, Testing, Validation Sets
• The data set needs to be divided into
training, validation, and testing datasets
(Beale et al., 2011)–
– Training set –used to compute the gradient
and updating weights
– Validation set – the data set used for
monitoring the error during training
– Testing set –out-of-sample test data set
29
Training, Testing, Validation Sets
• Walk forward testing –
t0 tn
Time
Window 1
Window 2
Window 3
Window 4
Window 5
Training Window
Validation Window
Testing Window
Testing Period
30
Network Implementation
Summing junction+ Tapped delay lineBias unitbx
Input WeightTDL IW LW Layer Weight Transfer function
Input Layer
5918 Fundamental Inputs
41 Technical Inputs
59
Tansig
b1
4 → 20
TDL
1
Hidden Layer
30-150
IW1,1
+
LW1,2
TDL
Output
Tansig
Output Layer
b2
LW2,1
1
1
+
31
Evaluation Criteria & Training
• Mean Squared Error (‘MSE’) selected as
error function to be minimised by the network
• Convergence method adopted for network
training. Training occurs until the earlier of –
– 100,000 training iterations occurs
OR
– 50 iterations with no reduction in validation set
error
32
Implementation
Parameter Options No. of
Options
Stocks/portfolios 20 stocks +
Equally and value weighted
portfolios of these stocks
22
Input type Price OR
Technical indicators OR
Fundamental indicators OR
Price + Technical OR
Price + Fundamental OR
Price + Technical +
Fundamental
6
Lookback window 4, 8, 12, 16, 20 periods 5
Hidden layer size 30, 60, 90, 120, 150 neurons 5
Training period
length
3mths, 6mths, 12mths, 24mths 4
Walk forward
testing
10 x 1-year testing periods from
2001 to 2011
10
Total networks = 132,000
• All testing undertaken on QUT
supercomputer ‘Lyra’
• At time of testing Lyra
consisted of -
• 106 compute nodes
• 1572 x 64 bit Intel Xeon
cores (consisting 11,736
core processors)
• 32.8 TeraFlop Theoretical
(double precision), 58.6
TeraFlop (single
precision)
• 10.4 TeraBytes of main
memory
33
Results
• The high number of network models specified (132,000)
meant that manual review and interpretation of individual
networks was not feasible.
• The algorithm selected the network configuration that
provides the lowest mean squared error over the one
year walk forward testing window. The algorithm
therefore provides 220 outputs. This is comprised as
follows:
– 20 stocks + 2 portfolios
– 10 walk forward testing windows for each
stock/portfolio
– Total outputs = 220
34
Input Type
0
5
10
15
20
25
30
35
40
45
50
Frequency
Input Type
Histogram of Input Type
17%
14%
10%
21%
22%
16%
35
Lookback Window
0
5
10
15
20
25
30
35
40
45
50
4 8 12 16 20
Frequency
Lookback Period (no. inputs)
Histogram of Lookback
19%
21%
19%
21%
20%
36
Hidden Layer Size
0
20
40
60
80
100
120
30 60 90 120 150
Frequency
Hidden Layer Size (no. of nodes)
Histogram of Hidden layer
Smaller hidden
layer produced
optimum network
in 47% of cases
47%
16%
9%
18%
10%
37
Training Period
0
10
20
30
40
50
60
70
80
3mths 6mths 12mths 24mths
Frequency
Training Period Length
Histogram of Training period
35%
20%
15%
30%
38
Predictive Capability
0%
5%
10%
15%
20%
25%
30%
2002 2003 2004 2005 2006 2007 2008 2009 2010 2011
1 2 3 4 5 6 7 8 9 10
RMSE%ofSharePrice
Year/
Timestep
RMSE as % of Share Price
AMC
ANZ
BHP
CBA
CCL
CSL
DJS
GPT
LEI
NAB
ORG
QAN
QBE
RIO
SUN
TLS
WBC
WDC
WES
WOW
Generally there is a major
spike in error in 2008 &
2009
Generally predictions are
fairly consistent from 2002-
2007
39
Predictive Capability
2,000
2,500
3,000
3,500
4,000
4,500
5,000
5,500
6,000
6,500
7,000
ASX200 - 2002 to 2011
ASX200 Adjusted Close Price
Sustained
benign trading
conditions
leading up to
GFC
2008 Major
reversal, Bear
Market
2009 Major
reversal, Bull
Market
2010 Return
to benign
trading
conditions
40
Directional Movement
0%
10%
20%
30%
40%
50%
60%
70%
80%
2002 2003 2004 2005 2006 2007 2008 2009 2010 2011
1 2 3 4 5 6 7 8 9 10
PredictionAccuracy
Year/
Timestep
Directional Movement Prediction Accuracy
AMC
ANZ
BHP
CBA
CCL
CSL
DJS
GPT
LEI
NAB
ORG
QAN
QBE
RIO
SUN
TLS
WBC
WDC
WES
WOW
EW
VW
41
Directional Movement
0%
10%
20%
30%
40%
50%
60%
70%
80%
90%
100%
PredictionAccuracy
Stocks & Portfolios
Directional Movement Prediction Accuracy
Over 10-year Testing Horizon
AMC
ANZ
BHP
CBA
CCL
CSL
DJS
GPT
LEI
NAB
ORG
QAN
QBE
RIO
SUN
TLS
WBC
WDC
WES
WOW
EW
VW
Significant at
5% level
42
Portfolio Selection
• The simulated price returns were then
used to form a long and a long-short
portfolio
• Portfolios formed at the beginning of each
month during the 10-year testing period
and held until the end of the month
• Stocks weighted according to the absolute
value of their simulated price return
43
Portfolio Price Returns
Portfolio Price Return
(2002-2011)
ASX200 19.6%
ANN Long Portfolio 205.5%
ANN Long-Short
Portfolio
196.6%
• Raw Results –
• Do not include transaction costs
• No performance measurement
framework
44
Regression
Results
Long Portfolio Long-Short
Portfolio
α 0.016** 0.013***
RMRF 0.712*** -0.011
SMB -0.317 -0.604**
HML -0.416** -0.306
WML -0.011 0.038
R2
0.283*** 0.215***
Performance Measurement
• The portfolio price returns were then
regressed against the Carhart 4-factor
model
Both portfolios
have a
significantly
positive alpha
Both models are
significant
45
Optimal Network Portfolios
• This analysis confirms that ANNs can be
used for portfolio selection to achieve positive
alpha
HOWEVER
• These results are based upon the most
accurate network specification – which is only
know post hoc
• The problem is whether ANNs can achieve
positive alpha without this post hoc
knowledge
46
Median Network Portfolios
• For each walk-forward testing window for
each stock, 600 different network
specifications were trialled
– 6 input options
– 5 lookback windows
– 5 hidden layer sizes
– 4 training period lengths
– 600 network options per stock per testing period
• We have confirmed that the most accurate
network achieves positive alpha – what about
the median network?
47
Regression
Results
Long Portfolio Long-Short
Portfolio
α 0.012** -0.001
RMRF 0.137 0.003
SMB -0.050 -0.011
HML -0.310 -0.064
WML -0.326 -0.115
R2
0.084* 0.012
Median Network Portfolios
Only the Long
portfolio
achieved a
significant alpha
Only the Long
portfolio model is
significant and
the Carhart style-
factors explain a
lower proportion
of variation in
data
48
Distribution of Alphas
• Network parameter combinations were
applied uniformly to all stocks and the long
portfolio based on expected returns was
formed.
• Mimics a real world trading situation
• This process resulted in 600
measurements of alpha.
49
Distribution of Alphas
0
20
40
60
80
100
120 Histogram for Alpha
Median = 0.0061
Only 12% of the
networks tested
have negative alpha
127 of 529
positive alphas
(27%) were
statistically
significant (p <
0.05)
50
Conclusions
• Network parameter testing was undertaken into
several network parameters –
– input type
– hidden layer size
– lookback window
– training period length.
• No useful guidance on heuristics for the input
type, lookback window, or training period length.
• Hidden layer size showed some promise. Almost
half of the most accurate networks models utilised
the smallest hidden layer of 30 nodes.
51
Conclusions
• Price prediction performance was mixed.
• Most networks performed poorly at the
beginning of the global financial crisis
(beginning of 2008) and at the end (end of
2009). The networks fail to predict major
reversals.
• The networks did not predict the portfolio
price returns well. The value weighted
portfolio error ranged from 9% (best year) to
32% (worst year) and only achieved a RMSE
below 10% in 1 of the 10 testing years.
52
Conclusions
• Stock directional movement prediction
performance was assessed.
• The ANN models predicted directional
movement with better than 50% accuracy
for all stocks and portfolios except for
BHP.
• Significant results were achieved for 13 of
the 22 stocks and portfolios tested.
53
Conclusions
• Price predictions used for long and long-
short portfolio selection
• Both portfolios achieved significantly
positive alpha.
• The optimal network specification was
made by comparing the simulated prices
to the target prices, thus requiring post hoc
knowledge.
54
Conclusions
• A distribution of alpha for all 600 different
network specifications was produced.
• Could not reject hypothesis that the
distribution was normal (p < 0.05).
• 88% of the networks produced positive
alpha (529 out of 600). 127 of which were
significant (27%).
• 12% of networks produced negative alpha
(71 out of 600). None of which were
significant.
55
Limitations & Future Research
• Research tested several network
parameters.
• Hidden layer size provided promising
results.
• Further research is required to develop
heuristics.
• Future research should look at individual
stocks and industry sectors.
56
Limitations & Future Research
• The analysis has been undertaken on a
price return basis, rather than an overall
returns basis.
• Study utilised 20 widely held stocks.
Needs testing over wider investment
universe. Survivorship bias.
• Impact of trading costs not included in
analysis
57
Questions and Discussion
58

Weitere ähnliche Inhalte

Andere mochten auch

Face recognition using artificial neural network
Face recognition using artificial neural networkFace recognition using artificial neural network
Face recognition using artificial neural network
Sumeet Kakani
 
Artificial intelligence in power systems
Artificial intelligence in power systems Artificial intelligence in power systems
Artificial intelligence in power systems
Riyas K H
 

Andere mochten auch (19)

Artificial neural networks in food industry
Artificial neural networks in food industryArtificial neural networks in food industry
Artificial neural networks in food industry
 
Artificial Neural Network in a Tic Tac Toe Symfony Console Application - Symf...
Artificial Neural Network in a Tic Tac Toe Symfony Console Application - Symf...Artificial Neural Network in a Tic Tac Toe Symfony Console Application - Symf...
Artificial Neural Network in a Tic Tac Toe Symfony Console Application - Symf...
 
Image Processing Based Signature Recognition and Verification Technique Using...
Image Processing Based Signature Recognition and Verification Technique Using...Image Processing Based Signature Recognition and Verification Technique Using...
Image Processing Based Signature Recognition and Verification Technique Using...
 
Compegence: Dr. Rajaram Kudli - An Introduction to Artificial Neural Network ...
Compegence: Dr. Rajaram Kudli - An Introduction to Artificial Neural Network ...Compegence: Dr. Rajaram Kudli - An Introduction to Artificial Neural Network ...
Compegence: Dr. Rajaram Kudli - An Introduction to Artificial Neural Network ...
 
Detection of heart diseases by data mining
Detection of heart diseases by data miningDetection of heart diseases by data mining
Detection of heart diseases by data mining
 
Face recognization using artificial nerual network
Face recognization using artificial nerual networkFace recognization using artificial nerual network
Face recognization using artificial nerual network
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 
Artificial neural network model & hidden layers in multilayer artificial neur...
Artificial neural network model & hidden layers in multilayer artificial neur...Artificial neural network model & hidden layers in multilayer artificial neur...
Artificial neural network model & hidden layers in multilayer artificial neur...
 
Character Recognition using Artificial Neural Networks
Character Recognition using Artificial Neural NetworksCharacter Recognition using Artificial Neural Networks
Character Recognition using Artificial Neural Networks
 
Handwritten character recognition using artificial neural network
Handwritten character recognition using artificial neural networkHandwritten character recognition using artificial neural network
Handwritten character recognition using artificial neural network
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
 
Convolutional Neural Network for Alzheimer’s disease diagnosis with Neuroim...
Convolutional Neural Network for Alzheimer’s disease diagnosis with Neuroim...Convolutional Neural Network for Alzheimer’s disease diagnosis with Neuroim...
Convolutional Neural Network for Alzheimer’s disease diagnosis with Neuroim...
 
artificial neural network
artificial neural networkartificial neural network
artificial neural network
 
Artificial Intelligence: Artificial Neural Networks
Artificial Intelligence: Artificial Neural NetworksArtificial Intelligence: Artificial Neural Networks
Artificial Intelligence: Artificial Neural Networks
 
Machine learning pt.1: Artificial Neural Networks ® All Rights Reserved
Machine learning pt.1: Artificial Neural Networks ® All Rights ReservedMachine learning pt.1: Artificial Neural Networks ® All Rights Reserved
Machine learning pt.1: Artificial Neural Networks ® All Rights Reserved
 
Face recognition using artificial neural network
Face recognition using artificial neural networkFace recognition using artificial neural network
Face recognition using artificial neural network
 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural network
 
Neural network
Neural networkNeural network
Neural network
 
Artificial intelligence in power systems
Artificial intelligence in power systems Artificial intelligence in power systems
Artificial intelligence in power systems
 

Ähnlich wie Portfolio Selection with Artificial Neural Networks

Estimize Bull speed using Back propagation
Estimize Bull speed using Back propagationEstimize Bull speed using Back propagation
Estimize Bull speed using Back propagation
IJMER
 
Estimize Bull speed using Back propagation
Estimize Bull speed using Back propagationEstimize Bull speed using Back propagation
Estimize Bull speed using Back propagation
IJMER
 
Quantitative Modeling and Trading
Quantitative Modeling and  TradingQuantitative Modeling and  Trading
Quantitative Modeling and Trading
Valapet Badri
 
Random Portfolio using R to create investment strategy for Hedge Fund
Random Portfolio using R to create investment strategy for Hedge FundRandom Portfolio using R to create investment strategy for Hedge Fund
Random Portfolio using R to create investment strategy for Hedge Fund
Krunal Naik, MBA, PMP, CSM
 
Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...
Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...
Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...
confluent
 
Data transformation and query management in personal health sensor network
Data transformation and query management in personal health sensor networkData transformation and query management in personal health sensor network
Data transformation and query management in personal health sensor network
TAIWAN
 

Ähnlich wie Portfolio Selection with Artificial Neural Networks (20)

Seminar.pptx
Seminar.pptxSeminar.pptx
Seminar.pptx
 
Liquidity Fragmentation & SOR
Liquidity Fragmentation & SORLiquidity Fragmentation & SOR
Liquidity Fragmentation & SOR
 
STOCK MARKET PRREDICTION WITH FEATURE EXTRACTION USING NEURAL NETWORK TEHNIQUE
STOCK MARKET PRREDICTION WITH FEATURE EXTRACTION USING NEURAL NETWORK TEHNIQUESTOCK MARKET PRREDICTION WITH FEATURE EXTRACTION USING NEURAL NETWORK TEHNIQUE
STOCK MARKET PRREDICTION WITH FEATURE EXTRACTION USING NEURAL NETWORK TEHNIQUE
 
Estimize Bull speed using Back propagation
Estimize Bull speed using Back propagationEstimize Bull speed using Back propagation
Estimize Bull speed using Back propagation
 
Estimize Bull speed using Back propagation
Estimize Bull speed using Back propagationEstimize Bull speed using Back propagation
Estimize Bull speed using Back propagation
 
Quantitative Modeling and Trading
Quantitative Modeling and  TradingQuantitative Modeling and  Trading
Quantitative Modeling and Trading
 
Random Portfolio using R to create investment strategy for Hedge Fund
Random Portfolio using R to create investment strategy for Hedge FundRandom Portfolio using R to create investment strategy for Hedge Fund
Random Portfolio using R to create investment strategy for Hedge Fund
 
Stock Price Prediction using ML Techniques
Stock Price Prediction using ML TechniquesStock Price Prediction using ML Techniques
Stock Price Prediction using ML Techniques
 
ICIS - Power price prediction with neural networks
ICIS - Power price prediction with neural networksICIS - Power price prediction with neural networks
ICIS - Power price prediction with neural networks
 
Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...
Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...
Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...
 
Data transformation and query management in personal health sensor network
Data transformation and query management in personal health sensor networkData transformation and query management in personal health sensor network
Data transformation and query management in personal health sensor network
 
Stock-market-prediction.pptx
Stock-market-prediction.pptxStock-market-prediction.pptx
Stock-market-prediction.pptx
 
Tsinghua University: Two Exemplary Applications in China
Tsinghua University: Two Exemplary Applications in ChinaTsinghua University: Two Exemplary Applications in China
Tsinghua University: Two Exemplary Applications in China
 
Predicting Stock Market Price Using Support Vector Regression
Predicting Stock Market Price Using Support Vector RegressionPredicting Stock Market Price Using Support Vector Regression
Predicting Stock Market Price Using Support Vector Regression
 
Overview of DuraMat software tool development
Overview of DuraMat software tool developmentOverview of DuraMat software tool development
Overview of DuraMat software tool development
 
Abc financial analyticsdemo
Abc financial analyticsdemoAbc financial analyticsdemo
Abc financial analyticsdemo
 
Stock Market Price Prediction Using Technical Analysis
Stock Market Price Prediction Using Technical AnalysisStock Market Price Prediction Using Technical Analysis
Stock Market Price Prediction Using Technical Analysis
 
NASDOQ 2015
NASDOQ 2015 NASDOQ 2015
NASDOQ 2015
 
Artificial Intelligence for Automated Decision Support Project
Artificial Intelligence for Automated Decision Support ProjectArtificial Intelligence for Automated Decision Support Project
Artificial Intelligence for Automated Decision Support Project
 
An intelligent framework using hybrid social media and market data, for stock...
An intelligent framework using hybrid social media and market data, for stock...An intelligent framework using hybrid social media and market data, for stock...
An intelligent framework using hybrid social media and market data, for stock...
 

Kürzlich hochgeladen

VIP Call Girl in Thane 💧 9920725232 ( Call Me ) Get A New Crush Everyday With...
VIP Call Girl in Thane 💧 9920725232 ( Call Me ) Get A New Crush Everyday With...VIP Call Girl in Thane 💧 9920725232 ( Call Me ) Get A New Crush Everyday With...
VIP Call Girl in Thane 💧 9920725232 ( Call Me ) Get A New Crush Everyday With...
dipikadinghjn ( Why You Choose Us? ) Escorts
 
VIP Independent Call Girls in Taloja 🌹 9920725232 ( Call Me ) Mumbai Escorts ...
VIP Independent Call Girls in Taloja 🌹 9920725232 ( Call Me ) Mumbai Escorts ...VIP Independent Call Girls in Taloja 🌹 9920725232 ( Call Me ) Mumbai Escorts ...
VIP Independent Call Girls in Taloja 🌹 9920725232 ( Call Me ) Mumbai Escorts ...
dipikadinghjn ( Why You Choose Us? ) Escorts
 
Call Girls in New Ashok Nagar, (delhi) call me [9953056974] escort service 24X7
Call Girls in New Ashok Nagar, (delhi) call me [9953056974] escort service 24X7Call Girls in New Ashok Nagar, (delhi) call me [9953056974] escort service 24X7
Call Girls in New Ashok Nagar, (delhi) call me [9953056974] escort service 24X7
9953056974 Low Rate Call Girls In Saket, Delhi NCR
 
VIP Independent Call Girls in Andheri 🌹 9920725232 ( Call Me ) Mumbai Escorts...
VIP Independent Call Girls in Andheri 🌹 9920725232 ( Call Me ) Mumbai Escorts...VIP Independent Call Girls in Andheri 🌹 9920725232 ( Call Me ) Mumbai Escorts...
VIP Independent Call Girls in Andheri 🌹 9920725232 ( Call Me ) Mumbai Escorts...
dipikadinghjn ( Why You Choose Us? ) Escorts
 
VIP Independent Call Girls in Mira Bhayandar 🌹 9920725232 ( Call Me ) Mumbai ...
VIP Independent Call Girls in Mira Bhayandar 🌹 9920725232 ( Call Me ) Mumbai ...VIP Independent Call Girls in Mira Bhayandar 🌹 9920725232 ( Call Me ) Mumbai ...
VIP Independent Call Girls in Mira Bhayandar 🌹 9920725232 ( Call Me ) Mumbai ...
dipikadinghjn ( Why You Choose Us? ) Escorts
 

Kürzlich hochgeladen (20)

Vasai-Virar Fantastic Call Girls-9833754194-Call Girls MUmbai
Vasai-Virar Fantastic Call Girls-9833754194-Call Girls MUmbaiVasai-Virar Fantastic Call Girls-9833754194-Call Girls MUmbai
Vasai-Virar Fantastic Call Girls-9833754194-Call Girls MUmbai
 
VIP Call Girl in Thane 💧 9920725232 ( Call Me ) Get A New Crush Everyday With...
VIP Call Girl in Thane 💧 9920725232 ( Call Me ) Get A New Crush Everyday With...VIP Call Girl in Thane 💧 9920725232 ( Call Me ) Get A New Crush Everyday With...
VIP Call Girl in Thane 💧 9920725232 ( Call Me ) Get A New Crush Everyday With...
 
Navi Mumbai Cooperetive Housewife Call Girls-9833754194-Natural Panvel Enjoye...
Navi Mumbai Cooperetive Housewife Call Girls-9833754194-Natural Panvel Enjoye...Navi Mumbai Cooperetive Housewife Call Girls-9833754194-Natural Panvel Enjoye...
Navi Mumbai Cooperetive Housewife Call Girls-9833754194-Natural Panvel Enjoye...
 
Indore Real Estate Market Trends Report.pdf
Indore Real Estate Market Trends Report.pdfIndore Real Estate Market Trends Report.pdf
Indore Real Estate Market Trends Report.pdf
 
VIP Independent Call Girls in Taloja 🌹 9920725232 ( Call Me ) Mumbai Escorts ...
VIP Independent Call Girls in Taloja 🌹 9920725232 ( Call Me ) Mumbai Escorts ...VIP Independent Call Girls in Taloja 🌹 9920725232 ( Call Me ) Mumbai Escorts ...
VIP Independent Call Girls in Taloja 🌹 9920725232 ( Call Me ) Mumbai Escorts ...
 
(INDIRA) Call Girl Srinagar Call Now 8617697112 Srinagar Escorts 24x7
(INDIRA) Call Girl Srinagar Call Now 8617697112 Srinagar Escorts 24x7(INDIRA) Call Girl Srinagar Call Now 8617697112 Srinagar Escorts 24x7
(INDIRA) Call Girl Srinagar Call Now 8617697112 Srinagar Escorts 24x7
 
Mira Road Awesome 100% Independent Call Girls NUmber-9833754194-Dahisar Inter...
Mira Road Awesome 100% Independent Call Girls NUmber-9833754194-Dahisar Inter...Mira Road Awesome 100% Independent Call Girls NUmber-9833754194-Dahisar Inter...
Mira Road Awesome 100% Independent Call Girls NUmber-9833754194-Dahisar Inter...
 
Vip Call US 📞 7738631006 ✅Call Girls In Sakinaka ( Mumbai )
Vip Call US 📞 7738631006 ✅Call Girls In Sakinaka ( Mumbai )Vip Call US 📞 7738631006 ✅Call Girls In Sakinaka ( Mumbai )
Vip Call US 📞 7738631006 ✅Call Girls In Sakinaka ( Mumbai )
 
WhatsApp 📞 Call : 9892124323 ✅Call Girls In Chembur ( Mumbai ) secure service
WhatsApp 📞 Call : 9892124323  ✅Call Girls In Chembur ( Mumbai ) secure serviceWhatsApp 📞 Call : 9892124323  ✅Call Girls In Chembur ( Mumbai ) secure service
WhatsApp 📞 Call : 9892124323 ✅Call Girls In Chembur ( Mumbai ) secure service
 
Kharghar Blowjob Housewife Call Girls NUmber-9833754194-CBD Belapur Internati...
Kharghar Blowjob Housewife Call Girls NUmber-9833754194-CBD Belapur Internati...Kharghar Blowjob Housewife Call Girls NUmber-9833754194-CBD Belapur Internati...
Kharghar Blowjob Housewife Call Girls NUmber-9833754194-CBD Belapur Internati...
 
Solution Manual for Financial Accounting, 11th Edition by Robert Libby, Patri...
Solution Manual for Financial Accounting, 11th Edition by Robert Libby, Patri...Solution Manual for Financial Accounting, 11th Edition by Robert Libby, Patri...
Solution Manual for Financial Accounting, 11th Edition by Robert Libby, Patri...
 
Top Rated Pune Call Girls Sinhagad Road ⟟ 6297143586 ⟟ Call Me For Genuine S...
Top Rated  Pune Call Girls Sinhagad Road ⟟ 6297143586 ⟟ Call Me For Genuine S...Top Rated  Pune Call Girls Sinhagad Road ⟟ 6297143586 ⟟ Call Me For Genuine S...
Top Rated Pune Call Girls Sinhagad Road ⟟ 6297143586 ⟟ Call Me For Genuine S...
 
Call Girls Koregaon Park Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Koregaon Park Call Me 7737669865 Budget Friendly No Advance BookingCall Girls Koregaon Park Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Koregaon Park Call Me 7737669865 Budget Friendly No Advance Booking
 
Call Girls in New Ashok Nagar, (delhi) call me [9953056974] escort service 24X7
Call Girls in New Ashok Nagar, (delhi) call me [9953056974] escort service 24X7Call Girls in New Ashok Nagar, (delhi) call me [9953056974] escort service 24X7
Call Girls in New Ashok Nagar, (delhi) call me [9953056974] escort service 24X7
 
Webinar on E-Invoicing for Fintech Belgium
Webinar on E-Invoicing for Fintech BelgiumWebinar on E-Invoicing for Fintech Belgium
Webinar on E-Invoicing for Fintech Belgium
 
Stock Market Brief Deck (Under Pressure).pdf
Stock Market Brief Deck (Under Pressure).pdfStock Market Brief Deck (Under Pressure).pdf
Stock Market Brief Deck (Under Pressure).pdf
 
Top Rated Pune Call Girls Lohegaon ⟟ 6297143586 ⟟ Call Me For Genuine Sex Se...
Top Rated  Pune Call Girls Lohegaon ⟟ 6297143586 ⟟ Call Me For Genuine Sex Se...Top Rated  Pune Call Girls Lohegaon ⟟ 6297143586 ⟟ Call Me For Genuine Sex Se...
Top Rated Pune Call Girls Lohegaon ⟟ 6297143586 ⟟ Call Me For Genuine Sex Se...
 
Booking open Available Pune Call Girls Wadgaon Sheri 6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Wadgaon Sheri  6297143586 Call Hot Ind...Booking open Available Pune Call Girls Wadgaon Sheri  6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Wadgaon Sheri 6297143586 Call Hot Ind...
 
VIP Independent Call Girls in Andheri 🌹 9920725232 ( Call Me ) Mumbai Escorts...
VIP Independent Call Girls in Andheri 🌹 9920725232 ( Call Me ) Mumbai Escorts...VIP Independent Call Girls in Andheri 🌹 9920725232 ( Call Me ) Mumbai Escorts...
VIP Independent Call Girls in Andheri 🌹 9920725232 ( Call Me ) Mumbai Escorts...
 
VIP Independent Call Girls in Mira Bhayandar 🌹 9920725232 ( Call Me ) Mumbai ...
VIP Independent Call Girls in Mira Bhayandar 🌹 9920725232 ( Call Me ) Mumbai ...VIP Independent Call Girls in Mira Bhayandar 🌹 9920725232 ( Call Me ) Mumbai ...
VIP Independent Call Girls in Mira Bhayandar 🌹 9920725232 ( Call Me ) Mumbai ...
 

Portfolio Selection with Artificial Neural Networks

  • 1. Portfolio Selection with Artificial Neural Networks Andrew J Ashwood Supervisor: Dr Anup Basu 1
  • 2. Outline • Introduction • Research Motivation • Neural Network Primer • Literature Review • Research Questions • Network specification • Results • Conclusions • Further work and limitations 2
  • 3. Research Motivation • Trading in equities is big business in Australia – Value of All Ordinaries c. $1.2 trillion – Stocks turned over in 2012 c. 244 billion – Funds invested in superannuation c. $1.4 trillion 3
  • 4. Research Motivation • Outperforming the broader market is difficult – From 1969 to 2010, an index investor would achieve returns 80 per cent greater than the average investor in a managed fund.’ (Malkiel, 2011) – ‘…if many managers have sufficient skill to cover costs, they are hidden by the mass of managers with insufficient skill. …true α in net returns to investors is negative for most if not all active funds’ (Fama & French, 2009) 4
  • 5. Research Motivation • The Australian experience is similar – 70% of active retail funds underperformed the benchmark over both the 1 and 3 year period (Karaban & Maguire, 2012) – Over a five year period 69% of active retail funds failed to outperform the ASX200AI (Karaban & Maguire, 2012) 5
  • 6. Research Motivation • ‘The attempt to predict accurately the future course of stock prices and thus the appropriate time to buy or sell a stock must rank as one of investors’ most persistent endeavours.’ (Malkiel, 2011) • Advances in computing power in combination with the widespread availability of historical datasets has provided investors with increased opportunity to test markets for predictable returns 6
  • 7. Research Motivation • Artificial Neural Networks (‘ANNs’) have several traits that suggest their suitability for stock price prediction – Ability to generalise and learn – Provide an analytic alternative to other techniques that rely on assumptions of normality, linearity, and variable independence 7
  • 8. Neural Network Primer • An artificial neural network is a mathematical model that mimics the structure and function of biological brains • A structure of highly interconnected processing neurons 8
  • 9. Neural Network Primer • Typical neural processing element X1 X2 Xn Neuron Inputs X1 to Xn W1 W2 Wn Input Weights W1 to Wn WX i n i i v 0 Neuron Activation Transform- ation )tanh(vy Output Output Learning Error Back propagation 9
  • 10. Neural Network Primer • Neural networks comprise interconnected structure of neurons Input Input Input Output Output Output Input Layer Hidden Layer Output Layer 10
  • 11. Neural Network Primer • Network typology – 2 fundamental classes of network architecture • Feed-forward network – information travels in the forward direction only – Single layer feed-forward network – Multi layer feed-forward network • Recurrent network – contains at least one feedback loop – Many types (Fully recurrent, Hopfield, Elman, Jordan, etc etc) – Several studies demonstrate that feed- forward networks outperform recurrent networks at predicting non-linear time series data 11
  • 12. Neural Network Primer • Network learning – The Back Propagation Algorithm – Input weights are randomly applied to the network – Network calculations occur and the network produces an output signal – Simulated output compared to target output to provide system error – Error is back-propagated through the network – Process repeats until error reaches desired level or maximum number of iterations is reached 12
  • 13. Literature Review • Neural networks are relatively new. – No published research in the financial domain prior to 1990. By 1993, over 10 studies per year were being published. (Wong & Yakup, 1998) 13
  • 14. Literature Review • Yoon & Swales (1991) used ANNs to predict prices of 98 companies and compare the ANN predictions to multiple discriminant analysis – The study concluded the ANN technique significantly outperformed multiple discriminant analysis 14
  • 15. Literature Review • Kryzanowski, Galler & Wright (1992) used ANNs to pick stocks – Fourteen fundamental ratios were used as inputs – coded as trending upwards/downwards or stable – The ANN model correctly predicted price directional movement over the next year in 66% of cases – The ANN did not predict extreme movements well 15
  • 16. Literature Review • Hall, 1994 used ANN to create style based portfolio – ANNs used to identify stocks with largest recent price increased and apply this style to the current market to select a portfolio – Utilised 1,000 largest US stocks – 37 inputs used to predict future stock prices – The portfolio has exceeded the S&P500 since inception after transaction costs, but, ‘Actual performance results and specific details of the system are proprietary’. 16
  • 17. Literature Review • Jai & Lang (1994) used ANNs to trade the TSE – Used 16 technical indicators to predict the trend of the price movement over the subsequent 6 trading days Year 1990 1991 Type of ANN Fixed structure Dual adaptive structure Fixed structure Dual adaptive structure Total return % -30.66% 21.61% 25.76% 29.20% Annual TSEWPI return% -39.52% 23.42% Note the large divergence in performance 17
  • 18. Literature Review • Freitas et al. (2001) used neural network to provide an estimate of future returns of 66 representative stocks listed on the Brazilian BOVESPA stock exchange. – Worst network prediction error 33% – Best network prediction error 7% – Average error almost 17% – Neural network portfolios achieved better performance in 19 out of 21 weeks and outperformed the benchmark portfolio by 12.39%. 18
  • 19. Literature Review • Ellis & Wilson (2005) used ANNs to classify Australian listed REITs as value stocks. – Value stocks were bought and portfolio returns measured – ANN portfolio outperformed the benchmark index 6.97% per month on average – Sharpe and Sortino ratios significantly greater than the benchmark indices 19
  • 20. Literature Review • Ko & Lin (2008) used ANNs for portfolio selection on Taiwan Stock Exchange (‘TSE’). – Focussed on 21 different equities – Used price, variance, covariance as input variables – ANN used to predict equity allocation ratio – 5yr data set including training & validation sets – ANN model outperformed TSE (no 20
  • 21. Literature Review • Vanstone et al. (2010) used neural networks based on a stock trading filter rule. The filter rule bought stocks when the following criteria were met • PE < 10 • Market Price < Book Value • ROE < 12 • Dividend Payout Ratio < 25% – System yielded low number of trading opportunities – The return achieved by the system exceeded the return available using the original filter rule – It is worth noting however that the higher return achieved, was at least partially due to increased risk. 21
  • 22. Research Questions • Core Research Question 1 – Can neural networks that utilise a combination of fundamental and technical inputs predict (a) future stock prices, (b) stock price directional movement? • Secondary Research Question 1 – Which network inputs contribute most to network accuracy? 22
  • 23. Research Questions • Secondary Research Question 2 – What network parameters lead to optimal predictive capability? • Secondary Research Question 3 – Do neural networks predict any of the known performance attribution factors? • All research questions are to be answered with reference to stocks listed on the ASX 23
  • 24. Network Specification • Kaastra & Boyd (1996) framework for design of financial and econometric time seriesStep 1: Variable selection Step 2: Data collection Step 3: Data preprocessing Step 4: Training, testing, and validation sets Step 7: Neural network training number of training iterations learning rate and momentum Step 6: Evaluation criteria Step 5: Neural network paradigms number of hidden layers number of hidden neurons number of output neurons transfer functions Step 8: Implementation Iterative Process 24
  • 25. Variable Selection FUNDAMENTAL INPUTS TECHNICAL INPUTS Profit Margin DPS Open High Low Close Dividend Yield CF per share 20d MA 50d MA 20d EMA 50d EMA BV per share BV per share growth MACD RSI CLV ADI Current ratio Quick Ratio 20d Momentum 50d Momentum 6mth Momentum 12mth Momentum Asset Turnover Debt to Equity Fast Stochastic Oscillator Fast Stochastic Indicator Slow Stochastic Oscillator Slow Stochastic Indicator Interest Cover Price to CF Rate of Change (Price 1wk) Rate of Change (Price 4wk) Rate of Change (Price 10wk) Rate of Change (MACD) PE Price to Sales Rate of Change (MACD Signal) Rate of Change (RSI) Rate of Change (ADI) Rate of Change (20d MA) ROA ROE Rate of Change (50d MA) Rate of Change (20d EMA) Rate of Change (50d EMA) Rate of Change (12mth Mom) FCF per share Sales growth Rate of Change (FSO) Rate of Change (FSI) Rate of Change (SSO) Rate of Change (SSI) - - Volume - - - 25
  • 26. Data Set • Dataset obtained from Bloomberg • Weekly data from Jan 1997 to Dec 2011 • Early data (1997-1999) used to calculate technical indicators • 20 widely held ASX50 stocks selected AMC ANZ BHP CBA CCA CSL DJS GPT LEI NAB ORG QAN QBE RIO SUN TLS WBC WDC WES WOW 26
  • 27. Data Preprocessing • 3 pre-processing algorithms utilised – – Remove duplicate data • Duplicate data provides no useful information to the ANN • Speeds calculation – Scale data • Ensures all inputs treated equally • Avoids saturation of the tansig transformation function – Mark missing data • Ensures that missing data is not used in the learning process 27
  • 28. Frequency of Data • Little theory to assist in determining optimum frequency of data for a given prediction horizon Deboeck (1994) – – Long term prediction are unreliable – Short term predictions are unreliable – There is some period into the future for which useful predictions can be made 28
  • 29. Training, Testing, Validation Sets • The data set needs to be divided into training, validation, and testing datasets (Beale et al., 2011)– – Training set –used to compute the gradient and updating weights – Validation set – the data set used for monitoring the error during training – Testing set –out-of-sample test data set 29
  • 30. Training, Testing, Validation Sets • Walk forward testing – t0 tn Time Window 1 Window 2 Window 3 Window 4 Window 5 Training Window Validation Window Testing Window Testing Period 30
  • 31. Network Implementation Summing junction+ Tapped delay lineBias unitbx Input WeightTDL IW LW Layer Weight Transfer function Input Layer 5918 Fundamental Inputs 41 Technical Inputs 59 Tansig b1 4 → 20 TDL 1 Hidden Layer 30-150 IW1,1 + LW1,2 TDL Output Tansig Output Layer b2 LW2,1 1 1 + 31
  • 32. Evaluation Criteria & Training • Mean Squared Error (‘MSE’) selected as error function to be minimised by the network • Convergence method adopted for network training. Training occurs until the earlier of – – 100,000 training iterations occurs OR – 50 iterations with no reduction in validation set error 32
  • 33. Implementation Parameter Options No. of Options Stocks/portfolios 20 stocks + Equally and value weighted portfolios of these stocks 22 Input type Price OR Technical indicators OR Fundamental indicators OR Price + Technical OR Price + Fundamental OR Price + Technical + Fundamental 6 Lookback window 4, 8, 12, 16, 20 periods 5 Hidden layer size 30, 60, 90, 120, 150 neurons 5 Training period length 3mths, 6mths, 12mths, 24mths 4 Walk forward testing 10 x 1-year testing periods from 2001 to 2011 10 Total networks = 132,000 • All testing undertaken on QUT supercomputer ‘Lyra’ • At time of testing Lyra consisted of - • 106 compute nodes • 1572 x 64 bit Intel Xeon cores (consisting 11,736 core processors) • 32.8 TeraFlop Theoretical (double precision), 58.6 TeraFlop (single precision) • 10.4 TeraBytes of main memory 33
  • 34. Results • The high number of network models specified (132,000) meant that manual review and interpretation of individual networks was not feasible. • The algorithm selected the network configuration that provides the lowest mean squared error over the one year walk forward testing window. The algorithm therefore provides 220 outputs. This is comprised as follows: – 20 stocks + 2 portfolios – 10 walk forward testing windows for each stock/portfolio – Total outputs = 220 34
  • 36. Lookback Window 0 5 10 15 20 25 30 35 40 45 50 4 8 12 16 20 Frequency Lookback Period (no. inputs) Histogram of Lookback 19% 21% 19% 21% 20% 36
  • 37. Hidden Layer Size 0 20 40 60 80 100 120 30 60 90 120 150 Frequency Hidden Layer Size (no. of nodes) Histogram of Hidden layer Smaller hidden layer produced optimum network in 47% of cases 47% 16% 9% 18% 10% 37
  • 38. Training Period 0 10 20 30 40 50 60 70 80 3mths 6mths 12mths 24mths Frequency Training Period Length Histogram of Training period 35% 20% 15% 30% 38
  • 39. Predictive Capability 0% 5% 10% 15% 20% 25% 30% 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 1 2 3 4 5 6 7 8 9 10 RMSE%ofSharePrice Year/ Timestep RMSE as % of Share Price AMC ANZ BHP CBA CCL CSL DJS GPT LEI NAB ORG QAN QBE RIO SUN TLS WBC WDC WES WOW Generally there is a major spike in error in 2008 & 2009 Generally predictions are fairly consistent from 2002- 2007 39
  • 40. Predictive Capability 2,000 2,500 3,000 3,500 4,000 4,500 5,000 5,500 6,000 6,500 7,000 ASX200 - 2002 to 2011 ASX200 Adjusted Close Price Sustained benign trading conditions leading up to GFC 2008 Major reversal, Bear Market 2009 Major reversal, Bull Market 2010 Return to benign trading conditions 40
  • 41. Directional Movement 0% 10% 20% 30% 40% 50% 60% 70% 80% 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 1 2 3 4 5 6 7 8 9 10 PredictionAccuracy Year/ Timestep Directional Movement Prediction Accuracy AMC ANZ BHP CBA CCL CSL DJS GPT LEI NAB ORG QAN QBE RIO SUN TLS WBC WDC WES WOW EW VW 41
  • 42. Directional Movement 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% PredictionAccuracy Stocks & Portfolios Directional Movement Prediction Accuracy Over 10-year Testing Horizon AMC ANZ BHP CBA CCL CSL DJS GPT LEI NAB ORG QAN QBE RIO SUN TLS WBC WDC WES WOW EW VW Significant at 5% level 42
  • 43. Portfolio Selection • The simulated price returns were then used to form a long and a long-short portfolio • Portfolios formed at the beginning of each month during the 10-year testing period and held until the end of the month • Stocks weighted according to the absolute value of their simulated price return 43
  • 44. Portfolio Price Returns Portfolio Price Return (2002-2011) ASX200 19.6% ANN Long Portfolio 205.5% ANN Long-Short Portfolio 196.6% • Raw Results – • Do not include transaction costs • No performance measurement framework 44
  • 45. Regression Results Long Portfolio Long-Short Portfolio α 0.016** 0.013*** RMRF 0.712*** -0.011 SMB -0.317 -0.604** HML -0.416** -0.306 WML -0.011 0.038 R2 0.283*** 0.215*** Performance Measurement • The portfolio price returns were then regressed against the Carhart 4-factor model Both portfolios have a significantly positive alpha Both models are significant 45
  • 46. Optimal Network Portfolios • This analysis confirms that ANNs can be used for portfolio selection to achieve positive alpha HOWEVER • These results are based upon the most accurate network specification – which is only know post hoc • The problem is whether ANNs can achieve positive alpha without this post hoc knowledge 46
  • 47. Median Network Portfolios • For each walk-forward testing window for each stock, 600 different network specifications were trialled – 6 input options – 5 lookback windows – 5 hidden layer sizes – 4 training period lengths – 600 network options per stock per testing period • We have confirmed that the most accurate network achieves positive alpha – what about the median network? 47
  • 48. Regression Results Long Portfolio Long-Short Portfolio α 0.012** -0.001 RMRF 0.137 0.003 SMB -0.050 -0.011 HML -0.310 -0.064 WML -0.326 -0.115 R2 0.084* 0.012 Median Network Portfolios Only the Long portfolio achieved a significant alpha Only the Long portfolio model is significant and the Carhart style- factors explain a lower proportion of variation in data 48
  • 49. Distribution of Alphas • Network parameter combinations were applied uniformly to all stocks and the long portfolio based on expected returns was formed. • Mimics a real world trading situation • This process resulted in 600 measurements of alpha. 49
  • 50. Distribution of Alphas 0 20 40 60 80 100 120 Histogram for Alpha Median = 0.0061 Only 12% of the networks tested have negative alpha 127 of 529 positive alphas (27%) were statistically significant (p < 0.05) 50
  • 51. Conclusions • Network parameter testing was undertaken into several network parameters – – input type – hidden layer size – lookback window – training period length. • No useful guidance on heuristics for the input type, lookback window, or training period length. • Hidden layer size showed some promise. Almost half of the most accurate networks models utilised the smallest hidden layer of 30 nodes. 51
  • 52. Conclusions • Price prediction performance was mixed. • Most networks performed poorly at the beginning of the global financial crisis (beginning of 2008) and at the end (end of 2009). The networks fail to predict major reversals. • The networks did not predict the portfolio price returns well. The value weighted portfolio error ranged from 9% (best year) to 32% (worst year) and only achieved a RMSE below 10% in 1 of the 10 testing years. 52
  • 53. Conclusions • Stock directional movement prediction performance was assessed. • The ANN models predicted directional movement with better than 50% accuracy for all stocks and portfolios except for BHP. • Significant results were achieved for 13 of the 22 stocks and portfolios tested. 53
  • 54. Conclusions • Price predictions used for long and long- short portfolio selection • Both portfolios achieved significantly positive alpha. • The optimal network specification was made by comparing the simulated prices to the target prices, thus requiring post hoc knowledge. 54
  • 55. Conclusions • A distribution of alpha for all 600 different network specifications was produced. • Could not reject hypothesis that the distribution was normal (p < 0.05). • 88% of the networks produced positive alpha (529 out of 600). 127 of which were significant (27%). • 12% of networks produced negative alpha (71 out of 600). None of which were significant. 55
  • 56. Limitations & Future Research • Research tested several network parameters. • Hidden layer size provided promising results. • Further research is required to develop heuristics. • Future research should look at individual stocks and industry sectors. 56
  • 57. Limitations & Future Research • The analysis has been undertaken on a price return basis, rather than an overall returns basis. • Study utilised 20 widely held stocks. Needs testing over wider investment universe. Survivorship bias. • Impact of trading costs not included in analysis 57

Hinweis der Redaktion

  1. Trading in equities is big business in Australia. The market capitalisation of the 500 largest publicly listed corporations in Australia is approximately $1.2 trillion dollars. In the twelve month period to January 2013, 244 billion stocks of these top 500 Australian companies were changed hands. Most Australians have a large proportion of their personal wealth invested in equities through superannuation – the current amount of funds invested in superannuation is over $1.4 trillion – much of which is invested in equities.By any measure, the value of Australia’s stock market is significant and the volume of trading activity demonstrates eagerness on the part of investors to buy and sell stocks in the hope of achieving above average returns.
  2. Burton Malkiel – A Random Walk Down Wall Street - ‘An investor with $10,000 at the start of 1969 who invested in a Standard &amp; Poor’s 500-Stock Index Fund would have had a portfolio worth $463,000 by 2010, assuming that all dividends were reinvested. A second investor who instead purchased shares in the average actively managed fund would have seen his investment grow to $258,000. The difference is dramatic. Through May 30, 2010, the index investor was ahead by $205,000, an amount almost 80 per cent greater than the final stake of the average investor in a managed fund.’ (Malkiel, 2011)Fama &amp; French - Luck vs skill in the cross section of mutual fund returns‘For fund investors the simulation results are disheartening. When α is estimated on net returns to investors, the cross-section of precision-adjusted α estimates suggests that few active funds produce benchmark adjusted expected returns that cover their costs. Thus, if many managers have sufficient skill to cover costs, they are hidden by the mass of managers with insufficient skill. On a practical level, our results on long-term performance say that true α in net returns to investors is negative for most if not all active funds, including funds with strongly positive α estimates for their entire histories.’ (Fama &amp; French, 2009)
  3. Karaban &amp; Maguire (Spiva Australian Scorecard) S&amp;P indices vs active funds scorecardA majority of active funds underperformed their respective benchmarks. Excluding small-cap funds, over 70% of active retail funds underperformed the benchmark over both the 1 and 3 year period (Karaban &amp; Maguire, 2012)
  4. These advances in computing power and the widespread availability of datasets raises the question – can advanced computational techniques provide an alternative method for portfolio selection?
  5. Ignoring transaction costs, my research intended to determine if neural networks can be successfully implemented to predict price movements of widely held Australian stocks. Further, the study seeks to examine whether these price predictions can be used for portfolio selection and whether these portfolios achieve positive alpha when measured using the Carhart 4-factor model.
  6. The neuron receives stimulus (input) from its network of dendrites. The input stimulus is then processed in the nucleus of the cell body. The neuron then sends out electrical impulses through the long thin axon. At the end of each strand of the axon is the synapse. The synapse is able to increase or decrease its strength of connection and causes excitation or inhibition in the connected neuron.In the biological brain, learning occurs through the changing of the effectiveness of the synapses, so that the influence that one neuron has on its connected neurons change.
  7. Just like in a biological neural network, the fundamental building block of an ANN is the individual neural processing element or neuron.Each neural processing element can have multiple inputs (the only real limiting factor is computational power). Each of these neuron inputs is assigned an input weight (which is randomly generated when the network is initialised).The next step is for the neuron activation to be determined – by summing the product of the input by its weight. The summation function calculates the activation level of the processing element. It is the artificial method of determining the strength of the input stimulus received by a biological neuron’s network of dendrites. Once neuron activation has occurred, the activation is transformed using a transformation function. There are numerous alternatives – but the function has to be a bounded, differentiable function. The purpose of the transformation function is to scale the output of the processing element into a useable form (generally between -1 and +1 or between 0 and +1). The neuron output that is then calculated using the transformation function can be considered a measure of neuron ‘excitement’.This is then compared to the target output to determine the neuron error – which can then be back-propagated through the neuron to update the input weights.
  8. Just like a biological brain is composed of numerous biological neurons, an ANN is composed of numerous artificial neural processing elements. This example shows an ANN with 9 neural processing elements (3 in the input layer, 3 in the hidden layer, and 3 in the output layer).As you can see the information travels in the forward direction only so this is a feedforward neural networks.
  9. Single-layer feed-forward networks –The simplest form is the single layer feed-forward network where the input stimulus is received directly into the output layer of neurons. Information travels only in the forward direction and feed-forward networks do not contain feedback loops.Multilayer feed-forward networks – The addition of extra layers to a single layer feed-forward network results in a multilayer feed-forward network. A multilayer feed-forward network contains one or more hidden layers. These hidden layers contain neurons that intervene between the input stimulus and the network output and enable the network to extra higher-order statistics (Haykin, 1999). In a multilayer feed-forward network the input layer feeds into the first hidden layer of neurons. This hidden layer undertakes computations and provides an output that is sent to the next hidden layer of neurons. This continues until the output layer is reached where the network provides a final output signal. The example on the previous slide was a multi-layer feed forward network.Recurrent networks – The recurrent network differs from feed-forward networks in that it contains at least one feedback loop. Several studies demonstrate that feed-forward back propagation networks outperform recurrent networks at predicting non-linear time series data – Hallas &amp; Dorffner – a comparative study using feedforwardvs recurrent neural networks to forecast a series of 8 different non-linear time series data Dematos et al – was a comparative study using feedforwardvs recurrent neural networks for forecasting monthly Japanese Yen exchange rates
  10. The back propagation algorithm relies on the delta rule which is a gradient descent learning rule used to update input weights applied to the network processing elements. Put simply, the input weights are updated according to the negative of the gradient of the error term for each node.
  11. The study utilised 9 inputs – confidence, economic factors outside the firm’s control, growth, strategic gains, new products, anticipated loss, anticipated gain, long-term optimism, short term optimism. The network output was a classification as either: a firm whose stock price performed well and a firm whose stock price performed poorly.
  12. 5 year datasetFourteen fundamental ratios were used as inputs – coded as trending upwards/downwards or stableThe ANN model correctly predicted price directional movement over the next year in 66% of cases The ANN did not predict extreme movements well (eg the ANN predicted a negative return for a stock which increased in price 150% and similarly predicted a positive return for a stock which lost 73%) – this is something that is relevant to my experimentation that we will revisit
  13. The study utilised a two year testing period only. Several of the results are significant. Firstly both the fixed structure and dual adaptive structure ANNs outperformed both the market and the comparison funds during both testing periods. Secondly, the dual adaptive structure outperformed the fixed structure ANN. Thirdly, both ANNs provided relative few trading signals (fixed structure 18 trades over 2 years; dual adaptive structure only 8 trades over 2 years). No performance measurement framework was utilised. Also note that while the two ANN models provided very different results in 1990, they both still outperformed the TSE.Many of the 16 technical indicators used in this study were utilised as input variables in my research.
  14. ANN that used past prices to predict future prices for each of the 66 stocksThese prices were used to calculate future returns and these predicted future returns were input as expected returns into the mean-variance modelThe worst performing network provided outputs that were on average 33% different from the target values. The best performing network achieved an average difference from the target values of just over 7%. The average error across the network for all 66 stocks was almost 17%. It was reported that when the profitability of the portfolios formed using the neural network predicted expected returns was compared to benchmark portfolios, the neural network portfolios achieved better performance in 19 out of 21 weeks and outperformed the benchmark portfolio by 12.39%. It should be noted that the research relies on a very short testing period – 21 weeks.The research by Freitas et al. provides some interesting insights and challenges to using neural networks for portfolio selection. While their research shows that neural networks have definite limits in terms of predictive accuracy, it also seems to suggest that notwithstanding the quantum of the error in the network, the relative expected return predicted can still be useful in the portfolio selection process.
  15. Ellis &amp; Wilson was an study where fundamental inputs were used in an ANN to classify Australian listed REITs as value stocks. The ANN portfolio outperformed the benchmark index by 6.97%/mth. The return of the benchmark index per month was approximately 1% while the ANN portfolio achieved about 8% per month – a dramatic difference. Of particular note was that the Sharpe and Sortino ratios for the ANN portfolios were significantly greater than for the benchmark indices
  16. While this study provides some interesting results, the 5 year dataset that includes training, validation, and testing datasets is considered too short to be robust. The returns have not been subject to a performance measurement framework.
  17. This research utilised an ANN that used the above fundamental measures as input variables to create an output signal that was proportional to expected return over a 1 year investment horizonThe investment strategy then determines what stocks to buy based upon whether the output signal strength is above an arbitrary threshold. The out of sample testing period of 5 years created only 38 trading opportunities. While the number of trading opportunities was low, the return achieved by the system exceeded the return available using the original filter rule. It is worth noting however that the higher return achieved, was at least partially due to increased risk.
  18. Kaastra and Boyd (1996) have proposed a framework that can be utilised when designing neural networks for financial and economic time series forecasting. This 8 step design process builds on previous work by Deboeck (1994), Masters (1993), Blum (1992). This 8-step process has been utilised for network specification.
  19. In order to fully address the research questions and in an attempt to produce an ANN with the highest possible predictive power, a range of different network inputs were utilised. 59 network inputs in total – 18 fundamental, 41 technicalBased upon fundamental inputs used in Kryzanowski, Galler &amp; Wright (1992) ANN study and the fundamental measured used in the fundamental indexation research by Arnott, Hsu, &amp; Moore (2005) and Ellis &amp; Wilson (2005)Many of the technical inputs from Jai &amp; Lang (1994) ANN study into predicting the TSE. Their inputs focussed on using moving averages, stochastic oscillators, and rates of change of these measures. They did not use price as an input. Momentum over different periods along with rates of change of momentum were used based on the Jegadeesh &amp; Titman studies demonstrating momentum in stock returns. PLUS OTHERS IN AN ATTEMPT TO ACHIEVE MAXIMUM PREDICTIVE CAPABILITY
  20. The 10 largest stocks included in this research represent over 50% of the All Ordinaries. Note that as these stocks were individually selected the study suffers from survivorship bias.
  21. Remove duplicate - Constant values do not provide the neural network with any information and can cause numerical problems for some network functions. This algorithm also assisted with efficient processing because it meant that the network was minimizing the number of calculations required. Scale – algorithm to scale each input series to values between -1 and 1. Two reasons for adopting this scaling process. (1) Ensures that all inputs are treated equally by the ANN. Some network inputs are many orders of magnitude greater than others (for example compare market capitalisation measured in billions to sales growth which generally will be less than 1. Scaling all inputs ensures that the network calculations aren’t dominated by inputs with large orders of magnitude. (2) The transformation function (tan sigmoid) squashes the neuron inputs to a value between -1 and 1. This function saturates quickly if input data isn’t pre-processed. For example when utilizing a hyperbolic tan function once an input value reaches 10, its hyperbolic tangent is essentially saturated. Therefore in order to keep input values within a meaningful range, the data must be pre-processed.Mark Missing - The algorithm utilized looked at all input series and located missing data. Wherever a missing data point was located the data point was marked with a string. During the network implementation phase, the learning algorithm recognised this string and skipped this data point so that no learning occurred as a result of this missing data.
  22. Doboeck, 1994 - ‘The evolutionary nature of the stock selection problem and the probability of uncontrollable external events make long-term predictions unreliable. The large amount of noise inherent in this type of system also makes most short-term predictions unreliable. However, there is some period into the future for which useful predictions can be made. Discovering the proper predictive period is a critical task in modelling evolutionary, complex NLD [non-linear dynamic] systems…For problems such as stock selection, an estimate of the predictive period must be made by experimentation over a period of several system cycles…For relationships of this complexity, there is no true optimum. The only method for discovering the proper predictive period is through repeated simulation of the entire system, including simulated trading, over an extensive period of time.’ (Deboeck, 1994)For the purposes of this research weekly input data has been used to predict stock prices four periods ahead.
  23. Training set – the data set used for computing the gradient and updating the network weights and biasesValidation set – the data set used for monitoring the error during training. Typically during training the error on the validation data set decreases until over-fitting starts to occur at which point error starts to increase again. Network learning parameters are set to ensure that over-fitting does not occur. Testing set – the data set is an out-of-sample test used when network training is complete to test how accurate the network is and therefore to compare network performance.
  24. While a static system can be easily modelled by using a random data division algorithm, time series prediction requires the use of moving window testing. …walk forward testing involves dividing the data into a series of overlapping training-testing-validation sets. Each set is moved forward through the time series…Walk-forward testing attempts to simulate real-life trading and tests the robustness of the model through its frequent retraining on a large out-of-sample data set. Walk forward testing is considered the best method for prediction of time series data. It simulates the real-life trading situation whereby the model is regularly retrained with data as it becomes available (an implicitly Bayesian approach). The frequent retraining is more time-consuming but allows the network to adapt to changing market conditions. One of the network parameters tested is training period. Several options were tested – 3, 6, 12, 24 months
  25. Input vector p which consists of 18 fundamental inputs and 41 technical indicators (59 inputs total). The size of this input layer varies depending on the input configuration. It ranges from 4 at a minimum when only price data is used to 59 when all technical and fundamental inputs are used. Six different input configurations are tested (1) Price (2) Technical indicators (3) Fundamental indicators (4) Price + Technical (5) Price + Fundamental (6) Price + Technical + FundamentalHidden layer – The use of hidden layers is what gives ANNs the ability to generalise. Previous research has concluded that generally network performance improves as the number of hidden nodes is increased up to a certain point, after which the addition of further units impairs network performance. There are some heuristics available but they provide a very large range of values (from 4 to 188 for this network). Several hidden layer sizes are tested (30, 60, 90, 120, 150). The hidden layer also includes a bias unit that acts as a neuron scalar and is used to teach the network the appropriate activation threshold – by shifting the activation function to the left or right. A non-linear transformation function has been utilised. Tansig scales each hidden layer output to a value between -1 and 1. The hidden layer also includes a tapped delay line that varies in size between 4 and 20 periods. This means that for each prediction made, the network is fed the last 4-20 pieces of input data. A larger tapped delay line (20 periods) means that the network has a longer memory upon which to make a prediction. Output layer – The output layer contains a single node as it provides a single predicted output – PRICE at t+4. Similar to the hidden layer, the output layer contains a bias node. The implementation diagram shows a tapped delay line from the output layer. This is how the network functions for the testing period only (it functions as an autoregressive network that contains feedback). The network is trained as a feed-forward only network as this creates the most accurate network possible. However this uses real values for price to make a one step ahead predication. Once training is complete and we wish to make multi period predictions we must let the network predicted prices. This is what the tapped delay line does. In effect, we alter the network from being a feedforward back propagation network during the training stage to using an autoregressive structure for testing.
  26. This configuration approach ensured that the network could not get trapped in a local minimum unless 50 sequential iterations occurred without reaching the boundary of the local minimum. Increasing the number of iterations in the validation stop rule increases the likelihood that the network will locate a global minimum.
  27. Testing occurred over an extended period of time utilising up to 88 computing nodes in parallel. Run times for individual networks varied widely. Some networks were trained in several minutes due to the validation stop rule leading to early training termination. Run times for other networks extended up to several days when close to the full 100,000 epochs were reached.While 132,000 networks were simulated in total – this equates to 600 networks per stock over the entire 10-year testing periodOf course price is a technical indicator, but for the purposes of this research it was treated separately so that its effect as an input variable could be measured.
  28. Instead an algorithm was created to determine the best network model for each of the 20 stocks and 2 portfolios. The optimum network configuration for each timestep in the walk-forward testing window was allowed to change each year so for each twelve month testing period a different network specification could be adopted.
  29. The histogram on this slide shows how frequently each of the input types specified the most accurate network model. 15% of the testing periods were most accurately predicted using past price information only. Further, 68% of the most accurate networks were predicted using either price alone or in combinations with technical and fundamental variables. This suggests that any successful price prediction system must contain past price information as an input variable. Interestingly, the network configurations that were most successful were a combination of price and technical indicators or price and both technical and fundamental indicators. This suggests that the addition of fundamental and technical indicators enhances the predictive capability of networks.
  30. The histogram on this slide shows how frequently each of the lookback window options specified the most accurate network model. Lookback is the number of previous inputs to give the network to make the next prediction.16 period lookback window was most successful, specifying the most accurate model in 22% of cases. The remaining lookback window options performed similarly, specifying the most accurate model between 19 – 20% of cases. While not uniform, the frequency distribution suggests that the optimum lookback window is idiosyncratic to the stock, or possibly the industry sector. This provides an interesting opportunity for future research.
  31. The histogram on this slide shows how frequently each of the hidden layer sizes specified the most accurate network model. optimum hidden layer size varies significantly between stocks. A smaller hidden layer of 30 nodes was the mode, producing the most accurate model in almost half of all cases (47%). The other hidden layer sizes produced the optimal model in similar frequencies ranging between 8-17%. The histogram suggests an inverse relationship between the number of hidden layer neurons and the network accuracy. While not uniform, there appears to be a general downward trend in predictive capability as the number of hidden layer neurons is increased.
  32. The histogram on this slide shows how frequently each of the training period lengths specified the most accurate network model. suggests little in the way of trend between network predictive capability and training period length. However from a practitioner’s perspective of trying to define reliable heuristics for network configuration, this result is consistent with the idea that different stocks and stock market segments react different to market and trading conditions. For example, consumer staples (such as Woolworths) are expected to react less to market contractions as consumer spending on staples is fairly inelastic to macro-economic conditions. This should be contrasted with the real estate investment trust sector which should be much more sensitive to adverse macro-economic conditions. The results of this testing provide a fertile ground for further research testing optimum training period length according to classification of stocks by industry segment.
  33. There is a significant degree of variation in network predictive capability both over time and between stocks. For some stocks, the best performing networks consistently achieve RMSE less than 10% of the average stock price over the period, but from time to time exceed this amount. For other stocks and portfolios, even the most accurate networks fail to achieve RMSE less than 10% of average stock price more than once or twice over the 10 year testing horizon. Interestingly there appears to be little evidence of reliability of predictive capability over long periods of time. For example, when predicting stock prices for ANZ, the networks perform very well from 2002 to 2007. Over this six year time period, the networks achieve &lt; 10% RMSE in all years and =&lt; 5% RMSE in five out of the six years. However in 2008, the RMSE as a proportion of stock price increases dramatically to over 13% (about four times the error in 2007) and remains similarly high during 2009 (14%) before returning to lower levels in 2010 and 2011 (5%, 7% respectively). This pattern of fairly reliable predictability appears in several of the stocks tested; and generally the volatility in prediction accuracy occurs in the same period. For example GPT, CBA, AMC, NAB, WES all exhibit a roughly similar pattern to the ANZ neural network models.
  34. This slide shows the adjusted close price of the ASX200 over the ten year testing horizon from 2002 to 2012. The figure shows several features which likely lead to the poor predictive performance of the neural network models over the 2008-2009 period. Firstly the network model adopted to predict stock prices during 2008, was trained and validated using data from the lead up period. Regardless of whether the short (6mth) or long (24mth) training period was utilised for this prediction, the trading conditions during the training and validation periods were vastly different from the trading condition during the testing period. The trading conditions from July 2005 (assuming the long training period was utilised) up until the end of 2007 were generally positive, with limited volatility. This meant that the neural network learned how to respond during these benign, positive trading conditions and had no knowledge or experience of other types of market conditions. The trading conditions during the prediction period from January 2008 to December 2008 were characterised by strongly negative price movements with significant volatility. A similar situation occurred during the 2009 testing period. The beginning of 2009 signalled the market bottom and turnaround from the strong bear market that occurred during 2008. So while the 2009 testing period networks were trained over the period leading up to the peak at the end of 2007 and during the bear market of 2008, it is little wonder that the prediction made at the beginning of the turnaround in 2009 should contain a higher degree of error. These results highlight the major shortcoming of existing network models viz. their limited ‘memory’ so that they can only make accurate predictions where they have encountered similar situations during network training. This finding is consistent with the findings of Kryzanowski et al. (1992). This area provides significant opportunities for further research. While this research focussed on relatively short training periods (from six months to two years) it would be interesting to see how a network would perform if much greater training periods were used. Presumably to predict accurately during significant market turnarounds, the training dataset would need to contain similar events.
  35. These results are encouraging, but these raw results do not include transaction costs and have not been subjected to a performance measurement framework.
  36. These results appear to suggest that while the most accurate neural networks tend to achieve higher abnormal returns, less optimal networks might still be capable of achieving abnormal risk adjusted returns.
  37. In order to gain an understanding of the distribution of alphas generated by the different ANN specifications the different network parameter combinations were applied uniformly to all stocks and the long portfolio based on expected returns was formed. This method mimics the real world situation where the practitioner does not have prior knowledge of which network specification to apply to achieve maximum predictive capability; this analysis is analogous to the practitioner selecting a particular network specification and then applying this network to all stocks in the investment universe. This process resulted in 600 measurements of alpha.
  38. The positive median risk adjusted return indicates that there is a greater than 50% chance of randomly specifying a network that achieves a positive alpha. A one-sample t test was conducted on alpha to evaluate whether the mean alpha was significantly greater than zero (one-tailed test). The one-sample t test confirmed that the mean alpha was significantly greater than zero, t(599) = 30.283, p &lt; .001. Only 12% (71 out of 600) network specifications returned negative alphas. Non of these negative alphas were statistically significant. On the other hand, 529 of the 600 specifications – 88% - achieved a positive alpha. Of these positive alphas, 127 – 27% - were statistically significant. Lilliefors test confirmed that the null hypothesis that the distribution is normal cannot be rejected at the 5% level.
  39. ANNs were created and then used to predict future price returns of 20 widely held ASX200 stocks along with equally weighted and value weighted portfolios of these stocks. This testing did not provide any useful guidance on heuristics for the input type, lookback window, or training period length. In contrast the hidden layer size, while still demonstrating a degree of variability between stocks, generally performed better at the low end of the testing spectrum. Almost half of the most accurate networks models utilised the smallest hidden layer of 30 nodes.
  40. The ANN predictive capability was tested ever a 10 year period from 2002-2011. On a RMSE basis network performance was mixed. For some stocks the networks performed relatively well, predicting future prices with a RMSE at or below 5% for several years in a row. However even these relatively well performing models generally performed poorly at the beginning of the global financial crisis (beginning of 2008) and at the end (end of 2009). This appears to suggest that network learning is responding well to the training data but fails to make accurate predictions when a major reversal or change in volatility occurs. While some networks performed relatively well, there were others that performed poorly. In particular the networks did not predict the portfolio price returns well. The value weighted portfolio error ranged from 9% (best year) to 32% (worst year) and only achieved a RMSE below 10% in 1 of the 10 testing years.
  41. Stock directional movement prediction performance was assessed. Over the 10 year testing horizon the ANN models predicted directional movement with better than 50% accuracy for all stocks and portfolios except for BHP. For some stocks the directional prediction was only marginally better than 50% and was not significant. Significant results were achieved for 13 of the 22 stocks and portfolios tested. Five of these results were significant at the p &lt; 0.001 level.
  42. Simulated price returns from the most accurate network specification for each stock was then used to form a long and a long-short portfolio. Stocks were weighted according to the absolute value of their simulated return. Both portfolios achieved significantly positive alpha. These positive results however must be viewed with caution as the selection of the correct network specification was made by comparing the simulated prices to the target prices, thus requiring post hoc knowledge.
  43. Analysis indicated that the hypothesis that the distribution of alpha was normally distributed could not be rejected at the 5% level. This distribution demonstrated that about 88% of the network models produced positive alpha which was significant in 27% of cases (127 out of 600 cases). None of the statistically significant values of alpha were negative.
  44. The research looked at how several network parameters affected network predictive capability. With the exception of hidden layer size, the analysis did not provide guidance heuristics that could be used for ex ante network specification.Further research is required to examine the importance of each of these parameters and to develop and test different network specification heuristics. This research should look at individual stocks and industry sectors.
  45. The results of the study are encouraging. The ANN model, when correctly specified, can achieve positive alpha. There remain however several limitations to the study and more work is required to further refine and understand the ANN capabilities. The analysis has been undertaken on a price basis, rather than an overall returns basis. The ANN model predicts future prices that have been converted to price-returns and these have been used for the portfolio selection and performance measurement functions. Typically total returns should be utilised for the performance measurement function. Further work is required to investigate this limitation. There are several options available. The network model could be reconfigured to try and predict total returns rather than future prices and the portfolio selection and performance measurement could occur on a total returns basis. However the arbitrary and unpredictable nature of dividends could make ANN predictions more prone to error which may detract from their overall predictive capability. An alternative approach may be to use the ANN to make future price predictions and use this for the portfolio selection function but then use total returns in the performance measurement.The study utilised 20 widely held stocks listed on the ASX. While this approach was reasonable for proof of concept purposes, a wider study adopting a larger universe of stocks should be undertaken (for example the ASX200). However given the lack of network specification heuristics, it is suggested that this body of work should proceed once heuristics have been developed and perhaps the performance of the heuristics can be measured using this greater stock universe. The study also did not examine the impact of trading costs on portfolio returns. As the approach utilised by this study relies on active portfolio management future studies that build on this research should make provision for trading costs.