1 / 43

Introduction to Algorithmic Trading Strategies Lecture 5

Introduction to Algorithmic Trading Strategies Lecture 5. Pairs T rading by Stochastic Spread Methods. Haksun Li haksun.li@numericalmethod.com www.numericalmethod.com. Outline. First passage time Kalman filter Maximum likelihood estimate EM algorithm. References.

charis
Télécharger la présentation

Introduction to Algorithmic Trading Strategies Lecture 5

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Introduction to Algorithmic Trading StrategiesLecture 5 Pairs Trading by Stochastic Spread Methods Haksun Li haksun.li@numericalmethod.com www.numericalmethod.com

  2. Outline • First passage time • Kalman filter • Maximum likelihood estimate • EM algorithm

  3. References • As the emphasis of the basic co-integration methods of most papers are on the construction of a synthetic mean-reverting asset, the stochastic spread methods focuses on the dynamic of the price of the synthetic asset. • Most referenced academic paper: Elliot, van der Hoek, and Malcolm, 2005, Pairs Trading • Model the spread process as a state-space version of Ornstein-Uhlenbeck process • Jonathan Chiu, Daniel WijayaLukman, KouroshModarresi, AvinayanSenthiVelayutham. High-frequency Trading. Stanford University. 2011 • The idea has been conceived by a lot of popular pairs trading books • Technical analysis and charting for the spread, Ehrman, 2005, The Handbook of Pairs Trading • ARMA model, HMM ARMA model, some non-parametric approach, and a Kalman filter model, Vidyamurthy, 2004, Pairs Trading: Quantitative Methods and Analysis

  4. Spread as a Mean-Reverting Process • The long term mean = . • The rate of mean reversion = .

  5. Sum of Power Series • We note that

  6. Unconditional Mean

  7. Long Term Mean

  8. Unconditional Variance

  9. Long Term Variance

  10. Observations and Hidden State Process • The hidden state process is: • The observations: • We want to compute the expected state from observations.

  11. First Passage Time • Standardized Ornstein-Uhlenbeck process • First passage time • The pdf of has a maximum value at

  12. A Sample Trading Strategy • , • Buy when unwind after time • Sell when unwind after time

  13. Kalman Filter • The Kalman filter is an efficient recursive filter that estimates the state of a dynamic system from a series of incomplete and noisy measurements.

  14. Conceptual Diagram as new measurements come in prediction at time t Update at time t+1 correct for better estimation

  15. A Linear Discrete System • : the state transition model applied to the previous state • : the control-input model applied to control vectors • : the noise process drawn from multivariate Normal distribution

  16. Observations and Noises • : the observation model mapping the true states to observations • : the observation noise

  17. Discrete System Diagram

  18. Prediction • predicted a prior state estimate • predicted a prior estimate covariance

  19. Update • measurement residual • residual covariance • optimal Kalman gain • updated a posteriori state estimate • updated a posteriori estimate covariance

  20. Computing the ‘Best’ State Estimate • Given , , , , we define the conditional variance • Start with , .

  21. Predicted (a Priori) State Estimation

  22. Predicted (a Priori) Variance

  23. Minimize Posteriori Variance • Let the Kalman updating formula be • We want to solve for K such that the conditional variance is minimized.

  24. Solve for K

  25. First Order Condition for k

  26. Optimal Kalman Filter

  27. Updated (a Posteriori) State Estimation • So, we have the “optimal” Kalman updating rule.

  28. Updated (a Posteriori) Variance

  29. Parameter Estimation • We need to estimate the parameters from the observable data before we can use the Kalman filter model. • We need to write down the likelihood function in terms of , and then maximize w.r.t. .

  30. Likelihood Function • A likelihood function (often simply the likelihood) is a function of the parameters of a statistical model, defined as follows: the likelihood of a set of parameter values given some observed outcomes is equal to the probability of those observed outcomes given those parameter values.

  31. Maximum Likelihood Estimate • We find such that is maximized given the observation.

  32. Example Using the Normal Distribution • We want to estimate the mean of a sample of size drawn from a Normal distribution.

  33. Log-Likelihood • Maximizing the log-likelihood is equivalent to maximizing the following. • First order condition w.r.t.,

  34. Nelder-Mead • After we write down the likelihood function for the Kalman model in terms of , we can run any multivariate optimization algorithm, e.g., Nelder-Mead, to search for . • The disadvantage is that it may not converge well, hence not landing close to the optimal solution.

  35. Marginal Likelihood • For the set of hidden states, , we write • Assume we know the conditional distribution of , we could instead maximize the following. • , or • The expectation is a weighted sum of the (log-) likelihoods weighted by the probability of the hidden states.

  36. The Q-Function • Where do we get the conditional distribution of from? • Suppose we somehow have an (initial) estimation of the parameters, . Then the model has no unknowns. We can compute the distribution of .

  37. EM Intuition • Suppose we know , we know completely about the mode; we can find • Suppose we know , we can estimate , by, e.g., maximum likelihood. • What do we do if we don’t know both and ?

  38. Expectation-Maximization Algorithm • Expectation step (E-step): compute the expected value of the log-likelihood function, w.r.t., the conditional distribution of under and . • Maximization step (M-step): find the parameters, , that maximize the Q-value.

  39. EM Algorithms for Kalman Filter • Offline: Shumway and Stoffer smoother approach, 1982 • Online: Elliott and Krishnamurthy filter approach, 1999

  40. A Trading Algorithm • From , , …, , we estimate . • Decide whether to make a trade at , unwind at , or some time later, e.g., . • As arrives, estimate . • Repeat.

  41. Results (1)

  42. Results (2)

  43. Results (3)

More Related