1 / 23

Solving Infinite Horizon Stochastic Optimization Problems

Solving Infinite Horizon Stochastic Optimization Problems. John R. Birge Northwestern University (joint work with Chris Donohue, Xiaodong Xu, and Gongyun Zhao). Motivating Problem. (Very) long-term investor (example: university endowment)

gavivi
Télécharger la présentation

Solving Infinite Horizon Stochastic Optimization Problems

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Solving Infinite Horizon Stochastic Optimization Problems John R. Birge Northwestern University (joint work with Chris Donohue, Xiaodong Xu, and Gongyun Zhao)

  2. Motivating Problem • (Very) long-term investor (example: university endowment) • Payout from portfolio over time (want to keep payout from declining) • Invest in various asset categories • Decisions: • How much to payout (consume)? • How to invest in asset categories? • Complication: restrictions on asset trades

  3. Problem Formulation • Notation: x – current state (x 2 X) u (or ux) – current action given x (u (or ux) 2 U(x)) d – single period discount factor Px,u – probability measure on next period state y depending on x and u c(x,u) – objective value for current period given x and u V(x) – value function of optimal expected future rewards given current state x • Problem: Find V such that V(x) = maxu2 U(x){c(x,u) + d EPx,u[V(y)] } for all x 2 X.

  4. Approach • Define an upper bound on the value function V0(x) ¸ V(x) 8 x 2 X • Iteration k: upper bound Vk Solve for some xk TVk(xk) = maxu c(xk,u) + d EPxk,u[Vk(y)] Update to a better upper bound Vk+1 • Update uses an outer linear approximation on Uk

  5. Successive Outer Approximation V0 V1 TV0 x0 V*

  6. Properties of Approximation • V*· TVk· Vk+1· Vk • Contraction || TVk – V* ||1·d ||Vk – V*||1 • Unique Fixed Point TV*=V* ) if TVk¸ Vk, then Vk=V*.

  7. Convergence • Value Iteration Tk V0! V* • Distributed Value Iteration If you choose every x2 X infinitely often, then Vk! V*. (Here, random choice of x, use concavity.) • Deepest Cut Pick xk to maximize Vk(x)-TVk(x) DC problem to solve Convergence again with continuity (caution on boundary of domain of V*)

  8. Details for Random Choice • Consider any x • Choose i and xi s.t. ||xi – x|| < ei • Suppose || r Vi || · K 8 i || Vk(x) – V*(x)|| · ||Vk(x)-Vk(xk)+Vk(xk)-V*(xk)+V*(xk)-V*(x)|| · 2 ek K + d ||Vk-1(xk)-V*(xk)|| · 2åiei K + dk ||V0(x0) – V*(x0)||

  9. Cutting Plane Algorithm Initialization: Construct V0(x)=maxu c0(x,u) + dEPx,u[V0(y)], where c0¸ c and c0 concave. V0 is assumed piecewise linear and equivalent to V0(x)=max {q |q· E0 x + e0}. k=0. Iteration: Sample xk2 X (in any way such that the probability of xk2 A is positive for any A½ X of positive measure) and solve TVk(xk) = maxu c(xk,u) + d EPxk,u[Vk(y)] where Vk(y) =max{q | q· El y + el,l=0,…,k.} Find supporting hyperplanes defined by Ek+1 and ek+1such that Ek+1 x + ek+1¸ TVk (x). kà k+1. Repeat.

  10. Specifying Algorithm Feasibility: Ax + Bu · b Transition: y=Fi u for some realization i with probability pi Iteration k Problem: TVk(xk) = maxu,q c(xk,u) + dåi piqi s.t. A xk + B u · b, - El(Fi u) - el + qi· 0, 8 i,l. From duality: TVk(xk) = infm,ll,i maxu,q c(xk,u) -m(Axk+Bu-b) + dåi (piqi+ålli,l(El(Fi u) + el - qi)) · maxu,q c(xk,u) -mk(Axk+Bu-b) + dåi (piqi+ålli,l,k(El(Fi u) +el - qi)) for optimal mk, li,l,k for xk · c(xk,uk) + r c(xk,uk)T (x-xk,-uk) -mk Ax +mk b +åi ( ålli,l,kel) Cuts: Ek+1= rx c(xk,uk)T -mk A ek+1 equal to the constant terms.

  11. Results • Sometimes convergence is fast

  12. Results • Sometimes convergence is slow

  13. Challenges • Constrain feasible region to obtain convergence results • Accelerate the DC search problem to find the deep cut • Accelerate overall algorithm using: • multiple simultaneous cuts? • nonlinear cuts? • bundles approach?

  14. Conclusions • Can formulate infinite-horizon investment problem in stochastic programming framework • Solution with cutting plane method • Convergence with some conditions • Results for trade-restricted assets significantly different from market assets with same risk characteristics

  15. Investment Problem • Determine asset allocation and consumption policy to maximize the expected discounted utility of spending • State and Action x=(cons, risky, wealth) u=(cons_new,risky_new) • Two asset classes • Risky asset, with lognormal return distribution • Riskfree asset, with given return rf • Power utility function • Consumption rate constrained to be non-decreasing cons_new ¸ cons

  16. Existing Research • Dybvig ’95* • Continuous-time approach • Solution Analysis • Consumption rate remains constant until wealth reaches a new maximum • The risky asset allocation a is proportional to w-c/rf, which is the excess of wealth over the perpetuity value of current consumption • a decreases as wealth decreases, approaching 0 as wealth approaches c/rf (which is in absence of risky investment sufficient to maintain consumption indefinitely). • Dybvig ’01 • Considered similar problem in which consumption rate can decrease but is penalized (soft constrained problem) * “Duesenberry's Ratcheting of Consumption: Optimal Dynamic Consumption and Investment Given Intolerance for any Decline in Standard of Living” Review of Economic Studies 62, 1995, 287-313.

  17. Objectives • Replicate Dybvig continuous time results using discrete time approach • Evaluate the effect of trading restrictions for certain asset classes (e.g., private equity) • Consider additional problem features • Transaction Costs • Multiple risky assets

  18. Results – Non-decreasing Consumption As number of time periods per year increases, solution converges to continuous time solution

  19. Results – Non-Decreasing Consumption with Transaction Costs

  20. Observations • Effect of Trading Restrictions • Continuously traded risky asset: 70% of portfolio for 4.2% payout rate • Quarterly traded risky asset: 32% of portfolio for same payout rate • Transaction Cost Effect • Small differences in overall portfolio allocations • Optimal mix depends on initial conditions

  21. Extensions • Soft constraint on decreasing consumption • Allow some decreases with some penalty • Lag on sales • Waiting period on sale of risky assets (e.g., 60-day period) • Multiple assets • Allocation bounds

  22. Approach • Application of typical stochastic programming approach complicated by infinite horizon • Initialization. • Define a valid constraint on Q(x) Requires problem knowledge. For optimal consumption problem, assume extremely high rate of consumption forever

  23. Approach (cont.) Expensive search over x, possible for the optimal consumption problem because of small number of variables • Iteration k

More Related