Presentation is loading. Please wait.

Presentation is loading. Please wait.

A Markov Chain Model of Baseball Used as a project for an undergraduate Stochastic Modeling course Eric Kuennen Department of Mathematics University of.

Similar presentations


Presentation on theme: "A Markov Chain Model of Baseball Used as a project for an undergraduate Stochastic Modeling course Eric Kuennen Department of Mathematics University of."— Presentation transcript:

1 A Markov Chain Model of Baseball Used as a project for an undergraduate Stochastic Modeling course Eric Kuennen Department of Mathematics University of Wisconsin Oshkosh Presented at: Joint Mathematics Meetings Washington, D.C. January 6, 2009

2 Markov Chain Model for Baseball View an inning of baseball as a stochastic process with 25 possible states. There are 8 different arrangements of runners on the bases: (bases empty, runner on 1 st, runner on 2 nd, runner on 3 rd, runners on 1 st and 2 nd, runners on 1 st and 3 rd, runners on 2 nd and 3 rd, bases loaded) and three possibilities for the number of outs (0 outs, 1 out, 2 outs), for a total of 24 non-absorbing states. The 25 th state (3 outs) is an absorbing state for the inning. 0,01,02,03,012,013,023,0123,0 0,11,12,13,112,113,123,1123,1 0,21,22,23,212,213,223,2123,2

3 Transition Probabilities A Markov Chain is a stochastic process in which the next state depends only on the present state. In other words, future states are independent of past states. Let P ij denote the probability the next state is j, given the current state is i. Form the Transition Matrix T = [P ij ]. w = probability of a walk s = probability of a single d = probability of a double t = probability of a triple h = probability of a home run out = probability of an out

4 0,01,02,03,012,013,023,0123,0 0,0 hs+wdt 1,0 hd/2tw+s/2s/2d/2 2,0 h3s/4dtws/4 3,0 hsdtw 12,0 hd/2ts/4s/2d/2w+s/4 13,0 hd/2ts/2 d/2w 23,0 hs/2dt w 123,0 hd/2ts/2 d/2w

5 Transition Matrix

6 Run Matrix 0,01,02,03,012,013,023,0123,0 0,0 1 1, , , , , , ,

7 Methods of Analysis Theoretical Calculations with Maple Expected Run Values for each state Steady State Probability Vector Expected Value of a given play in a given state or in general

8 Expected Run Values Let v i be the expected number of runs scored starting in state i Students use Maple’s linear algebra package to solve for the vector v

9 Expected Run Values w =.094s =.157d =.049 t =.005h =.029out =.661 From 2005 MLB: 0, , , , , , , , , , , , , , , , , , , , , , , ,2 0.84

10 Sacrifice Bunting 0, , , , , , , , , , , , , , , , , , , , , , , , Is it ever advantageous to sacrifice bunt?

11 Stealing Bases 0, , , , , , , , , , , , , , , , , , , , , , , , How successful does a base-stealer need to be on average in order for it to be worth-while to attempt to steal second base with a runner on first and no outs?

12 Methods of Analysis Experimental Simulations with Minitab Students write a Minitab macro that uses a random number generator to simulate the step by step evolution of the Markov Chain Large-scale simulations are used to estimate Expected Run Values and perform situational strategy analyses

13 Two Simulated Innings First Inning 1. Single 2. Out 3. Double 4. Single 5. Out 6. Single 7. Out Second Inning 8. Single 9. Homerun 10. Out 11. Out 12. Single 13. Out

14 Sacrificing with the game on the line Mean number of runs scored: Probability of scoring at least one run: Mean number of runs scored: Probability of scoring at least one run: In the ninth inning, your team needs one run to win or tie. Suppose the first batter reaches first. Should you bunt?

15 Reference Sokol, J.S. (2004) “An Intuitive Markov Chain Lesson From Baseball,” Informs Transactions on Education. 5 pp

16 Please contact me for: Sample Maple Worksheet Sample Minitab Macro Project Assignment Handout


Download ppt "A Markov Chain Model of Baseball Used as a project for an undergraduate Stochastic Modeling course Eric Kuennen Department of Mathematics University of."

Similar presentations


Ads by Google