The Community for Technology Leaders
Green Image
Issue No. 04 - July-Aug. (2012 vol. 38)
ISSN: 0098-5589
pp: 861-874
Mirco Tribastone , Ludwig-Maximilians-Universität, München
Jie Ding , Yangzhou University, Yangzhou
Stephen Gilmore , Edinburgh University, Edinburgh
Jane Hillston , Edinburgh University, Edinburgh
Reasoning about the performance of models of software systems typically entails the derivation of metrics such as throughput, utilization, and response time. If the model is a Markov chain, these are expressed as real functions of the chain, called reward models. The computational complexity of reward-based metrics is of the same order as the solution of the Markov chain, making the analysis infeasible when evaluating large-scale systems. In the context of the stochastic process algebra PEPA, the underlying continuous-time Markov chain has been shown to admit a deterministic (fluid) approximation as a solution of an ordinary differential equation, which effectively circumvents state-space explosion. This paper is concerned with approximating Markovian reward models for PEPA with fluid rewards, i.e., functions of the solution of the differential equation problem. It shows that 1) the Markovian reward models for typical metrics of performance enjoy asymptotic convergence to their fluid analogues, and that 2) via numerical tests, the approximation yields satisfactory accuracy in practice.
Modeling and prediction, ordinary differential equations, Markov processes

M. Tribastone, J. Ding, S. Gilmore and J. Hillston, "Fluid Rewards for a Stochastic Process Algebra," in IEEE Transactions on Software Engineering, vol. 38, no. , pp. 861-874, 2012.
1690 ms
(Ver 3.3 (11022016))