Issue No. 04 - July-Aug. (2012 vol. 38)

ISSN: 0098-5589

pp: 861-874

DOI Bookmark: http://doi.ieeecomputersociety.org/10.1109/TSE.2011.81

Mirco Tribastone , Ludwig-Maximilians-Universität, München

Jie Ding , Yangzhou University, Yangzhou

Stephen Gilmore , Edinburgh University, Edinburgh

Jane Hillston , Edinburgh University, Edinburgh

ABSTRACT

Reasoning about the performance of models of software systems typically entails the derivation of metrics such as throughput, utilization, and response time. If the model is a Markov chain, these are expressed as real functions of the chain, called reward models. The computational complexity of reward-based metrics is of the same order as the solution of the Markov chain, making the analysis infeasible when evaluating large-scale systems. In the context of the stochastic process algebra PEPA, the underlying continuous-time Markov chain has been shown to admit a deterministic (fluid) approximation as a solution of an ordinary differential equation, which effectively circumvents state-space explosion. This paper is concerned with approximating Markovian reward models for PEPA with fluid rewards, i.e., functions of the solution of the differential equation problem. It shows that 1) the Markovian reward models for typical metrics of performance enjoy asymptotic convergence to their fluid analogues, and that 2) via numerical tests, the approximation yields satisfactory accuracy in practice.

INDEX TERMS

Modeling and prediction, ordinary differential equations, Markov processes

CITATION

Mirco Tribastone, Jie Ding, Stephen Gilmore, Jane Hillston, "Fluid Rewards for a Stochastic Process Algebra",

*IEEE Transactions on Software Engineering*, vol. 38, no. , pp. 861-874, July-Aug. 2012, doi:10.1109/TSE.2011.81SEARCH