TY - JOUR AV - none SP - 861 EP - 874 TI - Fluid rewards for a stochastic process algebra N2 - Reasoning about the performance of models of software systems typically entails the derivation of metrics such as throughput, utilization, and response time. If the model is a Markov chain, these are expressed as real functions of the chain, called reward models. The computational complexity of reward-based metrics is of the same order as the solution of the Markov chain, making the analysis infeasible when evaluating large-scale systems. In the context of the stochastic process algebra PEPA, the underlying continuous-time Markov chain has been shown to admit a deterministic (fluid) approximation as a solution of an ordinary differential equation, which effectively circumvents state-space explosion. This paper is concerned with approximating Markovian reward models for PEPA with fluid rewards, i.e., functions of the solution of the differential equation problem. It shows that (1) the Markovian reward models for typical metrics of performance enjoy asymptotic convergence to their fluid analogues, and that (2) via numerical tests, the approximation yields satisfactory accuracy in practice. IS - 4 VL - 38 A1 - Tribastone, Mirco A1 - Ding, Jie A1 - Gilmore, Stephen A1 - Hillston, Jane ID - eprints2602 JF - IEEE Transactions on Software Engineering SN - 0098-5589 KW - Markov processes; Modeling and prediction; ordinary differential equations PB - IEEE Computer Society UR - http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=5975178&isnumber=6249691 Y1 - 2012/07// ER -