Tribastone, Mirco and Ding, Jie and Gilmore, Stephen and Hillston, Jane Fluid rewards for a stochastic process algebra. IEEE Transactions on Software Engineering, 38 (4). pp. 861-874. ISSN 0098-5589 (2012)
Full text not available from this repository.Abstract
Reasoning about the performance of models of software systems typically entails the derivation of metrics such as throughput, utilization, and response time. If the model is a Markov chain, these are expressed as real functions of the chain, called reward models. The computational complexity of reward-based metrics is of the same order as the solution of the Markov chain, making the analysis infeasible when evaluating large-scale systems. In the context of the stochastic process algebra PEPA, the underlying continuous-time Markov chain has been shown to admit a deterministic (fluid) approximation as a solution of an ordinary differential equation, which effectively circumvents state-space explosion. This paper is concerned with approximating Markovian reward models for PEPA with fluid rewards, i.e., functions of the solution of the differential equation problem. It shows that (1) the Markovian reward models for typical metrics of performance enjoy asymptotic convergence to their fluid analogues, and that (2) via numerical tests, the approximation yields satisfactory accuracy in practice.
Item Type: | Article |
---|---|
Identification Number: | https://doi.org/10.1109/TSE.2011.81 |
Uncontrolled Keywords: | Markov processes; Modeling and prediction; ordinary differential equations |
Subjects: | Q Science > QA Mathematics > QA75 Electronic computers. Computer science |
Research Area: | Computer Science and Applications |
Depositing User: | Ms T. Iannizzi |
Date Deposited: | 11 Feb 2015 14:17 |
Last Modified: | 11 Feb 2015 14:17 |
URI: | http://eprints.imtlucca.it/id/eprint/2602 |
Actions (login required)
Edit Item |