With the proliferation of the Internet-of-things (IoT), there is an undeniable consensus that cellular LTE networks will have to support a dramatically larger number of uplink connections. This is true since most of the devices to be added incur machine-type communications which is dominantly upstream. Can current LTE network withstand this challenge? To answer this question, the joint performance of random access process and the uplink data transmission should be investigated. These two problems have been classically treated in the literature in a disjoint fashion. In this paper, they are jointly analyzed as an inseparable couple. To do that, a tandem queuing model is adopted whereby devices are represented as spatially interacting queues. The interaction between queues is governed by the mutual inter-cell and intra-cell interference. To that end, a joint stochastic geometry and queueing theory model is exploited to study this problem and a spatiotemporal analytical model is developed accordingly. Network stability and scalability are two prime performance criteria for performance assessment. In light of these two criteria, the developed model is poised to offer valuable insights into efficient access and resource allocation strategies.
|Original language||English (US)|
|Title of host publication||2017 IEEE International Conference on Communications (ICC)|
|Publisher||Institute of Electrical and Electronics Engineers (IEEE)|
|State||Published - Jul 31 2017|