dc.contributor.author |
Sarji, Imad Elias, |
dc.date |
2013 |
dc.date.accessioned |
2015-02-03T09:49:59Z |
dc.date.available |
2015-02-03T09:49:59Z |
dc.date.issued |
2013 |
dc.date.submitted |
2013 |
dc.identifier.other |
b17910869 |
dc.identifier.uri |
http://hdl.handle.net/10938/9934 |
dc.description |
Thesis (M.E.)-- American University of Beirut, Department of Electrical and Computer Engineeering, 2013. |
dc.description |
Advisor : Dr. Ayman Kayssi, Professor, Electrical and Computer Engineering ; Co-Advisor : Dr. Ali Chehab, Associate Professor, Electrical and Computer Engineering ; Committee Member : Dr. Imad Elhajj, Associate Professor, Electrical and Computer Engineering. |
dc.description |
Includes bibliographical references (leaves 94-98) |
dc.description.abstract |
The world is rapidly moving towards the convergence of communication, computing and IP-based networks. With the smartphones invasion and the massive amount of services and applications available, The Long Term Evolution (LTE) is seen to be the key enabler for delivering the fourth generation of mobile broadband. Furthermore, LTE is huge step toward this convergence since it is the first pure IP-based cellular technology. As such, LTE networks are expected to achieve wide scale adoption worldwide. However, along with the proliferation of powerful smartphones and the diverse IP-based mobile applications, enormous amounts of data and signaling traffic are injected into the mobile network. Thus, mobile operators face additional burden in order to protect their networks, maintain good quality of service, and avoid any possible bottlenecks or node failures. This work sakes to study the effects of smartphone applications on the LTE core network. By profiling some of the most popular smartphone’s applications using Wireshark sniffer, and then inject the profiled data into Opnet simulator. Furthermore, in order to achieve less signaling overheads and more efficient battery consumption at the mobile device, we have proposed an optimization algorithm to decide when it is optimal for a UE to transit between LTE power states. This decision is based on probabilistic estimation using a traffic pattern model generated from an auto-learning algorithm running on the UE side. After experimentally evaluating the proposed algorithm, a decrease of 10 to 30percent was achieved. |
dc.format.extent |
xiv, 98 leaves : colored illustrations ; 30 cm |
dc.language.iso |
eng |
dc.relation.ispartof |
Theses, Dissertations, and Projects |
dc.subject.classification |
ET:005916 AUBNO |
dc.subject.lcsh |
Long-Term Evolution (Telecommunications) |
dc.subject.lcsh |
Signal theory (Telecommunication) |
dc.subject.lcsh |
Broadband communication systems. |
dc.subject.lcsh |
Smartphones. |
dc.subject.lcsh |
Cell phone systems. |
dc.subject.lcsh |
Mobile communication systems. |
dc.title |
The smartphone's signaling effect on LTE core network - |
dc.type |
Thesis |
dc.contributor.department |
American University of Beirut. Faculty of Engineering and Architecture. Department of Electrical and Computer Engineeering. |