File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/COMST.2025.3527641
- Find via

Supplementary
-
Citations:
- Appears in Collections:
Article: Mobile Edge Intelligence for Large Language Models: A Contemporary Survey
| Title | Mobile Edge Intelligence for Large Language Models: A Contemporary Survey |
|---|---|
| Authors | |
| Issue Date | 9-Jan-2025 |
| Publisher | Institute of Electrical and Electronics Engineers |
| Citation | IEEE Communications Surveys and Tutorials, 2025 How to Cite? |
| Abstract | On-device large language models (LLMs), referring to running LLMs on edge devices, have raised considerable interest since they are more cost-effective, latency-efficient, and privacy-preserving compared with the cloud paradigm. Nonetheless, the performance of on-device LLMs is intrinsically constrained by resource limitations on edge devices. Sitting between cloud and on-device AI, mobile edge intelligence (MEI) presents a viable solution by provisioning AI capabilities at the edge of mobile networks. This article provides a contemporary survey on harnessing MEI for LLMs. We begin by illustrating several killer applications to demonstrate the urgent need for deploying LLMs at the network edge. Next, we present the preliminaries of LLMs and MEI, followed by resource-efficient LLM techniques. We then present an architectural overview of MEI for LLMs (MEI4LLM), outlining its core components and how it supports the deployment of LLMs. Subsequently, we delve into various aspects of MEI4LLM, extensively covering edge LLM caching and delivery, edge LLM training, and edge LLM inference. Finally, we identify future research opportunities. We hope this article inspires researchers in the field to leverage mobile edge computing to facilitate LLM deployment, thereby unleashing the potential of LLMs across various privacy-and delay-sensitive applications. |
| Persistent Identifier | http://hdl.handle.net/10722/359208 |
| ISSN | 2023 Impact Factor: 34.4 2023 SCImago Journal Rankings: 15.966 |
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Qu, Guanqiao | - |
| dc.contributor.author | Chen, Qiyuan | - |
| dc.contributor.author | Wei, Wei | - |
| dc.contributor.author | Lin, Zheng | - |
| dc.contributor.author | Chen, Xianhao | - |
| dc.contributor.author | Huang, Kaibin | - |
| dc.date.accessioned | 2025-08-23T00:30:38Z | - |
| dc.date.available | 2025-08-23T00:30:38Z | - |
| dc.date.issued | 2025-01-09 | - |
| dc.identifier.citation | IEEE Communications Surveys and Tutorials, 2025 | - |
| dc.identifier.issn | 1553-877X | - |
| dc.identifier.uri | http://hdl.handle.net/10722/359208 | - |
| dc.description.abstract | <p>On-device large language models (LLMs), referring to running LLMs on edge devices, have raised considerable interest since they are more cost-effective, latency-efficient, and privacy-preserving compared with the cloud paradigm. Nonetheless, the performance of on-device LLMs is intrinsically constrained by resource limitations on edge devices. Sitting between cloud and on-device AI, mobile edge intelligence (MEI) presents a viable solution by provisioning AI capabilities at the edge of mobile networks. This article provides a contemporary survey on harnessing MEI for LLMs. We begin by illustrating several killer applications to demonstrate the urgent need for deploying LLMs at the network edge. Next, we present the preliminaries of LLMs and MEI, followed by resource-efficient LLM techniques. We then present an architectural overview of MEI for LLMs (MEI4LLM), outlining its core components and how it supports the deployment of LLMs. Subsequently, we delve into various aspects of MEI4LLM, extensively covering edge LLM caching and delivery, edge LLM training, and edge LLM inference. Finally, we identify future research opportunities. We hope this article inspires researchers in the field to leverage mobile edge computing to facilitate LLM deployment, thereby unleashing the potential of LLMs across various privacy-and delay-sensitive applications.<br></p> | - |
| dc.language | eng | - |
| dc.publisher | Institute of Electrical and Electronics Engineers | - |
| dc.relation.ispartof | IEEE Communications Surveys and Tutorials | - |
| dc.title | Mobile Edge Intelligence for Large Language Models: A Contemporary Survey | - |
| dc.type | Article | - |
| dc.identifier.doi | 10.1109/COMST.2025.3527641 | - |
| dc.identifier.eissn | 1553-877X | - |
| dc.identifier.issnl | 1553-877X | - |
