earticle

논문검색

Communication

Energy-Efficient Offloading with Distributed Reinforcement Learning for Edge Computing in Home Networks

초록

영어

This paper introduces a decision-making framework for offloading tasks in home network environments, utilizing Distributed Reinforcement Learning (DRL). The proposed scheme optimizes energy efficiency while maintaining system reliability within a lightweight edge computing setup. Effective resource management has become crucial with the increasing prevalence of intelligent devices. Conventional methods, including on-device processing and offloading to edge or cloud systems, need help to balance energy conservation, response time, and dependability. To tackle these issues, we propose a DRL-based scheme that allows flexible and enhanced decision-making regarding offloading. Simulation results demonstrate that the proposed method outperforms the baseline approaches in reducing energy consumption and latency while maintaining a higher success rate. These findings highlight the potential of the proposed scheme for efficient resource management in home networks and broader IoT environments.

목차

Abstract
1. Introduction
2. Related Work
2.1 Edge Computing and Offloading Techniques
2.2 Offloading Decisions Using Reinforcement Learning
2.3 Lightweight Edge Computing in Home Networks
3. System Model
3.1 Energy Consumption and Delay
3.2 Problem Statement
4. Proposed Scheme
4.1 DRL-based Offloading Decision
4.2 Training Process of the Proposed Scheme
5. Performance Evaluation
6. Conclusion
Acknowledgment
References

저자정보

  • Ducsun Lim Post-Doc, Department of Computer Software, Hanyang University, Korea
  • Dongkyun Lim Professor, Department of Computer Science Engineering, Hanyang Cyber University, Korea

참고문헌

자료제공 : 네이버학술정보

    함께 이용한 논문

      ※ 원문제공기관과의 협약기간이 종료되어 열람이 제한될 수 있습니다.

      0개의 논문이 장바구니에 담겼습니다.