An analysis of double Q-learning-based energy management strategies for TEG-powered IoT devices

Loading...
Thumbnail Image

Downloads

8

Date issued

Journal Title

Journal ISSN

Volume Title

Publisher

IEEE

Location

Signature

License

Abstract

The study presents a self-learning controller for managing the energy in an Internet of Things (IoT) device pow ered by energy harvested from a thermoelectric generator (TEG). The device’s controller is based on a double Q-learning (DQL) method; the hardware incorporates a TEG energy harvesting subsystem with a dc/dc converter, a load module with a microcon troller, and a LoRaWAN communications interface. The model is controlled according to adaptive measurements and transmis sion periods. The controller’s reward policy evaluates the level of charge available to the device. The controller applies and evaluates various learning parameters and reduces the learning rate over time. Using four years of historical soil temperature data in an experimental simulation of several controller config urations, the DQL controller demonstrated correct operation, a low learning rate, and high cumulative rewards. The best energy management controller operated with a completed cycle and missed cycle ratio of 98.5%. The novelty of the presented approach is discussed in relation to state-of-the-art methods in adaptive ability, learning processes, and practical applications of the device.

Description

Subject(s)

energy harvesting, energy management, Internet of Things (IoT), reinforcement learning, thermoelectric generator (TEG)

Citation

IEEE Internet of Things Journal. 2023, vol. 10, issue 21, p. 18919-18929.