A Power-Aware Reinforcement Learning Technique for Memory Allocation in Real-time Embedded Systems

dc.contributor.advisorPahlevani, Majid
dc.contributor.advisorFar, Behrouz Homayoun
dc.contributor.authorKarimi Fatemi, Masoud
dc.contributor.committeememberSanati-Nezhad, Amir
dc.contributor.committeememberMoshirpour, Mohammad
dc.date2021-02
dc.date.accessioned2021-01-06T22:44:33Z
dc.date.available2021-01-06T22:44:33Z
dc.date.issued2020-12-22
dc.description.abstractEmbedded systems are ubiquitous in today's world. They are used in a vast number of applications, from medical devices to spacecraft. Two of the main characteristics of such systems are real-time constraints and the lack of reliable energy sources. As cache memories negatively contribute to these two challenges, embedded systems have adopted a new concept called scratch-pad memories (SPMs). To further reduce power consumption, hybrid SPMs composed of Static RAMs (SRAMs) and non-volatile memories (NVMs) have been introduced. Compared to SRAMs, NVMs offer low leakage power and high density while suffering from high energy consumption during write activities. As such, numerous studies have been done on performing memory allocation in hybrid architectures. As memory allocation is an NP-complete problem, traditional approaches fall short due to their timing and memory complexities. Therefore, previous studies have offered alternative techniques to resolve this problem in significantly shorter times. In this work, I propose a reinforcement learning technique to tackle the memory allocation problem in a hybrid architecture. As the results showed, the proposed technique can provide a near-optimal solution - only 5.6% less efficient on average - in a substantially shorter time. To evaluate the algorithm's effectiveness, I compared the results with a greedy technique and a Genetic Algorithm (GA). The results showed that the proposed technique could reduce the average energy consumption by 35.84% and 5.1% compared to the greedy technique and GA, respectively. While the improvements gained by the proposed technique was negligible compared to the GA, it could outperform the GA in simulation time once the learning phase was complete. Particularly, the experiments on 15 soft real-time task frames showed a 35.1% reduction in the average simulation time.en_US
dc.identifier.citationKarimi Fatemi, M. (2020). A Power-Aware Reinforcement Learning Technique for Memory Allocation in Real-time Embedded Systems (Master's thesis, University of Calgary, Calgary, Canada). Retrieved from https://prism.ucalgary.ca.en_US
dc.identifier.doihttp://dx.doi.org/10.11575/PRISM/38530
dc.identifier.urihttp://hdl.handle.net/1880/112938
dc.language.isoengen_US
dc.publisher.facultySchulich School of Engineeringen_US
dc.publisher.institutionUniversity of Calgaryen
dc.rightsUniversity of Calgary graduate students retain copyright ownership and moral rights for their thesis. You may use this material in any way that is permitted by the Copyright Act or through licensing that has been assigned to the document. For uses that are not allowable under copyright legislation or licensing, you are required to seek permission.en_US
dc.subjectReinforcement Learningen_US
dc.subjectEmbedded Systemsen_US
dc.subjectMemory Allocationen_US
dc.subjectLow Poweren_US
dc.subjectReal-timeen_US
dc.subject.classificationArtificial Intelligenceen_US
dc.subject.classificationComputer Scienceen_US
dc.titleA Power-Aware Reinforcement Learning Technique for Memory Allocation in Real-time Embedded Systemsen_US
dc.typemaster thesisen_US
thesis.degree.disciplineEngineering – Electrical & Computeren_US
thesis.degree.grantorUniversity of Calgaryen_US
thesis.degree.nameMaster of Science (MSc)en_US
ucalgary.item.requestcopyfalseen_US
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
ucalgary_2020_karimi-fatemi_masoud.pdf
Size:
1.95 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.62 KB
Format:
Item-specific license agreed upon to submission
Description: