TY - GEN
T1 - Analysis of cache tuner architectural layouts for multicore embedded systems
AU - Adegbija, Tosiron
AU - Gordon-Ross, Ann
AU - Rawlins, Marisha
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2015/1/20
Y1 - 2015/1/20
N2 - Due to the memory hierarchy's large contribution to a microprocessor's total power, cache tuning is an ideal method for optimizing overall power consumption in embedded systems. Since most embedded systems are power and area constrained, the hardware and/or software that orchestrate cache tuning - the cache tuner - must not impose significant power and area overhead. Furthermore, as embedded systems increasingly trend towards multicore, inter-core data sharing, communication, and synchronization impose additional cache tuner design complexity, necessitating cross-core cache tuning coordination. In order to minimize cache tuner overhead, cache tuner design must consider these overheads and scalability. Whereas prior work proposes low-overhead cache tuners, scalability to multicore systems requires additional considerations. In this work, we present a low-overhead, scalable cache tuner and extensively evaluate various cache tuner design tradeoffs with respect to power and area for constrained multicore embedded systems. Based on our analysis, we formulate valuable insights and designer-assisted guidelines for modeling scalable and efficient cache tuners that best achieve optimization goals while maintaining power and area constraints.
AB - Due to the memory hierarchy's large contribution to a microprocessor's total power, cache tuning is an ideal method for optimizing overall power consumption in embedded systems. Since most embedded systems are power and area constrained, the hardware and/or software that orchestrate cache tuning - the cache tuner - must not impose significant power and area overhead. Furthermore, as embedded systems increasingly trend towards multicore, inter-core data sharing, communication, and synchronization impose additional cache tuner design complexity, necessitating cross-core cache tuning coordination. In order to minimize cache tuner overhead, cache tuner design must consider these overheads and scalability. Whereas prior work proposes low-overhead cache tuners, scalability to multicore systems requires additional considerations. In this work, we present a low-overhead, scalable cache tuner and extensively evaluate various cache tuner design tradeoffs with respect to power and area for constrained multicore embedded systems. Based on our analysis, we formulate valuable insights and designer-assisted guidelines for modeling scalable and efficient cache tuners that best achieve optimization goals while maintaining power and area constraints.
KW - cache memories
KW - cache tuning
KW - configurable hardware
KW - low-power design
KW - multicore embedded systems
UR - http://www.scopus.com/inward/record.url?scp=84923169527&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84923169527&partnerID=8YFLogxK
U2 - 10.1109/PCCC.2014.7017091
DO - 10.1109/PCCC.2014.7017091
M3 - Conference contribution
AN - SCOPUS:84923169527
T3 - 2014 IEEE 33rd International Performance Computing and Communications Conference, IPCCC 2014
BT - 2014 IEEE 33rd International Performance Computing and Communications Conference, IPCCC 2014
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 33rd IEEE International Performance Computing and Communications Conference, IPCCC 2014
Y2 - 5 December 2014 through 7 December 2014
ER -