Maximizing resource utilization by slicing of superscalar architecture
Superscalar architectural techniques increase instruction throughput by increasing resources and using complex control units that perform various functions to minimize stalls and to ensure a continuous feed of instructions to the execution units. This work proposes a dynamic scheme to increase efficiency of execution (throughput) by a methodology called block slicing. This takes advantage of instruction level parallelism (ILP) available in programs without increasing the number of execution units. Implementation of this concept in a wide, superscalar pipelined architecture introduces nominal additional hardware and delay, while offering power and area advantages. We present the design of the hardware required for the implementation of the proposed scheme and evaluate it for the metrics of speed-up, throughput and efficiency.
Bandwidth; Computer architecture; Embedded system; Field programmable gate arrays; Hardware; Logic design; Parallel processing; Processor scheduling; Resource management; Throughput
Electrical and Computer Engineering | Electrical and Electronics | Electronic Devices and Semiconductor Manufacturing | Signal Processing | VLSI and Circuits, Embedded and Hardware Systems
Use Find in Your Library, contact the author, or interlibrary loan to garner a copy of the item. Publisher policy does not allow archiving the final published version. If a post-print (author's peer-reviewed manuscript) is allowed and available, or publisher policy changes, the item will be deposited.
Maximizing resource utilization by slicing of superscalar architecture.
11th EUROMICRO Conference on Digital System Design Architectures, Methods and Tools, DSD 2008