Hyper-Threading Technology - PowerPoint PPT Presentation

1 / 27
About This Presentation
Title:

Hyper-Threading Technology

Description:

Hyper-Threading Technology Presented By Nagarajender Rao Katoori To Enhance Performance- Increase in clock rate Involves reducing clock cycle time Can increase the ... – PowerPoint PPT presentation

Number of Views:476
Avg rating:3.0/5.0
Slides: 28
Provided by: sitIitkg
Category:

less

Transcript and Presenter's Notes

Title: Hyper-Threading Technology


1
Hyper-Threading Technology
  • Presented By
  • Nagarajender Rao Katoori

2
  Introduction
  • To Enhance Performance- 
  • Increase in clock rate
  •  
  • Involves reducing clock cycle time
  • Can increase the performance by increasing
    number of instructions finishing per second
  • H/w limitations limit this feature
  • Cache hierarchies
  • Having frequently used data on the processor
    caches reduces average accesses time

3
  • Pipelining
  • Implementation Technique whereby multiple
    instructions are overlapped in execution
  • Limited by the dependencies between instructions
  • Effected by stalls and effective CPI is greater
    than 1
  • Instruction Level Parallelism 
  • It refers to techniques to increase the number
    of instructions executed in each clock cycle.
  • Exists whenever the machine instructions that
    make up a program are insensitive to the order in
    which they are executed if dependencies does not
    exist, they may be executed.

4
  • Thread level parallelism
  • Chip Multi Processing
  • Two processors, each with full set of execution
    and architectural resources, reside on a single
    die.
  • Time Slice Multi Threading
  • single processor to execute multiple threads by
    switching between them
  • Switch on Event Multi Threading
  • switch threads on long latency events such as
    cache misses

5
  • Thread level parallelism (cont..)
  • Simultaneous Multi Threading
  • Multiple threads can execute on a single
    processor without switching.
  • The threads execute simultaneously and make much
    better use of the resources.
  • It maximizes the performance vs. transistor count
    and power consumption.

6
  •    Hyper-Threading Technology
  • Hyper-Threading Technology brings the
    simultaneous multi-threading approach to the
    Intel architecture.
  •   Hyper-Threading Technology makes a single
    physical processor appear as two or more logical
    processors
  •  Hyper-Threading Technology first invented by
    Intel Corp.
  •  Hyper-Threading Technology provides
    thread-level-parallelism (TLP) on each processor
    resulting in increased utilization of processor
    and execution resources.
  • Each logical processor maintain one copy of the
    architecture state

7
Hyper-Threading Technology Architecture
Processor Execution Resources
Processor Execution Resources
Arch State
Arch State
Arch State
Processor with out Hyper-Threading Technology
Processor with Hyper-Threading Technology
Ref Intel Technology Journal, Volume 06 Issue
01, February 14, 2002
8
Following resources are duplicated to support
Hyper-Threading Technology
  • Register Alias Tables
  • Next-Instruction Pointer
  • Instruction Streaming Buffers and Trace Cache
    Fill Buffers
  • Instruction Translation Look-aside Buffer

9
Figure Intel Xeon processor pipeline
10
  • Sharing of Resources
  •  
  • Ø     Major Sharing Schemes are-
  • o Partition
  • o Threshold
  • o Full Sharing
  • Partition
  •  
  • Ø     Each logical processor uses half the
    resources
  • Ø     Simple and low in complexity
  • Ø     Ensures fairness and progress
  • Ø     Good for major pipeline queues

11
Partitioned Queue Example
  • Yellow thread It is faster thread
  • Green thread It is slower thread

12
(No Transcript)
13
(No Transcript)
14
(No Transcript)
15
Partitioned Queue Example
  • Partitioning resource ensures fairness and
  • ensures progress for both logical processors.

16
  •  Threshold
  •  Ø     Puts a threshold on number of resource
    entries a logical processor can use.
  • Ø     Limits maximum resource usage
  • Ø     For small structures where resource
    utilization in burst and time of utilization is
    short, uniform and predictable
  •  
  • Ø     Eg- Processor Scheduler

17
  •  Full Sharing  
  • Ø     Most flexible mechanism for resource
    sharing, do not limit the maximum uses for
    resource usage for a logical processor
  • Ø     Good for large structures in which working
    set sizes are variable and there is no fear of
    starvation
  • Ø     Eg All Processor caches are shared
  • o       Some applications benefit from a shared
    cache because they share code and data,
    minimizing redundant data in the caches

18
Netburst Microarchitectures execution pipeline
19
SINGLE-TASK AND MULTI-TASK MODES
  • Two modes of operations
  • single-task (ST)
  • multi-task (MT).
  • MT-mode- There are two active logical processors
    and some of the resources are partitioned.
  • There are two flavors of ST-mode single-task
    logical processor 0 (ST0) and single-task logical
    processor 1 (ST1).
  • In ST0- or ST1-mode, only one logical processor
    is active, and resources that were partitioned in
    MT-mode are re-combined to give the single active
    logical processor use of all of the resources

20
SINGLE-TASK AND MULTI-TASKMODES
21
  • HALT instruction that stops processor execution.
  • On a processor with Hyper-Threading Technology,
    executing HALT transition the processor from
    MT-mode to ST0- or ST1-mode, depending on which
    logical processor executed the HALT.
  • In ST0- or ST1-modes, an interrupt sent to the
    halted logical processor would cause a transition
    to MT-mode.

22
OPERATING SYSTEM
  • For best performance, the operating system should
    implement two optimizations.
  • The first is to use the HALT instruction if one
    logical processor is active and the other is not.
    HALT will allow the processor to transition MT
    mode to either the ST0- or ST1-mode.
  • The second optimization is in scheduling software
    threads to logical processors. The operating
    system should schedule threads to logical
    processors on different physical processors
    before scheduling two threads to the same
    physical processor.

23
  • Business Benefits of Hyper-Threading
  • Technology
  • Higher transaction rates for e-Businesses
  • Improved reaction and response times for
    end-users and customers.
  • Increased number of users that a server system
    can support
  • Handle increased server workloads
  • Compatibility with existing server applications
    and operating systems

24
Performance increases from Hyper-Threading
Technology on an OLTP workload
Web server benchmark performance
25
Conclusion
  • Intels Hyper-Threading Technology brings the
    concept of simultaneous multi-threading to the
    Intel Architecture.
  • It will become increasingly important going
    forward as it adds a new technique for obtaining
    additional performance for lower transistor and
    power costs.
  • The goal was to implement the technology at
    minimum cost while ensuring forward progress on
    logical processors, even if the other is stalled,
    and to deliver full performance even when there
    is only one active logical processor.

26
References
  • HYPER-THREADING TECHNOLOGY ARCHITECTURE AND
    MICROARCHITECTURE by Deborah T. Marr, Frank
    Binns, David L. Hill, Glenn Hinton,David A.
    Koufaty, J. Alan Miller, Michael Upton, intel
    Technology Journal, Volume 06 Issue 01, Published
    February 14, 2002. Pages 4 15.
  • HYPERTHREADING TECHNOLOGY IN THE NETBURST
    MICROARCHITECTURE by David Koufaty,Deborah T.
    Marr, IEEE Micro, Vol. 23, Issue 2, MarchApril
    2003. Pages 56 65.
  • http//cache-www.intel.com/cd/00/00/22/09/220943_
    220943.pdf
  • http//www.cs.washington.edu/research/smt/papers/t
    lp2ilp.final.pdf
  • http//mos.stanford.edu/papers/mj_thesis.pdf

27
  • Thank you
Write a Comment
User Comments (0)
About PowerShow.com