low power memory design

Upload: arun-singh

Post on 03-Apr-2018

223 views

Category:

Documents


0 download

TRANSCRIPT

  • 7/29/2019 low power memory design

    1/29

    1

    Table of ContentsLIST OF FIGURES ............................................................................................................................................ 2

    LIST OF TABLES .............................................................................................................................................. 3

    CHAPTER-1 .................................................................................................................................................... 4

    1.1 INTRODUCTION ................................................................................................................................... 4

    1.2 MOTIVATION ....................................................................................................................................... 5

    1.3 DRAM .................................................................................................................................................. 6

    1.3.1 Operations to read a data bit from a DRAM storage cell ............................................................ 6

    1.3.2 To write to memory ..................................................................................................................... 8

    1.3.3 Refresh rate .................................................................................................................................. 9

    1.4 SRAM ................................................................................................................................................. 10

    1.4.1 Design ......................................................................................................................................... 10

    1.4.2 SRAM Operation ........................................................................................................................ 11

    CHAPTER-2 .................................................................................................................................................. 13

    2.1 SOURCES AND REDUCTION OF POWER DISSIPATION IN MEMORY SUBSYSTEM ..................... 13

    2.1.2 Multi-data-bit Configuration Chip .............................................................................................. 13

    2.1.2 Small Package ............................................................................................................................. 15

    2.1.3 Low-voltage Data-bus Interface ................................................................................................. 162.2 SOURCES OF POWER DISSIPATION IN DRAM AND SRAM ................................................................. 18

    2.2.1 Active Power Sources ................................................................................................................. 18

    2.2.2 Data Retention Power Sources .................................................................................................. 21

    CHAPTER-3 .................................................................................................................................................. 23

    3.1 Low Power DRAM Circuits ................................................................................................................ 23

    3.1.1 Active power reduction .............................................................................................................. 23

    3.1.2 Operating Voltage Reduction ..................................................................................................... 25

    3.1.3 DC current reduction ................................................................................................................. 26

    3.1.4 Data Retention Power Reduction .............................................................................................. 27

    3.1.5 Refresh time extension .............................................................................................................. 27

    CONCLUSION ............................................................................................................................................... 28

    REFERENCES ................................................................................................................................................ 29

  • 7/29/2019 low power memory design

    2/29

    2

    LIST OF FIGURES

    Fig-1.2.1 Principle of operation of DRAM read ,for simple 4 by 4 array8

    Fig-1.3.1-A six transistor CMOS SRAM cell.11

    Fig-1.3.2-Four transistors SRAM cell11

    Fig-2.1.1-subsystem using multi-bit data configuration chip..14

    Fig-2.1.2-advancement in packages.16

    Fig-2.1.3-Typical TTL interface.17

    Fig-2.2.1-Sources of Power dissipation in a RAM chip19

    Fig-3.1.1-Low Power circuit advancement of DRAMs over the last decade.23

    Fig-3.1.2-Multi divided data line architecture with SA , shared I/O and shared y decoder.24

    Fig-3.1.3-Voltage down converter .26

    Fig-3.1.4-Pulse operation of column signal path circuitry.27

  • 7/29/2019 low power memory design

    3/29

    3

    LIST OF TABLES

    Table-2.1.1-5V DRAM subsystems with 32 bit data bit bus line.15

    Table-2.1.2-5V DRAM chip a package advancement..15

    Table-2.2.1-comparison of determinants of active cell current between DRAMs and SRAMs..21

  • 7/29/2019 low power memory design

    4/29

    4

    CHAPTER-1

    1.1 INTRODUCTION

    Much of the research efforts of the past years in the area of digital electronics

    has been directed towards increasing the speed of digital systems. Recently , the requirement of

    portability and the moderate improvement in battery performance indicate that the power

    dissipation is one of the most critical design parameters . The three most widely accepted metrics

    to measure the quality of a circuit or to compare various circuit styles are area, delay and power

    dissipation. Portability imposes a strict limitation on power dissipation while still demands high

    computational speeds. Hence, in recent VLSI systems the power-delay product becomes the most

    essential metric of performance.

    Energy-efficiency is one of the most required features for modern electronic

    systems designed for high-performance and/or portable applications. In one hand, the ever

    increasing market segment of portable electronic devices demands the availability of low-power

    building blocks that enable the implementation of long-lasting battery-operated systems. On the

    other hand, the general trend of increasing operating frequencies and circuit complexity, in order

    to cope with the throughput needed in modern high-performance processing applications,

    requires the design of very high-speed circuits. The power-delay product (PDP) metric relates theamount of energy spent during the realization of a determined task, and stands as the more fair

    performance metric when comparing optimizations of a module designed and tested using

    different technologies, operating frequencies, and scenarios .

    In any event , RAM chip power is a primary concern of the subsystem designer ,

    since it dominates subsystem power. Various low power circuit technologies concerning

    reductions in charging capacitance , operating voltage and static current have been developed .

    The ever decreasing supply voltage requirements could result in ultra low power and extended

    battery life. Thus the primary concern is for CMOS technology to realize high-speed with

    minimum power . In This report I will be discussing about the sources of power dissipation and

    the low power circuit designs of the DRAMs and SRAMs subsystems.

  • 7/29/2019 low power memory design

    5/29

    5

    1.2MOTIVATIONIn CMOS digital circuits there are two main contributors to power dissipation:

    dynamic and static. Static power dissipation is caused by leakage current drawn

    continuously from the supply. This is from the reverse bias leakage between diffusion

    regions and the substrate. Also, subthreshold currents can contribute to static power. In

    most of todays IC designs , static power dissipation is orders of magnitude below that of

    dynamic (or switching) power dissipation . ynamic power dissipation is a result of the

    charging and discharging of load capacitances in addition to switching transient currents.

    Dynamic power dissipation also includes short circuit currents; However, we can generally

    ignore this short circuit current compared to the current due to charging of capacitors.

    Dynamic power dissipation of CMOS circuits is given by-

    Pdyn = CLVdd2f

    where is the node transition activity factor, CL is the load capacitance, Vdd is the

    supply voltage, andf is the clock frequency. Thus, the minimization of power dissipation in

    CMOS digital circuits is achieved by reducing the components of this equation. Since the

    dynamic power has a quadratic dependence on the supply voltage Vdd ,it is easily seen

    that the lowest supply voltage is desirable. However, lowering the supply voltage induces

    a cost of increased gate delays. Thus we must consider both power reduction along with

    increased delay when looking at low power esign methodologies. Superior approaches are

    ones that provide more decrease in power for a given increase in delay.

  • 7/29/2019 low power memory design

    6/29

    6

    1.3 DRAM

    Dynamic random-access memory (DRAM) is a type ofrandom-access memory that

    stores each bit of data in a separate capacitor within an integrated circuit. The capacitor can beeither charged or discharged; these two states are taken to represent the two values of a bit,

    conventionally called 0 and 1. Since capacitors leak charge, the information eventually fades

    unless the capacitor charges refreshed periodically. Because of this refresh requirement, it is

    a dynamic memory as opposed to SRAM and other static memory . The main memory (the

    "RAM") in personal computers is dynamic RAM (DRAM). It is the RAM

    in laptop and workstation computers as well as some of the RAM ofvideo game consoles.

    The advantage of DRAM is its structural simplicity: only one transistor and a

    capacitor are required per bit, compared to four or six transistors in SRAM. This allows DRAM

    to reach very high densities. Unlike flash memory, DRAM is volatile memory, since it loses its

    data quickly when power is removed. The transistors and capacitors used are extremely small ;

    billions can fit on a single memory chip.

    1.3.1 Operations to read a data bit from a DRAM storage cell

    1. The sense amplifiers are disconnected.2. The bit-lines are precharged to exactly equal voltages that are in between high and low

    logic levels. The bit-lines are physically symmetrical to keep the capacitance equal, and

    therefore the voltages are equal.

    3. The precharge circuit is switched off. Because the bit-lines are relatively long, they haveenough capacitance to maintain the precharged voltage for a brief time. This is an

    example ofdynamic logic.

    4. The desired row's word-line is then driven high to connect a cell's storage capacitor to itsbit-line. This causes the transistor to conduct, transferring charge between the storage

    cell and the connected bit-line. If the storage cell's capacitor is discharged, it will greatly

    decrease the voltage on the bit-line as the precharge is used to charge the storage

    http://en.wikipedia.org/wiki/Random-access_memoryhttp://en.wikipedia.org/wiki/Bithttp://en.wikipedia.org/wiki/Capacitorhttp://en.wikipedia.org/wiki/Integrated_circuithttp://en.wikipedia.org/wiki/Memory_refreshhttp://en.wikipedia.org/wiki/Static_random_access_memoryhttp://en.wikipedia.org/wiki/Laptophttp://en.wikipedia.org/wiki/Workstationhttp://en.wikipedia.org/wiki/Video_gamehttp://en.wikipedia.org/wiki/Transistorhttp://en.wikipedia.org/wiki/Computer_storage_densityhttp://en.wikipedia.org/wiki/Flash_memoryhttp://en.wikipedia.org/wiki/Volatile_memoryhttp://en.wikipedia.org/wiki/Capacitancehttp://en.wikipedia.org/wiki/Dynamic_logic_(digital_logic)http://en.wikipedia.org/wiki/Electric_chargehttp://en.wikipedia.org/wiki/Electric_chargehttp://en.wikipedia.org/wiki/Dynamic_logic_(digital_logic)http://en.wikipedia.org/wiki/Capacitancehttp://en.wikipedia.org/wiki/Volatile_memoryhttp://en.wikipedia.org/wiki/Flash_memoryhttp://en.wikipedia.org/wiki/Computer_storage_densityhttp://en.wikipedia.org/wiki/Transistorhttp://en.wikipedia.org/wiki/Video_gamehttp://en.wikipedia.org/wiki/Workstationhttp://en.wikipedia.org/wiki/Laptophttp://en.wikipedia.org/wiki/Static_random_access_memoryhttp://en.wikipedia.org/wiki/Memory_refreshhttp://en.wikipedia.org/wiki/Integrated_circuithttp://en.wikipedia.org/wiki/Capacitorhttp://en.wikipedia.org/wiki/Bithttp://en.wikipedia.org/wiki/Random-access_memory
  • 7/29/2019 low power memory design

    7/29

    7

    capacitor. If the storage cell is charged, the bit-line's voltage only decreases very slightly.

    This occurs because of the high capacitance of the storage cell capacitor compared to the

    capacitance of the bit-line, thus allowing the storage cell to determine the charge level on

    the bit-line.

    5. The sense amplifiers are connected to the bit-lines. Positive feedback then occurs fromthe cross-connected inverters, thereby amplifying the small voltage difference between

    the odd and even row bit-lines of a particular column until one bit line is fully at the

    lowest voltage and the other is at the maximum high voltage. Once this has happened,

    the row is "open" (the desired cell data is available).

    6. All storage cells in the open row are sensed simultaneously, and the sense amplifieroutputs latched. A column address then selects which latch bit to connect to the external

    data bus. Reads of different columns in the same row can be performed without a row

    opening delay because, for the open row, all data has already been sensed and latched.

    7. While reading of columns in an open row is occurring, current is flowing back up the bit-lines from the output of the sense amplifiers and recharging the storage cells. This

    reinforces (i.e. "refreshes") the charge in the storage cell by increasing the voltage in the

    storage capacitor if it was charged to begin with, or by keeping it discharged if it was

    empty. Note that due to the length of the bit-lines there is a fairly long propagation delay

    for the charge to be transferred back to the cell's capacitor. This takes significant timepast the end of sense amplification, and thus overlaps with one or more column reads.

    8. When done with reading all the columns in the current open row, the word-line isswitched off to disconnect the storage cell capacitors (the row is "closed") from the bit-

    lines. The sense amplifier is switched off, and the bit lines are precharged again.

  • 7/29/2019 low power memory design

    8/29

    8

    Fig 1.2.1 - Principle of operation of DRAM read, for simple 4 by 4 array

    1.3.2 To write to memory

    To store data, a row is opened and a given column's sense amplifier is

    temporarily forced to the desired high or low voltage state, thus causing the bit-line to charge or

    discharge the cell storage capacitor to the desired value. Due to the sense amplifier's positive

    feedback configuration, it will hold a bit-line at stable voltage even after the forcing voltage is

    removed. During a write to a particular cell, all the columns in a row are sensed simultaneously

    just as during reading, so although only a single column's storage-cell capacitor charge is

    changed, the entire row is refreshed (written back in).

  • 7/29/2019 low power memory design

    9/29

    9

    1.3.3 Refresh rate

    Typically, manufacturers specify that each row must have its storage cell

    capacitors refreshed every 64 ms or less. Refresh logic is provided in a DRAM controller which

    automates the periodic refresh, that is no software or other hardware has to perform it. This

    makes the controller's logic circuit more complicated, but this drawback is outweighed by the

    fact that DRAM is much cheaper per storage cell and because each storage cell is very simple,

    DRAM has much greater capacity per unit of surface than SRAM.

    Some systems refresh every row in a burst of activity involving all rows every

    64 ms. Other systems refresh one row at a time staggered throughout the 64 ms interval. For

    example, a system with 213

    = 8192 rows would require a staggered refresh rate of one row every

    7.8 s which is 64 ms divided by 8192 rows. A few real-time systems refresh a portion of

    memory at a time determined by an external timer function that governs the operation of the rest

    of a system, such as the vertical blanking interval that occurs every 1020 ms in video

    equipment. All methods require some sort of counter to keep track of which row is the next to be

    refreshed. Most DRAM chips include that counter. Older types require external refresh logic to

    hold the counter.Under some conditions, most of the data in DRAM can be recovered even if the

    DRAM has not been refreshed for several minutes.

    http://en.wikipedia.org/w/index.php?title=Refresh_logic&action=edit&redlink=1http://en.wikipedia.org/wiki/Refresh_ratehttp://en.wikipedia.org/wiki/Vertical_blanking_intervalhttp://en.wikipedia.org/wiki/Vertical_blanking_intervalhttp://en.wikipedia.org/wiki/Refresh_ratehttp://en.wikipedia.org/w/index.php?title=Refresh_logic&action=edit&redlink=1
  • 7/29/2019 low power memory design

    10/29

    10

    1.4 SRAM

    Static random-access memory (SRAM) is a type ofsemiconductor memory that

    uses bistable latching circuitry to store each bit. The term static differentiates itfrom dynamic RAM (DRAM) which must be periodically refreshed. SRAM exhibits data

    remanence , but it is still volatile in the conventional sense that data is eventually lost when the

    memory is not powered.

    1.4.1 Design

    A typical SRAM cell is made up of six MOSFETs. Each bit in an SRAM is

    stored on four transistors(M1, M2, M3, M4) that form two cross-coupled inverters. This storage

    cell has two stable states which are used to denote 0 and 1. Two additional access transistorsserve to control the access to a storage cell during read and write operations. In addition to such

    six-transistor (6T) SRAM, other kinds of SRAM chips use 4T, 8T, 10T, or more transistors per

    bit. Four-transistor SRAM is quite common in stand-alone SRAM devices (as opposed SRAM

    used for CPU caches), implemented in special processes with an extra layer ofpolysilicon ,

    allowing for very high-resistance pull-up resistors . Four transistor SRAM provides advantages

    in density at the cost of manufacturing complexity. The resistors must have small dimensions

    and large values . This is sometimes used to implement more than one (read and/or write) port,

    which may be useful in certain types ofvideo memory and register files implemented with multi-

    ported SRAM circuitry. Generally, the fewer transistors needed per cell, the smaller each cell can

    be. Since the cost of processing a silicon wafer is relatively fixed, using smaller cells and so

    packing more bits on one wafer reduces the cost per bit of memory. Memory cells that use fewer

    than four transistors are possiblebut such 3T or 1T cells are DRAM, not SRAM (even the so-

    called 1T-SRAM).Access to the cell is enabled by the word line which controls the

    two access transistors M5 and M6 which, in turn, control whether the cell should be connected to

    the bit lines: BL and BL. They are used to transfer data for both read and write operations.

    Although it is not strictly necessary to have two bit lines, both the signal and its inverse are

    typically provided in order to improve noise margins . During read accesses, the bit lines are

    actively driven high and low by the inverters in the SRAM cell. This improves SRAM bandwidth

    compared to DRAMsin a DRAM, the bit line is connected to storage capacitors and charge

    http://en.wikipedia.org/wiki/Semiconductorhttp://en.wikipedia.org/wiki/Multivibratorhttp://en.wikipedia.org/wiki/Flip-flop_(electronics)http://en.wikipedia.org/wiki/DRAMhttp://en.wikipedia.org/wiki/Memory_refreshhttp://en.wikipedia.org/wiki/Data_remanencehttp://en.wikipedia.org/wiki/Data_remanencehttp://en.wikipedia.org/wiki/Volatile_memoryhttp://en.wikipedia.org/wiki/MOSFEThttp://en.wikipedia.org/wiki/Bithttp://en.wikipedia.org/wiki/Transistorhttp://en.wikipedia.org/wiki/Polysiliconhttp://en.wikipedia.org/wiki/Video_memoryhttp://en.wikipedia.org/wiki/Register_filehttp://en.wikipedia.org/wiki/DRAMhttp://en.wikipedia.org/wiki/1T-SRAMhttp://en.wikipedia.org/wiki/Noise_marginhttp://en.wikipedia.org/wiki/Dynamic_Random_Access_Memoryhttp://en.wikipedia.org/wiki/Charge_sharinghttp://en.wikipedia.org/wiki/Charge_sharinghttp://en.wikipedia.org/wiki/Dynamic_Random_Access_Memoryhttp://en.wikipedia.org/wiki/Noise_marginhttp://en.wikipedia.org/wiki/1T-SRAMhttp://en.wikipedia.org/wiki/DRAMhttp://en.wikipedia.org/wiki/Register_filehttp://en.wikipedia.org/wiki/Video_memoryhttp://en.wikipedia.org/wiki/Polysiliconhttp://en.wikipedia.org/wiki/Transistorhttp://en.wikipedia.org/wiki/Bithttp://en.wikipedia.org/wiki/MOSFEThttp://en.wikipedia.org/wiki/Volatile_memoryhttp://en.wikipedia.org/wiki/Data_remanencehttp://en.wikipedia.org/wiki/Data_remanencehttp://en.wikipedia.org/wiki/Memory_refreshhttp://en.wikipedia.org/wiki/DRAMhttp://en.wikipedia.org/wiki/Flip-flop_(electronics)http://en.wikipedia.org/wiki/Multivibratorhttp://en.wikipedia.org/wiki/Semiconductor
  • 7/29/2019 low power memory design

    11/29

    11

    sharing causes the bit line to swing upwards or downwards. The symmetric structure of SRAMs

    also allows for differential signaling, which makes small voltage swings more easily detectable.

    Another difference with DRAM that contributes to making SRAM faster is that commercial

    chips accept all address bits at a time. By comparison, commodity DRAMs have the address

    multiplexed in two halves, i.e. higher bits followed by lower bits, over the same package pins in

    order to keep their size and cost down.The size of an SRAM with m address lines and n data

    lines is 2m words, or 2m n bits.

    Fig 1.3.1 -A six-transistor CMOS SRAM cell.

    Fig 1.3.2 - Four transistor SRAM

    1.4.2 SRAM Operation

    An SRAM cell has three different states. It can be in: standby (the circuit is

    idle), reading (the data has been requested) and writing (updating the contents). The SRAM to

    operate in read mode and write mode should have "readability" and "write stability" respectively.

    The three different states work as follows:

    http://en.wikipedia.org/wiki/Charge_sharinghttp://en.wikipedia.org/wiki/Differential_signalinghttp://en.wikipedia.org/wiki/Differential_signalinghttp://en.wikipedia.org/wiki/Charge_sharing
  • 7/29/2019 low power memory design

    12/29

    12

    Standby

    If the word line is not asserted, the access transistors M5 and M6 disconnect the cell from

    the bit lines. The two cross-coupled inverters formed by M1M4 will continue to reinforce eachother as long as they are connected to the supply.

    Reading

    Assume that the content of the memory is a 1, stored at Q. The read cycle is started by

    precharging both the bit lines to a logical 1, then asserting the word line WL, enabling both

    the access transistors. The second step occurs when the values stored in Q and Q are transferred

    to the bit lines by leaving BL at its precharged value and discharging BL through M1 and M5 to a

    logical 0 (i. e. eventually discharging through the transistor M1 as it is turned on because the Q is

    logically set to 1). On the BL side, the transistors M4 and M6 pull the bit line toward VDD, a

    logical 1 (i. e. eventually being charged by the transistor M4 as it is turned on because Q is

    logically set to 0). If the content of the memory was a 0, the opposite would happen

    and BL would be pulled toward 1and BL toward 0. Then these BL and BL will have a small

    difference of delta between them and then these lines reach a sense amplifier, which will sense

    which line has higher voltage and thus will tell whether there was 1 stored or 0. The higher the

    sensitivity of sense amplifier, the faster the speed of read operation is.

    Writing

    The start of a write cycle begins by applying the value to be written to the bit lines. If

    we wish to write a 0, we would apply a 0 to the bit lines, i.e. setting BL to 1 and BL to 0. This is

    similar to applying a reset pulse to an SR-latch, which causes the flip flop to change state. A 1 is

    written by inverting the values of the bit lines. WL is then asserted and the value that is to be

    stored is latched in. Note that the reason this works is that the bit line input-drivers are designed

    to be much stronger than the relatively weak transistors in the cell itself, so that they can easilyoverride the previous state of the cross-coupled inverters. Careful sizing of the transistors in an

    SRAM cell is needed to ensure proper operation.

    http://en.wikipedia.org/wiki/Latch_(electronic)http://en.wikipedia.org/wiki/Latch_(electronic)
  • 7/29/2019 low power memory design

    13/29

    13

    CHAPTER-2

    2.1 SOURCES AND REDUCTION OF POWER DISSIPATION IN MEMORY

    SUBSYSTEM

    Major sources of active power in a modern CMOS memory subsystem are

    active chips , whose power is a product of the activated chip power and the number of

    simultaneously activated chips . Other minor sources are the capacitances of the data-bus lines ,

    address lines and control lines on a memory board . In particular , data-bus lines are of special

    concern . They cause a heavy capacitance compared with other lines because of there inherently

    large values in number and capacitance . Active chip power and AC capacitive power have both

    been reduced by -

    lowpower circuits miniaturized device technology multi-data-bit chip configuration small package technology low voltage interfaces

    2.1.2 Multi-data-bit Configuration Chip

    The multi-data-bit configuration is becoming important and more favorable for power

    reduction , as well as design flexibility , as chip capacity increases .This is due to the resultant

    lower chip count combined with low power chip technology , for a fixed memory capacity of the

    subsystem . Examplea 4MByte , 32-bit data bus subsystem necessitates 32 memory chips for 1

    Mb chip of 1-Mword x 1 bit configuration .With multi-data-bit configuration , the number of

    chips necessary is 8 for a 4-Mb chip(1-Mword x 4-bit) and 2 for a 1-Mb chip(1-Mword x 16-bit).

    As shown in fig 2.2.1 .

  • 7/29/2019 low power memory design

    14/29

    14

    Fig 2.1.1 Subsystem using multi-bit-data configuration chip

    Source - K.Itoh,K. Sasaki and Y .Nakagome,Trends in Low-Power RAM Circuit Technologies Proc IEEE ,

    April 1995

    In addition to increasing memory capacity of a chip , widening data-bit width

    increases chip power because of resultant increase increased number of column and sense

    circuits . However , chip power has been suppressed with low power circuits . This arrangement

    dramatically reduces subsystem power with less chip count . On the contrary , a fixed data-bit

    configuration gives a disadvantages such as large power as well as design inflexibility . drastic

    reduction in subpower system is achieved with less chip count resulting from multi-data-bit

    configuration combined with large memory capacity and low-power chip technology

    advancement as shown in table 2.1.2 .

  • 7/29/2019 low power memory design

    15/29

    15

    Table2.1.15V DRAM subsystems with 32 data-bit bus lines

    Table 2.1.25V DRAM chip and package advancement

    2.1.2 Small Package

    Small package technology reduces the AC capacitive power by reducing

    the capacitance on the memory board . Common I/O pin assignment in which the data input

    (Din)pin and the data output (Dout) pin are common enables the use of small packages by halving

    the number of data pins that increase with multi-data-bit configuration . Furthermore ,

    remarkable progress in package structure accommodates ever-enlarged chips at every successive

    generation ina small packages. A typical example is the LOC (LEAD ON CHIPS) package

    featuring a vertical structure of chip and inner leads as shown in fig 2.1.2 .

  • 7/29/2019 low power memory design

    16/29

    16

    Fig 2.1.2advancement in packages

    2.1.3 Low-voltage Data-bus Interface

    Further AC power reduction has been required because the ever-increased data-

    throughput of memory subsystems , resulting from the rapidly improved performance of

    microprocessors , increases AC power . In traditional address multiplexed DRAM chip , as

    shown in fig a selected word line , WL , is activated after strobing the corresponding set of row

    addresses with an external clock called Row addressing strobe(RAS ).As a result all the stored

    information of memory calls along the selected WL are read on the corresponding data lines ,

    and then amplified by the corresponding amplifiers .any read information can be read out to the

    data output (Dout ) pin by strobing corresponding set of column address with another clock called

    column address strobe (CAS ) .

    Now evaluating AC capacitive power , AC capacitive power of each data-

    bus line in 5 V TTL (Transistor -Transistor Logic) as shown in fig 2.1.3 is expressed by -

    P=()C(V)2f

  • 7/29/2019 low power memory design

    17/29

    17

    where C, V andfare loading capacitance , voltage swing and frequency respectively.

    It is found that AC power of an 8-Mb,32-bit data-bus-line subsystem becomes relatively large as

    the width of data I/O increases , although it is still small portion of total power . V = VDDVT ,

    since the gate voltage of the Dout transistor is usually equals VDD . Chip power increases as the

    DOUT transistor enlarge according to the to high speed driving of bus capacitance , increases

    the internal node capacitance of the chip . Recently new interfaces called GTL(Gunning

    Transceiver Logic) , CTT(Center Tapped Termination ) and Rambus interfaces featuring small

    amplitude impedance matched bus lines have been proposed .Small amplitude minimizes AC

    capacitance power while impedance matching allows extremely high speed transmission .The

    Power consumption of each CTT interface is roughly expressed as

    P=(IOLVOL+RTI2OL)(1duty)+(IOH(VDD-VOH)+RTI2OH)duty+()C(V)2f

    Where IOL and IOH are currents for the low signal voltage(VOL) and the high signal voltage (VOH)

    respectively and RT is the termination resistance . Even for new interfaces , the chip power

    component dominates the total power , this is due to additional on- chip interface circuits while

    the interface power component is still low enough in high speed region .

    Fig -2.1.3 Typical TTL interface

    Source- M. Taguchi High-Speed ,Small-Amplitude I/O Interface Circuits for memory bus application

    IEICE Trans. Electron ., Dec 1994

  • 7/29/2019 low power memory design

    18/29

    18

    2.2 SOURCES OF POWER DISSIPATION IN DRAM AND SRAM

    2.2.1 Active Power Sources

    The chip comprises three major blocks of power sources :

    memory cell array , decoders(row and column ) periphery .

    All the m cells on one word line are simultaneously activated in the logical array model . An

    unified active power equation for modern CMOS DRAMs and SRAMs is approximately given

    by

    P= VDDIDD

    IDD = miact+m(n-1)ihld+(n+m)CDEVINT+CPTVINT+IDCP

    where VDD is external supply voltage

    IDD is the current of VDD

    Iact is effective current of a active or selected cells

    Ihld is effective data retention current of inactive or non selected cell

    CDE is the output node capacitance of each decoder

    VINT is the internal supply voltage

    CPT is the total capacitance of the CMOS logic and the driving circuitry

    IDCP total static(DC) or quasi-static current of the periphery

    Major sources of IDCP are the column circuitry and the differential amplifiers

    on the input output lines .other contributors are refresh related circuits and on chip voltage

    converters essential to DRAM operations : these includes a substrate back-bias generator , a

  • 7/29/2019 low power memory design

    19/29

    19

    voltage-down converter , a voltage up converter , a voltage reference circuit and a half V DD

    generator .At high frequency the data retention current m(n-1) ihld , is negligible evidence by the

    small cell leakage current and small periphery current necessary for the refresh operation in

    DRAMs.IDD increases with increasing memory capacity that is with increased m and n.

    Fig-2.2.1 Sources of power dissipation in a RAM chip

    Source- M. Taguchi High-Speed ,Small-Amplitude I/O Interface Circuits for memory bus application

    IEICE Trans. Electron ., Dec 1994

    DRAM - DRAM-Destructive read-out characteristics of a DRAM cell necessitate successive

    operations of amplification and restoration for a selected cell on every data line . Consequently ,

    a data line is charged with a large voltage swing of V D(usually 1.5-2.5V) and with charging

    current of CDVD where CD is the data line capacitance . Hence current is expressed as -

  • 7/29/2019 low power memory design

    20/29

    20

    IDD = [mCDVD + CPTVINT] + IDCP

    Key to reduce active power is

    (i)reducing the charging capacitance(mCD,CPT),

    (ii)lowering the external and internal voltages (VDD,VINT,VD) ,

    (iii)reducing the active current(IDCP).

    In particular , emphasis must be placed on reduction of the total data-line

    dissipation charge(mCDVD). Since it dominates the total active power .However dissipation

    charge must be reduced while maintaining an acceptable S/N ratio for stable operation . The

    signal Vs is approximately expressed as -

    Vs = (Cs/CD)VDD/2 = (Cs/CD)VD = QS/CD

    Reducing CD is effective for both reducing IDD and increasing VS, while reducing VD degrades

    VS , despite the IDD reduction .

    SRAM- The non destructive read-out characteristics of SRAM never require restoration of cell

    data , allowing the elimination of a sense amplifier on each data line . In SRAM V D is very

    small(=0.1V-0.3V) although it is prominent for write operation . Thus the current for readoperation is expressed as -

    IDD =[miDCt + CPTVINT]f+ IDCP

    To reduce power the three key points are same as was there in DRAMs . However , the

    static current charge , miDCt , should differs between SRAMs and DRAMs .S/N ratio is not so

    serious problem in SRAM because of ratio operations .

    Eventually ,DRAMs and SRAMs have evolved to use similar circuit technology ,

    although emphasis on each of the three issues is different between both types of RAM . A partial

    activation scheme of a multi divided word line reduces the DRAM current down to the SRAM

    level , although it is in experimental stage .In addition to the active current described , the

    subthreshold DC current of a MOSFET will be a source of active power dissipation for a future

  • 7/29/2019 low power memory design

    21/29

    21

    DRAMs and SRAMs .As VT becomes small enough to no longer cut off the transistor , the

    subthreshold current increases exponentially with decreasing VT .The subthreshold current

    generated from an overwhelming large number of inactive circuits would eventually be larger

    than AC current from the small number of inactive circuits , dominating the total chip current .

    Table -2.2.1 -Comparison of determinants of active cell current between DRAMs and SRAMs

    Source- K.Itoh,K. Sasaki and Y .Nakagome,Trends in Low-Power RAM Circuit TechnologiesProc IEEE , April 1995

    2.2.2 Data Retention Power Sources

    DRAMIn the data retention mode , a memory chip is not accessed from outside and the data

    are retained by the refresh operation . The refresh operation is performed by reading data of the

    m cells on a word line and restoring them for each of the n word lines in order . Note that n in the

    logic array responds to the number of refresh cycle . A current flows every time m cells arerefreshed at the same time . The frequencyfat which the refresh current flows is n/tREF , where

    tREF is the refresh time of the cells in the retention mode and increases with reducing junction

    temperature . Data retention current is given by

    IDD = [mCDVD + CPTVINT](n/ tREF) + IDCP

  • 7/29/2019 low power memory design

    22/29

    22

    This is because tREFmax is for the active mode when operating the memory at the

    maximum frequency of around 10MHz and where cell leakage current is maximized with highest

    junction temperature . In any event ,IDCP becomes relatively large for other AC current

    components because of small n/ tREF .This implies the necessity of reducing both AC and DC

    components .

    SRAM In low power CMOS SRAMs , the static cell leakage current , mnIhld , is the major

    source of retention current because of the negligibly small IDCP . A high resistance polysilicon

    load cell has been widely used for its high-density and low Ihld characteristics , although a full

    CMOS 6-T cell provides the smallest Ihld . For ultralow voltage operation involving scaled VT

    , subthreshold currents from almost all circuits in DRAMs and SRAMs could drastically increase

    the retention current , again posing a serious concern .

  • 7/29/2019 low power memory design

    23/29

    23

    CHAPTER-3

    3.1 Low Power DRAM Circuits

    3.1.1 Active power reduction

    For a given memory capacity chip , successive circuit advancements have

    produced a power reduction equivalent to 2 to 3 orders of magnitude over the last decade . Fig

    3.1.1 shows the power dissipation of a 64-Mb DRAM , hypothetically designed with the NMOS

    circuit of the 64- Kb generation in 1980 ,compared with the CMOS circuit presented in 1990 .

    Almost the same process and device technology are assumed. The drastic reduction in power by

    about two orders of magnitude is due to many sophisticated circuits : partial activation of a

    multi-divided data-line and shared I/O which reduces mCD .

    Fig- 3.1.1 Low power circuit advancement of DRAMs over the last decade

    Source- M. Taguchi High-Speed ,Small-Amplitude I/O Interface Circuits for memory bus application

    IEICE Trans. Electron ., Dec 1994

    3.1.1.1 Charging Capacitance Reduction

    Partial activation of multi-divided data lines is widely accepted .In this methodshared I/O divides a multi-divided data line into two , which are selected by the isolation

    switches ,ISO and ISO .The shared SA provides an almost doubled cell signal with halved CD .

    The partial activation is performed by activating only one sense amplifier along the data line .

  • 7/29/2019 low power memory design

    24/29

    24

    Partialactivation of Multi-divided Data Line

    A practical solution is to divide one data line into several sections and to active

    only one section . In the early years , the number of divisions was increased by simply increasing

    the number of Y decoders by using additional chip area . This Y decoder division was applied to

    various sense amplifier (SA) arrangement such as one SA at each division and a shared SA .

    However , this division was almost exhausted at the 16-Mb generation with the combination of

    shared SA , shared and shared Y decoder . Partial activation of multi-divided data lines is widely

    accepted .In this method shared I/O divides a multi-divided data line into two , which are

    selected by the isolation switches . The shared SA provides an almost doubled cell signal with

    halved CD . The partial activation is performed by activating only one sense amplifier along the

    data line .

    Fig 3.1.2 Multi divided data line architecture with SA , shared I/O and shared Y decoder

    Refresh time increase

    Reduction of m is never achieved without without the help of increasing the

    maximum refresh time of the cell , tREFmax . This resulted to preserve the busy rate expressed as

    = tRCmin/(tREFmax/n) = (M/m)(tRCmin/tREFmax)

  • 7/29/2019 low power memory design

    25/29

    25

    Where trcmin and M are the minimum cycle time and memory capacity , respectively .A smaller

    is preferred since it involves less conflicts between refresh and normal operation . Hence , for a

    fixed M , it is necessary to maintain mtREFmax to keep constant assuming a fixed tRCmin .

    3.1.2 Operating Voltage Reduction

    VDD reduction from 12 to 5 V at the 64-Kb generation and then to 3.3 V at the 64-

    Mb generation , half VDD data-line-precharge which has been widely used since commercial 1-

    Mb DRAMs are well known as contributors to low power .

    Half VDD Data Line Precharge

    Half Vdd precharge halves the data-line power of full-Vdd precharge with halved

    data-line voltage swing . In full VDD precharge , the voltage difference difference on the data

    lines is amplified by applying pulse and then resultant degraded high level is restored to full

    VDD precharge with halved data-line voltage swing .

    On Chip Voltage down Converter

    An internal voltage , VINT , reduced by scaling the same electric field . This

    eventually provides a low power (reduced by 1/k) with higher speed and smaller chip area . For

    accuracy and load current driving capabilities , it consists of a current-mirror differentialamplifier (Q1-Q4,Qs)and common-source drive transistors(Q6) as shown in Fig -3.1.2 In order

    tpo minimize the output voltage drop YVDP , the gate voltage of Q6 , VG has to be respond

    quickly when the output goes low . An amplifier currents , Is of 2 to 3 mA enables such a fast

    response time . Bias current source Ib is needed to clamp the output voltage when the load

    current becomes almost zero .

  • 7/29/2019 low power memory design

    26/29

    26

    Fig 3.1.3Vottage down converter

    SourceH.Tanaka, Stabilization of voltage limiter circuit for high density DRAM,IEICE

    ,Nov.1992

    3.1.3 DC current reduction

    Fig 3.1.3 shows column signal path circuitry which is a main source of static

    current . It consists of a pair of data lines , a column switch , a pair of I/O lines . DC current

    flows from I/O line load to data lines while the column switch is on . Also , the main amplifier

    consumes DC current for amplifying the signal , because it usually employes the conventional

    current mirror differential amplifier . The current reductions are essential especially for multi

    data bit chip because of the increased number of column circuitaries . The DC current are shut

    down with pulse-operation technique .

  • 7/29/2019 low power memory design

    27/29

    27

    Fig 3.1.4 Pulse operation of column signal path circuitary

    3.1.4 Data Retention Power Reduction

    Reduction of both DC and AC current components in data-retention mode is a

    prime concern. Minimizing the power of on-chip voltage converters such as VDC , voltage

    up(VDH)converter , substrate back bias(VBB) generator . VREF generator, and half-VDD

    generator reduces the DC current component .Extending the refresh time and reducing refresh

    charge reduce the AC current component .

    3.1.5 Refresh time extension

    To extend the refresh time tREF according to a reduced junction temperature in

    data retention mode , a self- refresh control with an on-chip temperature detection circuit and the

    use of a cell-leakage monitor circuit on the chip have been proposed .

  • 7/29/2019 low power memory design

    28/29

    28

    CONCLUSION

    We have presented a methodology for low power design of DRAM

    memories. This method can be applied to electronic design automation tools to ease the

    burden on industry designers. We first examined some of the previous methods in

    designing low power RAMs. We looked at the architecture level issues such as partial

    activation of divided word and data lines and partial activation of divided power lines.

    Also we explored the pulsed operation of word line circuitry. Next we looked at some

    circuit issues such as current mode operation and an on-chip voltage down converter . The

    development till date has been discussed in respect to DRAM and SRAM . In this report every

    aspect of loses have been taken care of so that while designing DRAM , one optimize it for theminimum power loses Future work looks promising .

  • 7/29/2019 low power memory design

    29/29

    REFERENCES

    [1] Low Power Design Methodologies by Jan M.Rabaey and Massoud Pedram springer

    publications .year of publication 1996

    [2] - K.Itoh,K. Sasaki and Y .Nakagome,Trends in Low-Power RAM Circuit Technologies

    Proc IEEE , April 1995

    [3] M. Taguchi High-Speed ,Small-Amplitude I/O Interface Circuits for memory bus

    application IEICE Trans. Electron ., Dec 1994

    [4] Tools and Methodologies for Low Power DesignJerry Frenkil Sente, Inc. Proc IEEE ,

    April 1998