US12332720B2 - Low power optimization based upon host exit latency - Google Patents
Low power optimization based upon host exit latency Download PDFInfo
- Publication number
- US12332720B2 US12332720B2 US18/218,864 US202318218864A US12332720B2 US 12332720 B2 US12332720 B2 US 12332720B2 US 202318218864 A US202318218864 A US 202318218864A US 12332720 B2 US12332720 B2 US 12332720B2
- Authority
- US
- United States
- Prior art keywords
- actions
- data storage
- storage device
- group
- low power
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3234—Power saving characterised by the action undertaken
- G06F1/325—Power saving in peripheral device
- G06F1/3268—Power saving in hard disk drive
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3206—Monitoring of events, devices or parameters that trigger a change in power modality
- G06F1/3215—Monitoring of peripheral devices
- G06F1/3221—Monitoring of peripheral devices of disk drive devices
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3206—Monitoring of events, devices or parameters that trigger a change in power modality
- G06F1/3215—Monitoring of peripheral devices
- G06F1/3225—Monitoring of peripheral devices of memory devices
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3234—Power saving characterised by the action undertaken
- G06F1/325—Power saving in peripheral device
- G06F1/3275—Power saving in memory, e.g. RAM, cache
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3234—Power saving characterised by the action undertaken
- G06F1/3296—Power saving characterised by the action undertaken by lowering the supply or operating voltage
Definitions
- Embodiments of the present disclosure generally relate to data storage devices, such as solid state drives (SSDs), and, more specifically, optimizing device-exit-latency of the data storage device.
- SSDs solid state drives
- the data storage device When a data storage device has no commands to perform and/or when there are no input/output (I/O) related activities over a peripheral component interconnect (PCI) express (PCIe) coupling the data storage device to a host device, the data storage device may be transitioned to a firmware (FW) active idle state.
- FW firmware
- the data storage device may attempt to reduce its power consumption, such as by turning off one or more clocks, turning off certain components of the data storage device, or transitioning the PCIe link into a low power state (e.g., L1.2), which may effectively place the data storage device in a low power state.
- L1.2 low power state
- the data storage device will also send a long term retention (LTR) notification to the host device indicating a new latency requirement associated with the power state transition.
- LTR long term retention
- the host device In order to wake up the data storage device from the low power state, the host device sends a wake up indication to the data storage device.
- the time it takes for the host device to pull the data storage device out of the low power mode i.e., when the host device sends a doorbell associated with a command
- the host exit latency is the host exit latency.
- the data storage device may still be initializing or restoring one or more clocks, components, and the like that were turned off in order to reduce power consumption. Thus, completing the command sent by the host device may be delayed, which may result in a lower quality of service (QOS) provided by the data storage device.
- QOS quality of service
- the present disclosure generally relates to data storage devices, such as solid state drives (SSDs), and, more specifically, optimizing device-exit-latency of the data storage device.
- SSDs solid state drives
- a controller of the data storage device initiates a counter in order to determine a host exit latency. Based on the host exit latency, the controller determines a group of low power state entrance actions from a plurality of groups to perform during a next entrance into a firmware active idle state based on an associated completion wake up time and the host exit latency. The controller selects the group whose completion wake up time is closest to the host exit latency and less than or equal to the host exit latency. The controller performs the selected groups low power state entrance actions during a next entrance into the firmware active idle state.
- a data storage device includes a memory device and a controller coupled to the memory device.
- the controller is configured to receive a wake up indication from a host device, where the data storage device is in a low power state, receive a doorbell from the host device indicating that a new command is in a submission queue of the host device, determine a host exit latency, where the host exit latency is a time between the wake up indication and the doorbell, and utilize the host exit latency to select a group of actions from a plurality of group of actions based on the determining, where each of the plurality of group of actions comprises a different amount of low power state entrance actions to complete during a next low power state entrance.
- a data storage device in another embodiment, includes a memory device and a controller coupled to the memory device.
- the controller includes a firmware-active detector and a latency program selector coupled to the firmware-active detector.
- the firmware-active detector is configured to determine a host exit latency, where the host exit latency is a time between receiving a wake up indication from a host device and receiving a doorbell subsequent to receiving the wake up indication from the host device.
- the latency program selector is configured to select a group of actions from a plurality of group of actions based on the determining, where each of the plurality of group of actions comprises a different amount of low power state entrance actions to complete during a next low power state entrance.
- a data storage device in another embodiment, includes a means for storing data and a controller coupled to the means for storing data.
- the controller is configured to determine that an idle time timeout threshold has been reached, perform a number of low power state entrance actions based on a previously determined host exit latency, where a completion wake up time to complete the number of low power state entrance actions is less than the previously determined host exit latency, and transition the data storage device into a low power mode.
- FIG. 1 is a schematic block diagram illustrating a storage system in which a data storage device may function as a storage device for a host device, according to certain embodiments.
- FIG. 2 is an illustration of a link power management flow diagram, according to certain embodiments.
- FIG. 3 is a flow diagram illustrating a method of FW active idle power flow, according to certain embodiments.
- FIG. 4 is a flow diagram illustrating a method of measuring host exit latency, according to certain embodiments.
- FIG. 5 is an exemplary illustration of power saving groups, according to certain embodiments.
- FIG. 6 is an exemplary illustration showing exit latency vs power consumption, according to certain embodiments.
- FIG. 7 is an exemplary graph of selecting a group of low power state actions based on a host exit latency, according to certain embodiments.
- FIG. 8 is a flow diagram illustrating a method of selecting a group of low power state actions based on a host exit latency, according to certain embodiments.
- FIG. 9 is a schematic block diagram illustrating a storage system in which a data storage device may function as a storage device for a host device, according to certain embodiments.
- the present disclosure generally relates to data storage devices, such as solid state drives (SSDs), and, more specifically, optimizing device-exit-latency of the data storage device.
- SSDs solid state drives
- a controller of the data storage device initiates a counter in order to determine a host exit latency. Based on the host exit latency, the controller determines a group of low power state entrance actions from a plurality of groups to perform during a next entrance into a firmware active idle state based on an associated completion wake up time and the host exit latency. The controller selects the group whose completion wake up time is closest to the host exit latency and less than or equal to the host exit latency. The controller performs the selected groups low power state entrance actions during a next entrance into the firmware active idle state
- FIG. 1 is a schematic block diagram illustrating a storage system 100 having a data storage device 106 that may function as a storage device for a host device 104 , according to certain embodiments.
- the host device 104 may utilize a non-volatile memory (NVM) 110 included in data storage device 106 to store and retrieve data.
- the host device 104 comprises a host DRAM 138 .
- the storage system 100 may include a plurality of storage devices, such as the data storage device 106 , which may operate as a storage array.
- the storage system 100 may include a plurality of data storage devices 106 configured as a redundant array of inexpensive/independent disks (RAID) that collectively function as a mass storage device for the host device 104 .
- RAID redundant array of inexpensive/independent disks
- the host device 104 may store and/or retrieve data to and/or from one or more storage devices, such as the data storage device 106 . As illustrated in FIG. 1 , the host device 104 may communicate with the data storage device 106 via an interface 114 .
- the host device 104 may comprise any of a wide range of devices, including computer servers, network-attached storage (NAS) units, desktop computers, notebook (i.e., laptop) computers, tablet computers, set-top boxes, telephone handsets such as so-called “smart” phones, so-called “smart” pads, televisions, cameras, display devices, digital media players, video gaming consoles, video streaming device, or other devices capable of sending or receiving data from a data storage device.
- NAS network-attached storage
- the host DRAM 138 may optionally include a host memory buffer (HMB) 150 .
- the HMB 150 is a portion of the host DRAM 138 that is allocated to the data storage device 106 for exclusive use by a controller 108 of the data storage device 106 .
- the controller 108 may store mapping data, buffered commands, logical to physical (L2P) tables, metadata, and the like in the HMB 150 .
- the HMB 150 may be used by the controller 108 to store data that would normally be stored in a volatile memory 112 , a buffer 116 , an internal memory of the controller 108 , such as static random access memory (SRAM), and the like.
- the controller 108 may utilize the HMB 150 as the DRAM of the data storage device 106 .
- the data storage device 106 includes the controller 108 , NVM 110 , a power supply 111 , volatile memory 112 , the interface 114 , a write buffer 116 , and an optional DRAM 118 .
- the data storage device 106 may include additional components not shown in FIG. 1 for the sake of clarity.
- the data storage device 106 may include a printed circuit board (PCB) to which components of the data storage device 106 are mechanically attached and which includes electrically conductive traces that electrically interconnect components of the data storage device 106 or the like.
- the physical dimensions and connector configurations of the data storage device 106 may conform to one or more standard form factors. Some example standard form factors include, but are not limited to, 3 .
- data storage device e.g., an HDD or SSD
- 2.5′′ data storage device 2.5′′ data storage device
- 1.8′′ data storage device peripheral component interconnect (PCI), PCI-extended (PCI-X), PCI Express (PCIe) (e.g., PCIe x1, x4, x8, x16, PCIe Mini Card, MiniPCI, etc.).
- PCIe PCI Express
- the data storage device 106 may be directly coupled (e.g., directly soldered or plugged into a connector) to a motherboard of the host device 104 .
- Interface 114 may include one or both of a data bus for exchanging data with the host device 104 and a control bus for exchanging commands with the host device 104 .
- Interface 114 may operate in accordance with any suitable protocol.
- the interface 114 may operate in accordance with one or more of the following protocols: advanced technology attachment (ATA) (e.g., serial-ATA (SATA) and parallel-ATA (PATA)), Fibre Channel Protocol (FCP), small computer system interface (SCSI), serially attached SCSI (SAS), PCI, and PCIe, non-volatile memory express (NVMe), OpenCAPI, GenZ, Cache Coherent Interface Accelerator (CCIX), Open Channel SSD (OCSSD), or the like.
- ATA advanced technology attachment
- SATA serial-ATA
- PATA parallel-ATA
- FCP Fibre Channel Protocol
- SCSI small computer system interface
- SAS serially attached SCSI
- PCI PCI
- NVMe non-volatile memory express
- OpenCAPI OpenCAPI
- Interface 114 (e.g., the data bus, the control bus, or both) is electrically connected to the controller 108 , providing an electrical connection between the host device 104 and the controller 108 , allowing data to be exchanged between the host device 104 and the controller 108 .
- the electrical connection of interface 114 may also permit the data storage device 106 to receive power from the host device 104 .
- the power supply 111 may receive power from the host device 104 via interface 114 .
- the NVM 110 may include a plurality of memory devices or memory units. NVM 110 may be configured to store and/or retrieve data. For instance, a memory unit of NVM 110 may receive data and a message from controller 108 that instructs the memory unit to store the data. Similarly, the memory unit may receive a message from controller 108 that instructs the memory unit to retrieve data. In some examples, each of the memory units may be referred to as a die. In some examples, the NVM 110 may include a plurality of dies (i.e., a plurality of memory units).
- each memory unit may be configured to store relatively large amounts of data (e.g., 128 MB, 256 MB, 512 MB, 1 GB, 2 GB, 4 GB, 8 GB, 16 GB, 32 GB, 64 GB, 128 GB, 256 GB, 512 GB, 1 TB, etc.).
- relatively large amounts of data e.g., 128 MB, 256 MB, 512 MB, 1 GB, 2 GB, 4 GB, 8 GB, 16 GB, 32 GB, 64 GB, 128 GB, 256 GB, 512 GB, 1 TB, etc.
- each memory unit may include any type of non-volatile memory devices, such as flash memory devices, phase-change memory (PCM) devices, resistive random-access memory (ReRAM) devices, magneto-resistive random-access memory (MRAM) devices, ferroelectric random-access memory (F-RAM), holographic memory devices, and any other type of non-volatile memory devices.
- non-volatile memory devices such as flash memory devices, phase-change memory (PCM) devices, resistive random-access memory (ReRAM) devices, magneto-resistive random-access memory (MRAM) devices, ferroelectric random-access memory (F-RAM), holographic memory devices, and any other type of non-volatile memory devices.
- the NVM 110 may comprise a plurality of flash memory devices or memory units.
- NVM Flash memory devices may include NAND or NOR-based flash memory devices and may store data based on a charge contained in a floating gate of a transistor for each flash memory cell.
- the flash memory device may be divided into a plurality of dies, where each die of the plurality of dies includes a plurality of physical or logical blocks, which may be further divided into a plurality of pages.
- Each block of the plurality of blocks within a particular memory device may include a plurality of NVM cells. Rows of NVM cells may be electrically connected using a word line to define a page of a plurality of pages.
- Respective cells in each of the plurality of pages may be electrically connected to respective bit lines.
- NVM flash memory devices may be 2D or 3D devices and may be single level cell (SLC), multi-level cell (MLC), triple level cell (TLC), or quad level cell (QLC).
- the controller 108 may write data to and read data from NVM flash memory devices at the page level and erase data from NVM flash memory devices at the block level.
- the power supply 111 may provide power to one or more components of the data storage device 106 .
- the power supply 111 may provide power to one or more components using power provided by an external device, such as the host device 104 .
- the power supply 111 may provide power to the one or more components using power received from the host device 104 via interface 114 .
- the power supply 111 may include one or more power storage components configured to provide power to the one or more components when operating in a shutdown mode, such as where power ceases to be received from the external device. In this way, the power supply 111 may function as an onboard backup power source.
- the one or more power storage components include, but are not limited to, capacitors, super-capacitors, batteries, and the like.
- the amount of power that may be stored by the one or more power storage components may be a function of the cost and/or the size (e.g., area/volume) of the one or more power storage components. In other words, as the amount of power stored by the one or more power storage components increases, the cost and/or the size of the one or more power storage components also increases.
- the volatile memory 112 may be used by controller 108 to store information.
- Volatile memory 112 may include one or more volatile memory devices.
- controller 108 may use volatile memory 112 as a cache. For instance, controller 108 may store cached information in volatile memory 112 until the cached information is written to the NVM 110 .
- volatile memory 112 may consume power received from the power supply 111 .
- Examples of volatile memory 112 include, but are not limited to, random-access memory (RAM), dynamic random access memory (DRAM), static RAM (SRAM), and synchronous dynamic RAM (SDRAM (e.g., DDR1, DDR2, DDR3, DDR3L, LPDDR3, DDR4, LPDDR4, and the like)).
- RAM random-access memory
- DRAM dynamic random access memory
- SRAM static RAM
- SDRAM synchronous dynamic RAM
- the optional DRAM 118 may be utilized to store mapping data, buffered commands, logical to physical (L2P) tables, metadata, cached data, and the like in the optional DRAM 118 .
- the data storage device 106 does not include the optional DRAM 118 , such that the data storage device 106 is DRAM-less. In other examples, the data storage device 106 includes the optional DRAM 118 .
- Controller 108 may manage one or more operations of the data storage device 106 . For instance, controller 108 may manage the reading of data from and/or the writing of data to the NVM 110 . In some embodiments, when the data storage device 106 receives a write command from the host device 104 , the controller 108 may initiate a data storage command to store data to the NVM 110 and monitor the progress of the data storage command. Controller 108 may determine at least one operational characteristic of the storage system 100 and store at least one operational characteristic in the NVM 110 . In some embodiments, when the data storage device 106 receives a write command from the host device 104 , the controller 108 temporarily stores the data associated with the write command in the internal memory or write buffer 116 before sending the data to the NVM 110 .
- the controller 108 may include an optional second volatile memory 120 .
- the optional second volatile memory 120 may be similar to the volatile memory 112 .
- the optional second volatile memory 120 may be SRAM.
- the controller 108 may allocate a portion of the optional second volatile memory to the host device 104 as controller memory buffer (CMB) 122 .
- the CMB 122 may be accessed directly by the host device 104 .
- the host device 104 may utilize the CMB 122 to store the one or more submission queues normally maintained in the host device 104 .
- the host device 104 may generate commands and store the generated commands, with or without the associated data, in the CMB 122 , where the controller 108 accesses the CMB 122 in order to retrieve the stored generated commands and/or associated data.
- FIG. 2 is an illustration of a link state flow diagram 200 , according to certain embodiments. Aspects of the storage system 100 may be referenced in the description herein for exemplary purposes.
- the data storage device 106 includes several link states. For example, the data storage device 106 may have the following 5 link states: L0, L0s, L1, L2, and L3, where L1 includes a L1.1 sub-state and a L1.2 sub-state. Each of the link states are associated with a distinct operation of the data storage device 106 .
- Link states L0, L0s, and L1 are considered operational link states and utilize a first range of power, whereas link states L2 and L3 are considered non-operational link states, utilizing a second range of power, where the first range of power is greater than the second range of power.
- An operational link state refers to the ability of the host device 104 to communicate with the NVM 110 of the data storage device 106 .
- a non-operational link state refers to the inability of the host device 104 to communicate with the NVM 110 of the data storage device 106 due to a shut down or disconnection of a link between the host device 104 and the controller 108 .
- the listed non-operational link states are not intended to be limiting and may include other link states, such as the L1.1 and L1.2 link states.
- more or less link states than the number of link states shown in the link state flow diagram 200 may be available and more or less low power link states may be applicable to the embodiments described herein.
- Link states are numbered sequentially, where higher numbers represent lower power requirements due to a greater number of offline circuits and corresponding higher exit latencies. Furthermore, each link state has an associated power requirement and an exit latency. L0 and L0s may require 4.5 W with the lowest exit latency. L1 may require less power than L0, such as 3 W, and may have an exit latency equal to or higher than the exit latency of L0. L2 may require less power than L1 and may have an exit latency equal to or higher than the exit latency of L1. L3 may require less power than L2 and may have an exit latency equal to or higher than the exit latency of L2. The values for the link states and exit latencies are not intended to be limiting, but to provide an example of possible embodiments.
- L0 is referred to as a fully operational state, where I/O commands are enabled, and the device may generate interrupts.
- L0 is a link state where the link is operating normally. Interrupts are an automatic transfer of firmware execution due to a system timer or a user command.
- Link states L0s and L1 are also operational states; however, L0s and L1 may have a lower functionality than that of L0. For example, L0s has a similar power requirement as that of the L0, but only allows for a serial link in one direction. In the L0s link state, data may be transferred in one direction, but not the other. Thus, when a first device is coupled to a second device through a link, the first device may idle a transmitter of the first device independently and separately of the second device idling a transmitter of the second device, and/or vice-versa.
- L1 allows for a bidirectional serial link and allows for a greater reduction in the power requirement, but has a higher exit latency than that of L0 and L0s.
- Link states L2 and L3 are non-operational link states have a power requirement less than that of the operational link states. The difference between the L2 link state and the L3 link state is that power has not been yet removed from the L2 link state.
- the memory devices of the NVM 110 that are not used are placed in a non-operational link state, L2 and L3, to limit the idle power consumption to a minimal value.
- the link such as a data bus
- the controller 108 changes the link state of the link between the host device 104 and the controller 108 from the operational link states, such as L0, L0s, or L1, to a different operational link state, such as L0, L0s, or L1, or to a non-operational link state, such as L2 or L3, depending on the situation.
- the link will need to be in link state L2/L3 ready, which is a pseudo-state to prepare the component for a loss in power and reference clock(s).
- the controller 108 allocates the appropriate amount of power to return all link states L0s, L1, L2, L3 into link state L0 when a full operational state is required. For example, to return to L0 from L2 or L3, the link transitions to a transient pseudo-state, LDn, before transitioning to L0.
- the LDn state may be a fundamental reset state, a hot reset state, or a link disable transmission state by the upstream component (e.g., the host device 104 ).
- the link state L1 includes additional sub-states, L1.1 and L1.2, where the link state L1 may be referred to as L1.0.
- the L1 sub-states (L1SS), L1.1 and L1.2, may require more power for operation than L2 and L3; however, the L1SS utilizes less power than the L1.0 state.
- L1SS the link remains operational and requires less power to return to a more operational state, such as L1.0 or L0.
- the L1SS requires less time than the L2 and/or the L3 link states to return to a full active link state L0.
- link state may be referred to as “power state” herein for exemplary purposes.
- FIG. 3 is a flow diagram illustrating a method 300 of FW active idle power flow, according to certain embodiments.
- Method 300 may be executed by a controller, such as the controller 108 of FIG. 1 .
- a controller such as the controller 108 of FIG. 1 .
- aspects of the storage system 100 of FIG. 1 may be referenced herein.
- the data storage device is in a FW active idle state.
- the controller 108 starts an idle counter.
- the idle counter may be referred to as a “counter” herein for simplification purposes.
- the idle counter may be a generic or dedicated counter of a timer of the controller 108 .
- the controller determines if the counter is greater than or equal to an idle time timeout threshold, where the idle time timeout threshold is a maximum time before the controller 108 transitions the data storage device 106 into a low power mode. If the counter is greater than or equal to the idle time timeout threshold at block 306 , then the controller 108 stops the counter at block 308 .
- the controller 108 transitions the data storage device 106 into a low power mode.
- the transitioning may include one or more of powering down one or more hardware modules of the data storage device 106 while retaining memories, powering down one or more hardware modules of the data storage device 106 without retaining memories, turning off one or more clocks, turning off one or more phase lock loops (PLLs), and the like.
- PLLs phase lock loops
- the controller 108 transitions may transition the data storage device 106 to the L1.2 link state as described in FIG. 2 . Transitioning to the low power mode may also include notifying the host device 104 , by the controller 108 , that the data storage device 106 is entering the low power mode.
- the controller 108 receives a wake up indication from the host device 104 to return form the low power state (e.g., L1.2) to an operational power state (e.g., L0, L0s, or L1).
- the controller 108 may restore each of the one or more hardware modules, each of the one or more PLLs, each of the one or more clocks, and/or the like that were turned off or powered down during the previous low power state entrance.
- the controller 108 may restore any tables stored in the volatile memories that were flushed to the non-volatile memory during the previous low power state entrance.
- a new command arrives at the controller 108 .
- the host device 104 may send a doorbell to the controller 108 indicating that a command has been placed in a submission queue of the host device 104 to be fetched by the controller 108 .
- the data storage device 108 has been restored to an operational power state and is no longer FW active idle.
- the controller 108 determines if a new command has arrived. If a new command has not arrived at block 318 , then method 300 returns to block 306 . However, if a new command has arrived at block 318 , then the controller 108 stops the counter at block 320 . At block 322 , the data storage device 106 is no longer FW active idle due to receiving the new command at block 318 .
- FIG. 4 is a flow diagram illustrating a method 400 of measuring host exit latency, according to certain embodiments.
- Method 400 may be executed by a controller, such as the controller 108 of FIG. 1 .
- a controller such as the controller 108 of FIG. 1 .
- aspects of the storage system 100 of FIG. 1 may be referenced herein.
- Method 400 may be similar to method 300 , however, method 400 additionally measures a host exit latency.
- the data storage device is in a FW active idle state.
- the controller 108 starts a counter.
- the idle counter may be referred to as a “counter” herein for simplification purposes.
- the idle counter may be a generic or dedicated counter of a timer of the controller 108 .
- starting the counter may include resetting the counter to a value of zero.
- the controller determines if the counter is greater than or equal to an idle time timeout threshold, where the idle time timeout threshold is a maximum time before the controller 108 transitions the data storage device 106 into a low power mode. If the counter is greater than or equal to the idle time timeout threshold at block 406 , then the controller 108 stops the counter at block 408 .
- the controller 108 transitions the data storage device 106 into a low power mode.
- the transitioning may include one or more of powering down one or more hardware modules of the data storage device 106 while retaining memories, powering down one or more hardware modules of the data storage device 106 without retaining memories, turning off one or more clocks, turning off one or more phase lock loops (PLLs), and the like.
- PLLs phase lock loops
- the controller 108 transitions may transition the data storage device 106 to the L1.2 link state as described in FIG. 2 . Transitioning to the low power mode may also include notifying the host device 104 , by the controller 108 , that the data storage device 106 is entering the low power mode.
- the controller 108 receives a wake up indication from the host device 104 to return form the low power state (e.g., L1.2) to an operational power state (e.g., L0, L0s, or L1).
- the controller 108 may restore each of the one or more hardware modules, each of the one or more PLLs, each of the one or more clocks, and/or the like that were turned off or powered down during the previous low power state entrance.
- the controller 108 may restore any tables stored in the volatile memories that were flushed to the non-volatile memory during the previous low power state entrance.
- the controller 108 resets and restarts the counter in order to determine a host exit latency.
- the host exit latency is a time elapsed between the wake up indication and receiving a new command (or doorbell, in some examples).
- a new command arrives at the controller 108 .
- the host device 104 may send a doorbell to the controller 108 indicating that a command has been placed in a submission queue of the host device 104 to be fetched by the controller 108 .
- the controller 108 determines the host exit latency, which is equal to the counter sampled when the new command arrived at blocked 416 .
- the host exit latency for a specific host/data storage device/connection pairing may be constant or about constant, where “about” refers to a range of +/ ⁇ 5%. In other words, the host exit latency for the pairing may be considered as a constant.
- the data storage device 108 has been restored to an operational power state and is no longer FW active idle.
- the controller 108 determines if a new command has arrived. If a new command has not arrived at block 422 , then method 400 returns to block 406 . However, if a new command has arrived at block 422 , then the controller 108 stops the counter at block 424 . At block 426 , the data storage device 106 is no longer FW active idle due to receiving the new command at block 422 .
- FIG. 5 is an exemplary illustration of power saving groups (e.g., group 1, group 2, group 3, and group 4), according to certain embodiments.
- Each of the power saving groups (e.g., group 1, group 2, group 3, and group 4) includes one or more low power state entrance actions (e.g., A-K), where the controller 108 may determine one of power saving groups (e.g., group 1, group 2, group 3, or group 4) to perform when entering the low power state.
- Each of the power saving groups (e.g., group 1, group 2, group 3, and group 4) may include at least one common low power state entrance action, such that the at least one common low power state entrance action may be a required action.
- each of the low power state entrance actions (e.g., A-K) is associated with a circle, where, for exemplary purposes, an area of the circle is proportional to an amount of power saved upon completing the respective low power state entrance action. It is to be understood that the number of low power state entrance actions and the area of the circles of each low power state entrance actions are not intended to be limiting, but to provide an example of a possible embodiment.
- low power state entrance action A may include turning off part of the device NAND power
- low power state entrance action B may include lowering clock frequency
- low power state entrance action C may include turning off some clock entirely
- low power state entrance action D may include turning off local oscillator
- low power state entrance action E may include copying data from volatile range to a non-volatile range and turning off the volatile memories
- low power state entrance action F may include shutting off about 50% of the CPUs
- low power state entrance action G may include shutting off PLLs
- low power state entrance action H may include lowering the device operating voltage
- low power state entrance action I may include stopping some peripheral circuitry
- low power state entrance action J may include powering down other parts of the circuitry
- low power state entrance action K may include turning off analog circuitry power. It is to be understood that the term “about” above may refer to plus or minus 5%.
- Group 1 includes low power state entrance actions A-K.
- Group 2 includes low power state entrance actions A-C, G, and H.
- Group 3 includes low power state entrance actions A-C and K.
- Group 4 includes low power state entrance actions D, E, G, and K.
- Group 1 includes all of the low power state entrance actions and groups 2-4 includes less than all of the low power state entrance actions.
- groups 2-4 may include at least one common low power state entrance action and/or at least one different low power state entrance action.
- Group 1 may be associated with a completion wake up time of about 100 usec.
- a completion wake up time is the amount of time needed to restore a data storage device, such as the data storage device 106 of FIG. 1 , from a low power state (e.g., L1.2) to an operational power state (e.g., L0, L0s, or L1).
- Group 2 may be associated with a completion wake up time of about 50 usec.
- Group 3 may be associated with a completion wake up time of about 10 usec.
- Group 4 may be associated with a completion wake up time of about 1 usec. It is to be understood that the previously listed times are not intended to be limiting but to provide an example of a possible embodiment.
- group 1 has the most amount of power saved
- group 2 has the second most amount of power saved
- group 3 has the third most amount of power saved
- group 4 has the least amount of power saved.
- the amount of power saved may be proportional to the completion wake up time. In other words, the greater the exit latency from the low power state, the more amount of power saved.
- a controller such as the controller 108 may determine which of the power saving groups to perform during a next low power state entrance operation. For example, if the host exit latency is 60 usec, then the controller 108 may choose to perform group 2 actions during a next low power state entrance operation. The controller 108 may choose a group of the power saving groups by comparing each completion wake up time to the host exit latency and choosing a group whose completion wake up time is less than or equal to the host exit latency. The group chosen may be one that is closest to the host exit latency.
- the controller 108 may choose a group whose completion wake up time is nearest to the host exit latency. In other words, the wake up operation is more optimized when the completion wake up time is closest to the host exit latency. In some examples, the controller 108 may select a number of actions to perform out of all possible low power state entrance actions to perform based on the host exit latency and the completion wake up time of the selected number of actions to perform.
- FIG. 6 is an exemplary illustration showing exit latency vs power consumption, according to certain embodiments. Aspects of the storage system 100 of FIG. 1 may be referenced herein for exemplary purposes. Each numeral shown (e.g., 1, 2, 3, and 4) may refer to a respective group of the power saving groups (e.g., group 1, group 2, group 3, and group 4) of FIG. 5 . As shown in FIG. 6 , the amount of consumed power during a low power state of the data storage device 106 is inversely proportional to the exit latency to exit the low power state.
- group 1 has the largest exit latency from the low power state and the lowest amount of consumed power during the low power state and group 4 has the lowest exit latency from the low power state and the highest amount of consumed power during the low power state.
- Table 1 below describes various power saving options.
- FIG. 7 is an exemplary graph 700 of selecting a group of low power state actions based on a host exit latency, according to certain embodiments. Aspects of the storage system 100 of FIG. 1 may be referenced herein for exemplary purposes. Each group (i.e., group 1, group 2, group 3, and group 4) may be a respective group described in FIG. 5 and FIG. 6 . For example, group 1 of FIG. 7 may be group 1 of FIG. 5 and group 1 of FIG. 6 .
- group 1 has an exit latency (i.e., completion wake up time) of 100 ⁇ sec
- group 2 has an exit latency of 50 ⁇ sec
- group 3 has an exit latency of 10 ⁇ sec
- group 4 has an exit latency of 1 ⁇ sec.
- the host exit latency is shown to be 30 ⁇ sec.
- the controller 108 may determine to perform group 3 low power state entrance actions or group 4 low power state entrance actions. Because the exit latency of group 3 is closest to the host exit latency, the controller 108 selects the low power state entrance actions of group 3 to perform during a next low power state entrance operation.
- FIG. 8 is a flow diagram illustrating a method 800 of selecting a group of low power state actions based on a host exit latency, according to certain embodiments.
- Method 800 may be implemented by a controller, such as the controller 108 of FIG. 1 .
- a controller such as the controller 108 of FIG. 1 .
- aspects of the storage system 100 of FIG. 1 may be referenced herein.
- an idle time timeout threshold is reached.
- the controller selects a group of low power state entrance actions from a plurality of groups of low power state entrance actions (e.g., group 1, group 2, group 3, or group 4 of FIG. 5 , FIG. 6 , and FIG. 7 ) to complete based on a previously determined host exit latency for the host device/data storage device/connection pairing. It is to be understood that each group of the plurality of groups of low power state entrance actions may be fine-tuned.
- the fine-tuning is based on one or more of a number of NVMe submission queues and/or completion queues, a number of un-updated logical-to-physical (L2P) entries, a current state of the data storage device 106 (i.e., link state or power state), and the like.
- the fine-tuning may be based on a setup time or configuration time of the data storage device from the low power state.
- the setup time or configuration time of the data storage device may include initialization aspects other than the abovementioned examples.
- the controller performs the selected group of low power state entrance actions.
- the controller 108 transitions the data storage device into the low power state.
- FIG. 9 is a schematic block diagram illustrating a storage system 900 in which a data storage device 904 may function as a storage device for a host device 902 , according to certain embodiments.
- controller 906 may be used to implement method 300 of FIG. 3 , method 400 of FIG. 4 , and method 800 of FIG. 8 .
- embodiments described above may utilize aspects of the storage system 900 , such as controller 906 , counter 924 , and the like.
- aspects of the storage system 900 may be similar to the storage system 100 of FIG. 1 .
- the data storage device 904 includes the controller 906 and the NVM 934 .
- the controller 906 includes a PCIe 908 , a control path 910 , and a data path 926 .
- the command is transferred to the control path 910 .
- the control path 910 includes a FW-active detector 912 , a latency program selector 914 , one or more state tables 916 , one or more configuration tables 918 , one or more L2P tables 920 , a central processing unit (CPU) 922 , and a counter 924 .
- the FW-active detector 912 may receive a doorbell sent by the host device 902 to wake up the data storage device 904 from a low power state, such as the L1.2 link state.
- the counter 924 is used to determine when to transition the data storage device into the low power state as well as to determine the host exit latency.
- the counter 924 provides the host exit latency to the latency program selector 914 , where the latency program selector 914 determines which group of low power state entrance actions from a plurality of groups of low power state entrance actions to complete during a next low power state entrance operation.
- the latency program selector 914 may fine-tune a completion wake up time associated with each of the groups based on one or more of the one or more state tables 916 , the one or more configuration tables 918 , and the one or more L2P tables 920 .
- the one or more state tables 916 , one or more configuration tables 918 , and one or more L2P tables 920 may be a portion of a setup time or configuration time of the data storage device from the low power state. It is to be understood that the setup time or configuration time of the data storage device may include other setup tables or configuration tables other than the one or more state tables 916 , one or more configuration tables 918 , and one or more L2P tables 920 .
- the control path 910 passes instructions, generated by the CPU 922 , to the data path 926 . If data was received by the controller 906 from the host device, then the data is passed from the direct memory access (DMA) 928 to the error correction module 930 to be encoded and from the error correction module 930 to the flash interface module (FIM) 932 . The data is programmed to the NVM 934 from the FIM 932 . Likewise, if the command received is a read command, the data is read from the NVM 934 and transferred to the FIM 932 , from the FIM 932 to the error correction module 930 to be decoded, and from the error correction module 930 to the DMA 928 . The data is then transferred through the PCIe 908 to the host device 902 .
- DMA direct memory access
- FIM flash interface module
- a data storage device includes a memory device and a controller coupled to the memory device.
- the controller is configured to receive a wake up indication from a host device, where the data storage device is in a low power state, receive a doorbell from the host device indicating that a new command is in a submission queue of the host device, determine a host exit latency, where the host exit latency is a time between the wake up indication and the doorbell, and utilize the host exit latency to select a group of actions from a plurality of group of actions based on the determining, where each of the plurality of group of actions comprises a different amount of low power state entrance actions to complete during a next low power state entrance.
- the different amount of low power state entrance actions is a configurable number of low power state entrance actions. At least one of the low power state entrance actions of one or more group of actions of the plurality of group of actions is the same. At least one of the low power state entrance actions of one or more group of actions of the plurality of group of actions is different. Each group of the plurality of group of actions is associated with a completion wake up time. The completion wake up time corresponds to a time to restore the data storage device from a low power state to a normal operating power state.
- the controller is further configured to, when utilizing the host exit latency to select the group of actions from the plurality of group of actions, determine that there is at least one completion wake up time that is less than or equal to the host exit latency.
- the controller is further configured to select the group of actions that has the completion wake up time that is closest to the host exit latency when there is at least one completion wake up time that is less than or equal to the host exit latency.
- the controller is further configured to select the group of actions that has the completion wake up time that is closest to the host exit latency upon determining that all of the completion wake up times are greater than the host exit latency.
- the completion wake up time is dependent on the low power state entrance actions of the group of actions selected and a setup time of the data storage device.
- the controller is further configured to perform the low power state entrance actions of the group of actions selected upon determining that an idle time timeout threshold has been reached and transition the data storage device to the low power state.
- a data storage device in another embodiment, includes a memory device and a controller coupled to the memory device.
- the controller includes a firmware-active detector and a latency program selector coupled to the firmware-active detector.
- the firmware-active detector is configured to determine a host exit latency, where the host exit latency is a time between receiving a wake up indication from a host device and receiving a doorbell subsequent to receiving the wake up indication from the host device.
- the latency program selector is configured to select a group of actions from a plurality of group of actions based on the determining, where each of the plurality of group of actions comprises a different amount of low power state entrance actions to complete during a next low power state entrance.
- the firmware active detector is further configured to initiate a counter responsive to receiving the wake up indication from the host device and stop the counter responsive to receiving the doorbell from the host device, where the time is determined based on the counter.
- the data storage device is in a low power state during the determining.
- a first group of actions of the plurality of group of actions comprises all of the low power state entrance actions.
- Each group of actions other than the first group of actions of the plurality of group of actions comprises less than all of the low power state entrance actions.
- a number of the low power state entrance actions of each group of actions other than the first group of actions is different from each other.
- the group of actions selected has a completion wake up time less than the host exit latency.
- the completion wake up time is calculated based on a time to complete a restoration from each of the low power state entrance actions of the group of actions.
- the completion wake up time is further calculated based on a fine-tuning parameter.
- the fine-tuning parameter is dependent on a setup time of the data storage device.
- a data storage device in another embodiment, includes a means for storing data and a controller coupled to the means for storing data.
- the controller is configured to determine that an idle time timeout threshold has been reached, perform a number of low power state entrance actions based on a previously determined host exit latency, where a completion wake up time to complete a restoration from the number of low power state entrance actions is less than the previously determined host exit latency, and transition the data storage device into a low power mode.
- the number of low power state actions is dynamic.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Power Sources (AREA)
Abstract
Description
| TABLE 1 | ||||
| Time to | Time to | |||
| Power | Initiate | Complete | Depen- | |||
| # | Description | Saved | | Exit | dencies | |
| 1 | Power down a | 3 | |
100 |
1 | | # | 3 |
| hardware module | completed | |||||||
| while retaining | ||||||||
| |
||||||||
| 2 | Power down a | 3.5 | |
400 nsec | 1.2 | | # | 3 |
| hardware module | completed | |||||||
| without retaining | ||||||||
| |
||||||||
| 3 | Turn off clocks | 5 | |
80 |
200 | | # | 4 |
| completed | ||||||||
| 4 | Turn off |
4 | |
100 nsec | 600 | nsec | ||
Claims (20)
Priority Applications (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US18/218,864 US12332720B2 (en) | 2023-03-20 | 2023-07-06 | Low power optimization based upon host exit latency |
| PCT/US2024/011408 WO2024196458A1 (en) | 2023-03-20 | 2024-01-12 | Low power optimization based upon host exit latency |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US202363453288P | 2023-03-20 | 2023-03-20 | |
| US18/218,864 US12332720B2 (en) | 2023-03-20 | 2023-07-06 | Low power optimization based upon host exit latency |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20240319779A1 US20240319779A1 (en) | 2024-09-26 |
| US12332720B2 true US12332720B2 (en) | 2025-06-17 |
Family
ID=92803548
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US18/218,864 Active 2043-09-05 US12332720B2 (en) | 2023-03-20 | 2023-07-06 | Low power optimization based upon host exit latency |
Country Status (2)
| Country | Link |
|---|---|
| US (1) | US12332720B2 (en) |
| WO (1) | WO2024196458A1 (en) |
Citations (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US5650762A (en) * | 1994-03-29 | 1997-07-22 | Nohmi Bosai Ltd. | Fire alarm system |
| US20110154090A1 (en) * | 2009-12-22 | 2011-06-23 | Dixon Martin G | Controlling Time Stamp Counter (TSC) Offsets For Mulitple Cores And Threads |
| US20120096293A1 (en) * | 2010-10-19 | 2012-04-19 | International Business Machines Corporation | Directed Resource Folding for Power Management |
| US20130290606A1 (en) | 2012-04-30 | 2013-10-31 | Apple Inc. | Power management for a system having non-volatile memory |
| US9116694B2 (en) | 2012-09-26 | 2015-08-25 | Intel Corporation | Efficient low power exit sequence for peripheral devices |
| US9417961B2 (en) | 2014-11-18 | 2016-08-16 | HGST Netherlands B.V. | Resource allocation and deallocation for power management in devices |
| US20170168853A1 (en) | 2015-12-14 | 2017-06-15 | Qualcomm Incorporated | Dynamic predictive wake-up techniques |
| US20190086995A1 (en) | 2017-09-21 | 2019-03-21 | Toshiba Memory Corporation | Memory system, method of controlling memory system, and controller circuit |
| US20190235612A1 (en) | 2018-12-12 | 2019-08-01 | Ang Li | Securely providing multiple wake-up time options for pci express |
| CN111142653A (en) | 2019-12-26 | 2020-05-12 | 江苏芯盛智能科技有限公司 | PCIe device low-power-consumption control method and device and electronic device |
| US10732871B2 (en) | 2017-11-01 | 2020-08-04 | Western Digital Technologies, Inc. | Storage device and method for adaptive sleep mode transitions |
| US20210018975A1 (en) | 2019-07-15 | 2021-01-21 | Micron Technology, Inc. | Storage system deep idle power mode |
| US20220327074A1 (en) | 2021-04-13 | 2022-10-13 | SK Hynix Inc. | PERIPHERAL COMPONENT INTERCONNECT EXPRESS (PCIe) SYSTEM AND METHOD OF OPERATING THE SAME |
| US20220326885A1 (en) | 2021-04-13 | 2022-10-13 | SK Hynix Inc. | Peripheral component interconnect express (pcie) interface system and method of operating the same |
-
2023
- 2023-07-06 US US18/218,864 patent/US12332720B2/en active Active
-
2024
- 2024-01-12 WO PCT/US2024/011408 patent/WO2024196458A1/en not_active Ceased
Patent Citations (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US5650762A (en) * | 1994-03-29 | 1997-07-22 | Nohmi Bosai Ltd. | Fire alarm system |
| US20110154090A1 (en) * | 2009-12-22 | 2011-06-23 | Dixon Martin G | Controlling Time Stamp Counter (TSC) Offsets For Mulitple Cores And Threads |
| US20120096293A1 (en) * | 2010-10-19 | 2012-04-19 | International Business Machines Corporation | Directed Resource Folding for Power Management |
| US20130290606A1 (en) | 2012-04-30 | 2013-10-31 | Apple Inc. | Power management for a system having non-volatile memory |
| US9116694B2 (en) | 2012-09-26 | 2015-08-25 | Intel Corporation | Efficient low power exit sequence for peripheral devices |
| US9417961B2 (en) | 2014-11-18 | 2016-08-16 | HGST Netherlands B.V. | Resource allocation and deallocation for power management in devices |
| US20170168853A1 (en) | 2015-12-14 | 2017-06-15 | Qualcomm Incorporated | Dynamic predictive wake-up techniques |
| US20190086995A1 (en) | 2017-09-21 | 2019-03-21 | Toshiba Memory Corporation | Memory system, method of controlling memory system, and controller circuit |
| US10732871B2 (en) | 2017-11-01 | 2020-08-04 | Western Digital Technologies, Inc. | Storage device and method for adaptive sleep mode transitions |
| US20190235612A1 (en) | 2018-12-12 | 2019-08-01 | Ang Li | Securely providing multiple wake-up time options for pci express |
| US20210018975A1 (en) | 2019-07-15 | 2021-01-21 | Micron Technology, Inc. | Storage system deep idle power mode |
| CN111142653A (en) | 2019-12-26 | 2020-05-12 | 江苏芯盛智能科技有限公司 | PCIe device low-power-consumption control method and device and electronic device |
| US20220327074A1 (en) | 2021-04-13 | 2022-10-13 | SK Hynix Inc. | PERIPHERAL COMPONENT INTERCONNECT EXPRESS (PCIe) SYSTEM AND METHOD OF OPERATING THE SAME |
| US20220326885A1 (en) | 2021-04-13 | 2022-10-13 | SK Hynix Inc. | Peripheral component interconnect express (pcie) interface system and method of operating the same |
Non-Patent Citations (1)
| Title |
|---|
| International Search Report and Written Opinion for Application No. PCT/US2024/011408 dated May 17, 2024. |
Also Published As
| Publication number | Publication date |
|---|---|
| WO2024196458A1 (en) | 2024-09-26 |
| US20240319779A1 (en) | 2024-09-26 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11640251B2 (en) | Early transition to low power mode for data storage devices | |
| US11531473B1 (en) | Selective HMB backup in NVM for low power mode | |
| US11537193B2 (en) | Shaped and optimized power cycles | |
| US11543996B2 (en) | Systems and methods for power management in a data storage device | |
| US12166505B2 (en) | Partial speed changes to improve in-order transfer | |
| US11966613B2 (en) | Selective device power state recovery method | |
| US11853603B2 (en) | Host memory buffer cache management | |
| US12045501B2 (en) | Sideband information over host interface considering link states | |
| US20210294408A1 (en) | Power allocation management for external storage | |
| US11763040B2 (en) | Enhanced D3-cold and faster recovery | |
| US11137823B1 (en) | Systems and methods for power management in a data storage device | |
| US20240427407A1 (en) | Data Storage With Real Time Dynamic Clock Frequency Control | |
| US12265738B2 (en) | Reordering commands to optimize execution of the commands | |
| US12360686B2 (en) | Highly efficient SSD power management through PCIe data rate modulation | |
| US12112048B2 (en) | Adaptive tuning of memory device clock rates based on dynamic parameters | |
| US12332720B2 (en) | Low power optimization based upon host exit latency | |
| US11640252B2 (en) | Idle-power mitigation circuit | |
| US11755238B2 (en) | Continuous NAND data-transfer using filling rate evaluation approach | |
| US12474841B1 (en) | Pipeline optimization in low power mode using PMR | |
| US11893253B1 (en) | Dynamic TD-PPM state and die mapping in multi-NAND channels | |
| US11726911B2 (en) | NVMe persistent memory region quick copy | |
| WO2024054274A1 (en) | Asymmetric time division peak power management (td-ppm) timing windows |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| AS | Assignment |
Owner name: WESTERN DIGITAL TECHNOLOGIES, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ELMALEH, NISSIM;SEGEV, AMIR;BENISTY, SHAY;SIGNING DATES FROM 20230309 TO 20230316;REEL/FRAME:064197/0745 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| AS | Assignment |
Owner name: JPMORGAN CHASE BANK, N.A., ILLINOIS Free format text: PATENT COLLATERAL AGREEMENT - DDTL;ASSIGNOR:WESTERN DIGITAL TECHNOLOGIES, INC.;REEL/FRAME:065657/0158 Effective date: 20231117 Owner name: JPMORGAN CHASE BANK, N.A., ILLINOIS Free format text: PATENT COLLATERAL AGREEMENT- A&R;ASSIGNOR:WESTERN DIGITAL TECHNOLOGIES, INC.;REEL/FRAME:065656/0649 Effective date: 20231117 |
|
| AS | Assignment |
Owner name: SANDISK TECHNOLOGIES, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:WESTERN DIGITAL TECHNOLOGIES, INC.;REEL/FRAME:067567/0682 Effective date: 20240503 Owner name: SANDISK TECHNOLOGIES, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNOR'S INTEREST;ASSIGNOR:WESTERN DIGITAL TECHNOLOGIES, INC.;REEL/FRAME:067567/0682 Effective date: 20240503 |
|
| AS | Assignment |
Owner name: SANDISK TECHNOLOGIES, INC., CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:SANDISK TECHNOLOGIES, INC.;REEL/FRAME:067982/0032 Effective date: 20240621 |
|
| AS | Assignment |
Owner name: JPMORGAN CHASE BANK, N.A., AS THE AGENT, ILLINOIS Free format text: PATENT COLLATERAL AGREEMENT;ASSIGNOR:SANDISK TECHNOLOGIES, INC.;REEL/FRAME:068762/0494 Effective date: 20240820 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| AS | Assignment |
Owner name: SANDISK TECHNOLOGIES, INC., CALIFORNIA Free format text: PARTIAL RELEASE OF SECURITY INTERESTS;ASSIGNOR:JPMORGAN CHASE BANK, N.A., AS AGENT;REEL/FRAME:071382/0001 Effective date: 20250424 Owner name: JPMORGAN CHASE BANK, N.A., AS COLLATERAL AGENT, ILLINOIS Free format text: SECURITY AGREEMENT;ASSIGNOR:SANDISK TECHNOLOGIES, INC.;REEL/FRAME:071050/0001 Effective date: 20250424 |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |