US20200293191A1 - Electronic system with storage management mechanism and method of operation thereof - Google Patents
Electronic system with storage management mechanism and method of operation thereof Download PDFInfo
- Publication number
- US20200293191A1 US20200293191A1 US16/887,341 US202016887341A US2020293191A1 US 20200293191 A1 US20200293191 A1 US 20200293191A1 US 202016887341 A US202016887341 A US 202016887341A US 2020293191 A1 US2020293191 A1 US 2020293191A1
- Authority
- US
- United States
- Prior art keywords
- key value
- volatile memory
- user data
- memory array
- interface circuit
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0602—Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
- G06F3/061—Improving I/O performance
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F12/00—Accessing, addressing or allocating within memory systems or architectures
- G06F12/02—Addressing or allocation; Relocation
- G06F12/0223—User address space allocation, e.g. contiguous or non contiguous base addressing
- G06F12/023—Free address space management
- G06F12/0238—Memory management in non-volatile memory, e.g. resistive RAM or ferroelectric memory
- G06F12/0246—Memory management in non-volatile memory, e.g. resistive RAM or ferroelectric memory in block erasable memory, e.g. flash memory
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0602—Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
- G06F3/0614—Improving the reliability of storage systems
- G06F3/0616—Improving the reliability of storage systems in relation to life time, e.g. increasing Mean Time Between Failures [MTBF]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0628—Interfaces specially adapted for storage systems making use of a particular technique
- G06F3/0655—Vertical data movement, i.e. input-output transfer; data movement between one or more hosts and one or more storage devices
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0628—Interfaces specially adapted for storage systems making use of a particular technique
- G06F3/0655—Vertical data movement, i.e. input-output transfer; data movement between one or more hosts and one or more storage devices
- G06F3/0659—Command handling arrangements, e.g. command buffers, queues, command scheduling
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0668—Interfaces specially adapted for storage systems adopting a particular infrastructure
- G06F3/0671—In-line storage system
- G06F3/0673—Single storage device
- G06F3/0679—Non-volatile semiconductor memory device, e.g. flash memory, one time programmable memory [OTP]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0668—Interfaces specially adapted for storage systems adopting a particular infrastructure
- G06F3/0671—In-line storage system
- G06F3/0683—Plurality of storage devices
- G06F3/0688—Non-volatile semiconductor memory arrays
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2212/00—Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
- G06F2212/10—Providing a specific technical effect
- G06F2212/1016—Performance improvement
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2212/00—Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
- G06F2212/10—Providing a specific technical effect
- G06F2212/1032—Reliability improvement, data loss prevention, degraded operation etc
- G06F2212/1036—Life time enhancement
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2212/00—Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
- G06F2212/72—Details relating to flash memory management
- G06F2212/7205—Cleaning, compaction, garbage collection, erase control
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2212/00—Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
- G06F2212/72—Details relating to flash memory management
- G06F2212/7211—Wear leveling
Definitions
- An embodiment of the present invention relates generally to an electronic system, and more particularly to a system for data storage functions for electronic systems.
- Big Data refers to huge volumes of data having a diverse and complex nature that is captured very quickly.
- Data is exploding at an astonishing rate: up until 2003 approximately 5 exabytes of data was created, 90% of the data was structured data that was gathered through rational database management systems (RDBMS), and the rest was unstructured data.
- RDBMS rational database management systems
- unstructured data is growing faster than structured data. Predictions are made that data will grow to 5 zettabytes by 2015. Unstructured data is growing faster than structured data.
- IDC International Data Corporation
- Storage cell wear issues can include a limited number of reads allowed between each write, a limited number of writes before the cell becomes inoperative, erasures count as a write, erase required prior to writing, erasure can only be performed in larger capacity than that of write, and Flash memory can include imperfect cells, which need to be managed as bad blocks.
- the storage cell wear issues require constant maintenance as well as over-provisioning, which reserves additional capacity of the storage cells that is not available in the rated capacity available to the user. In order to accommodate the requirement of the Big Data databases and their uses, large capacity storage devices must be used. The typical uses of the Big Data can very quickly cause an SSD to experience degraded performance due to the background maintenance required to protect the data.
- An embodiment of the present invention provides an electronic system includes: a key value storage device, configured to transfer user data, the key value storage device including: a non-volatile memory array accessed by a key value address, an interface circuit, coupled to the non-volatile memory array, configured to receive a key value transfer command, a volatile memory, coupled to the interface circuit and the non-volatile memory array, configured to reduce a number of copies of the user data in the non-volatile memory array, and a device processor, coupled to the interface circuit, configured to manage the non-volatile memory array, the volatile memory, and the interface circuit by a key value index tree, including a key value translation block, to access the user data.
- a key value storage device configured to transfer user data
- the key value storage device including: a non-volatile memory array accessed by a key value address, an interface circuit, coupled to the non-volatile memory array, configured to receive a key value transfer command, a volatile memory, coupled to the interface circuit and the non-volatile memory
- An embodiment of the present invention provides a method of operation of an electronic system including: transferring user data by a key value storage device including: receiving a key value transfer command through an interface circuit includes connecting a device coupling structure to the interface circuit, selecting a non-volatile memory array, a volatile memory, or a combination thereof for reducing a number of copies of the user data in the non-volatile memory array accessed by a key value address, and managing a key value index tree, by a device processor, for transferring the user data between the interface circuit, the volatile memory, the non-volatile memory array, or a combination thereof includes accessing a key value translation block.
- FIG. 1 is an architectural block diagram of an electronic system with storage management mechanism in an embodiment.
- FIG. 2 is a hardware block diagram of a flash defined key value store device in an embodiment.
- FIG. 3 is a software functional block diagram of an electronic system with storage management mechanism in an embodiment.
- FIG. 4 is a functional block diagram of an operating software of a key value storage device in an embodiment.
- FIG. 5 is a flow chart of a MVCC operation in an embodiment.
- FIG. 6 is a flow chart of a method of operation of an electronic system in a further embodiment of the present invention.
- Various embodiments provide a data management mechanism for Big Data databases that can simplify the distributed input/output (I/O) interface and maximize the execution efficiency of the electronic system by establishing a less burdensome I/O architecture.
- the data management mechanism can be configured to process the data used in a user program without utilizing the I/O storage stack structure provided in the operating system.
- the execution of the data management mechanism can be executed at a user program level to provide efficient access to remote storage, local storage, and memory and provide the maximum program execution efficiency.
- Various embodiments provide a data management mechanism for Big Data databases by omitting the customary I/O storage stacks required in operating system architectures.
- the data management mechanism also allows manipulation of the data based on a key value assigned to the data rather than a logical block address.
- the data management mechanism can operate in a host central processing unit (CPU) system memory, without making I/O calls through the operating system.
- CPU central processing unit
- the execution of the data management mechanism from the user defined space of the system memory can provide higher priority operation, while managing the operational characteristics of a key value storage device. This combination can increase the efficiency and operational performance of the data management mechanism.
- module can include software, hardware, or a combination thereof in an embodiment of the present invention in accordance with the context in which the term is used.
- the software can be machine code, firmware, embedded code, and application software.
- the hardware can be circuitry, processor, computer, integrated circuit, integrated circuit cores, a pressure sensor, an inertial sensor, a microelectromechanical system (MEMS), passive devices, or a combination thereof. Further, if a module is written in the apparatus claims section below, the modules are deemed to include hardware circuitry for the purposes and the scope of apparatus claims.
- the term “unit” referred to herein is a circuit formed of hardware components or hardware state machines used for specific functions.
- the “unit” can be for timing critical functions and does not necessarily include software functions or support.
- the term “customary I/O storage stacks” referred to herein are memory segments within the host system memory that are used to hold address pointers and interrupt flags for input/output (I/O) operations performed by the operating system.
- the customary I/O storage stacks insert a latency to the I/O in order to load the memory with interrupt pointers to service the transfer of data to or from the I/O device based on servicing the interrupt associated with the transfer.
- the electronic system 100 includes a key value storage device 102 , examples of such devices can include solid state storage devices (SSSD) based on Flash memory, magnetic random access memory (MRAM), Resistive Random Access Memory (RRAM), and Phase Change Memory (PCM), as well as hybrid disk storage devices, and key value storage adapters, which are addressed by way of a key value for identifying file structures.
- the key value storage device 102 can be a non-volatile data storage mechanism that stores and accesses user data by way of key value addressing.
- the key value storage device 102 can be used for data processing in a computing environment for managing big data databases.
- a device coupling structure 104 can link the key value storage device 102 to a host processor 106 .
- the device coupling structure 104 can be an input/output interface structure connected between the host processor 106 and the key value storage device 102 .
- the device coupling structure 104 can include a peripheral computer interface express (PCI-e), fiber channel (FC), small computer system interface (SCSI), or serial attached SCSI (SAS), and the host's memory channel interface.
- the device coupling structure 104 can be implemented as a memory bus for host internal application of the electronic system 100 .
- the device coupling structure 104 can provide local or remote connectivity between the host processor 106 and the key value storage device 102 . The connection of the device coupling structure 104 between the key value storage device 102 and the host processor 106 is performed in a manner that meets the specification of the device coupling structure 104 .
- the device coupling structure 104 can provide direct coupling and communication to the key value storage device 102 as well as standard storage devices (not shown). It is understood that the communication with standard storage devices requires the I/O calls provided by the operating system and may utilize a logical block address (LBA) architecture. Since the key value storage device 102 does not use the LBA architecture, the communication from the host processor 106 must be structured in a key value based linear storage architecture.
- LBA logical block address
- the host processor 106 can be a general purpose computer, a computer network, a server, a storage processor, GPU, ASIC, FPGA, PLD, or the like.
- the host processor 106 can include a system memory 108 , a hardware interface controller 110 , and a control processor (not shown).
- the host processor 106 can execute a key value device application program interface (API) 112 and a key value device driver 114 . It is understood that the system memory 108 can be outside the physical bounds of the host processor 106 .
- the hardware interface controller 110 is a hardware adapter that controls the electrical and timing parameters associated with the device coupling structure 104 .
- the key value device API 112 provides an application interface that allows user applications to access the key value storage device 102 without having a description of the operational characteristics of the key value storage device 102 .
- the key value device API 112 is a program that translates standard Application instructions into a form that is recognized by the key value device driver 114 .
- the key value device API 112 passes the application requirements to the key value device driver 114 for conversion to interface commands that are recognized by the key value storage device 102 .
- the key value device driver 114 can communicate with a physical interface conditioning block 116 , such as Non-Volatile Memory Express (NVMe) or the Linux kernel virtual file system (sysfs).
- the physical interface conditioning block 116 can provide formatted commands to a universal Input/Output (UIO) kernel driver 118 , such as an NVMe UIO kernel driver or a UIO framework with internal kernel functions provided by another block.
- the physical interface conditioning block 116 is a program that can add required information used by the hardware interface controller 110 to access the device coupling structure 104 .
- the UIO kernel driver 118 is a program that can manage and regulate the interrupt timing for the hardware interface controller 110 .
- the UIO kernel driver 118 can communicate through the hardware interface controller 110 to the device coupling structure 104 and on to the key value storage device 102 .
- the physical interface conditioning block 116 , the UIO kernel driver 118 , the hardware interface controller 110 , and the device coupling structure 104 provide a communication transfer mechanism between the key value device driver 114 and the key value storage device 102 .
- the key value storage device 102 can include a hardware interface 120 for coupling to the device coupling structure 104 .
- the hardware interface 120 can be hardware structure that manages the communication and timing between the key value storage device 102 and the device coupling structure 104 .
- the hardware interface 120 can be controlled by a device processor 122 , which can implement a key value flash translation layer (KVFTL), for managing the geometry of a non-volatile memory array 124 , and support functions 126 , which can include garbage collection, wear leveling, bad block management, and command execution.
- KVFTL key value flash translation layer
- the key value storage device 102 can provide the support functions 126 for expiry, multi-version concurrency control (MVCC), and available storage collection and management.
- the support functions 126 can include hardware and microcode for supporting the execution of the KVFTL.
- the hardware associated with the support functions 126 can include memory, register files, sequential state machines, multiplexers, error checking devices, and power monitoring devices.
- the device processor 122 can include a microprocessor, an embedded processor, a sequential state machine, or a microcomputer.
- the non-volatile memory array 124 can be an array of non-volatile memory devices including flash memory, magnetoresistive random access memory, Ferroelectric random access memory, or phase-change random access memory.
- the key value storage device 102 can provide increased performance by not utilizing the standard block level architecture.
- the user applications that access the key value storage device 102 are not required to maintain indexing structures in order to search for specific data.
- a key value that is unique to the data is stored by the user application to locate specific data structures and all indexing functions are performed by the key value storage device 102 .
- the support functions 126 of the key value storage device 102 , can manage the non-volatile memory array 124 as a linear addressable space with bad blocks masked out.
- the device processor 122 can maintain a key value index list to facilitate the indexing of user data without support from the host processor 106 .
- the electronic system 100 can enhance performance of the storage management by reducing the I/O overhead in the host processor 106 by removing the concept of a logical block address in favor of a key value that can define the user data while leaving the identification of the location and extent of the user data to the key value storage device 102 .
- the device processor of the key value storage device 102 can maintain a key value index list that can accommodate expiry, MVCC, and read/write access.
- FIG. 2 therein is shown a hardware block diagram of a flash defined key value store device in an embodiment.
- the hardware block diagram of the flash defined key value store device 102 depicts the interface circuit 120 , such as a peripheral computer interface express (PCI-e), fiber channel (FC), small computer system interface (SCSI), serial attached SCSI (SAS), or the like, that can deliver the key value storage commands and data through the device coupling structure 104 of FIG. 1 .
- PCI-e peripheral computer interface express
- FC fiber channel
- SCSI small computer system interface
- SAS serial attached SCSI
- the device processor 122 such as a bit slice processor, a microprocessor, a hardware sequential state machine, an embedded microprocessor, an FPGA, an ASIC, a microcomputer, or the like, can be coupled to the interface circuit 120 in order to receive key value commands, prepare the interface circuit 120 for routing user data, manage stored data, or providing status to the host processor 106 of FIG. 1 .
- the device processor 122 can be coupled to a device memory 202 , which can include volatile memory, non-volatile memory, or a combination thereof. It is understood that the interface circuit 120 , the device processor 122 , and the device memory 202 can be integrated into a single component, such as a system-on-a-chip (SOC), a gate array, a custom logic chip, a field programmable gate array, or an application specific integrated circuit (ASIC).
- SOC system-on-a-chip
- ASIC application specific integrated circuit
- the device memory 202 can store the key value associated with the user data received through the interface circuit 120 .
- a key value index list 204 can be stored in the device memory 202 and maintained by the device processor 122 for all of the support functions 126 of FIG. 1 .
- the key value index list 204 will link the key value with a physical location of the user data.
- a volatile memory 206 can be coupled to the interface circuit 120 .
- the volatile memory 206 can be used as a data cache and can be accessed by the device processor 122 through the interface circuit 120 .
- a non-volatile memory array 124 can be coupled to the volatile memory 206 and the interface circuit 120 in order to provide reduced latency access to user data written to the volatile memory 206 , the non-volatile memory array 124 , or both.
- the interface circuit 120 can be configured to concurrently write both the volatile memory 206 and the non-volatile memory array 124 in order to support the MVCC function.
- the device processor 122 can manipulate the user data in the volatile memory 206 and the non-volatile memory array 124 through the interface circuit 120 .
- the interface circuit 120 can actively transfer user data to or from one of the volatile memory 206 and the non-volatile memory array 124 while the device processor 122 can concurrently access the other of the volatile memory 206 and the non-volatile memory array 124 .
- the device processor 122 can execute the support functions 126 of FIG. 1 while the interface circuit 120 is actively transferring user data to the volatile memory 206 or the non-volatile memory array 124 without adding latency to the transfer of the user data.
- the non-volatile memory array 124 can include a number of flash memory integrated circuits coupled in the array format on a key value device circuit board 208 .
- an embodiment of the key value storage device 102 can reduce the write amplification factor as compared to a logical block address (LBA) device.
- the support functions 126 can produce fewer copies of the user data during garbage collection and wear leveling.
- the support functions 126 can take advantage of the device architecture by utilizing the volatile memory 206 as a bridge device to hold temporary copies of the key value data blocks, rather than making multiple copies within the non-volatile memory array 124 . This technique can reduce the number of write operations applied to the non-volatile memory array 124 and reduce the burden of the garbage collection because the volatile memory 206 can be used without generating additional cell utilization within the non-volatile memory array 124 .
- FIG. 3 therein is shown an architectural block diagram of an electronic system 301 with storage management mechanism in an embodiment.
- the architectural block diagram of the electronic system 301 can depict a user space 302 and an operating system kernel space 304 within the system memory 108 of FIG. 1 .
- the user space 302 can be utilized by user applications for preparing user data 306 for transfer to the key value storage device 102 .
- the key value device API 112 can pass the user data 306 to the key value device driver 114 for configuration to an I/O compatible format.
- the key value device driver 114 can access an I/O description 308 and an I/O configuration 310 in order to correctly format the user data 306 for the UIO kernel driver 118 for direct transfer to the key value storage device 102 .
- the electronic system 301 can reduce the latency of the transfer of the user data 306 by bypassing the system calls 312 and the I/O storage stack 314 that are normally associated with I/O operations.
- the system calls 312 and the I/O storage stack 314 operate in the operating system kernel space 304 of the system memory 108 .
- the operating system kernel space 304 is normally utilized during system interrupts, which can add operating system latency to the I/O operation.
- the key value storage device 102 can monitor the indexing of the user data 306 based on the unique key value associated with the user data 306 , which can reduce the amount of the system memory 108 and latency needed to save the information from the I/O storage stack 314 . By reducing both the latency and the amount of the system memory 108 utilized by the I/O storage stack 314 , the electronic system 301 can enhance performance and reduce the power required for the transfer of the user data 306 .
- the UIO kernel driver 118 can convert the key value transfer command 316 to a key value transfer 318 .
- the key value transfer 318 can be a key value command for the key value storage device 102 for a transfer of the user data 306 to satisfy the key value transfer command 316 .
- the device processor 122 of FIG. 2 can interpret the key value transfer 318 in order to manage the transaction of the user data 306 with the volatile memory 206 of FIG. 2 , the non-volatile memory array 124 of FIG. 1 , or both concurrently.
- the interface circuit 120 can be configured to send or receive the user data 306 .
- the user data 306 can be transferred between the interface circuit 120 and the volatile memory 206 , the non-volatile memory array 124 , or both concurrently, in the case of a multi-version concurrency control (MVCC) operation.
- MVCC multi-version concurrency control
- FIG. 4 therein is shown an operational block diagram of a key value storage device 401 in an embodiment.
- the operational block diagram of the key value storage device 401 depicts the support functions 126 executed by the device processor 122 of FIG. 2 including a command processing queue 402 , a key value index tree 404 , and a device real-time operating system (RTOS) 406 .
- RTOS device real-time operating system
- the command processing queue 402 can manage queued commands from the key value device driver 114 , while coordinating access of the key value index tree 404 by the device processor 122 of FIG. 1 .
- the command processing queue 402 can configure the device coupling structure 104 for direct communication to the key value device driver 114 by addressing the requests and responses to the address of the key value device driver 114 in user space 302 of FIG. 2 .
- the command processing queue 402 can set-up the interface circuit 120 of FIG. 2 to optimize the access for the device processor 122 without adding latency to the transfer of the user data 306 of FIG. 3 by implementing a direct communication to the key value device driver 114 of FIG. 3 and bypassing the system calls 312 of FIG. 3 and the storage stack 314 of FIG. 3 the electronic system 100 of FIG. 1 .
- the command processing queue 402 can significantly reduces the latency between the key value transfers 318 because the transfers occur as operations in user space 302 that do not wait for system interrupt processing to occur.
- the key value index tree 404 can include a garbage collection block 408 , a wear leveling block 410 , a write operation monitor 412 , a read operation monitor 414 , a device monitor 416 , a key value translation block 418 , and a bad block management module 420 for supporting the non-volatile memory array 124 of FIG. 1 .
- the key value index tree 404 can maintain the operational readiness of the key value storage device 102 of FIG. 1 without adding latency to the transfer of the user data 306 .
- the key value index tree 404 can be managed by the device processor 122 in order to maintain the current location of the user data 306 contained in the volatile memory 206 of FIG. 2 and the non-volatile memory array 124 .
- the device processor 122 can establish rules governing the execution of the garbage collection block 408 and the wear leveling block 410 in order to preserve multiple versions of the user data 306 associated with the same key.
- the garbage collection block 408 can reduce the write amplification factor by utilizing a portion of the volatile memory 206 of FIG. 2 to store still valid data while performing an erase of expired or invalid blocks of the user data 306 .
- the reduction in the write amplification factor can extend the operable life of the key value storage device 102 , well beyond that of a solid state drive (SSD), while maintaining the high performance and reduced latency that makes the key value storage device 102 desirable.
- SSD solid state drive
- Any of the valid data displaced due to the garbage collection block 408 is readily available for transfer to the interface circuit 120 if required. This process is necessary because the user data 306 can be written in 4 KB to 8 KB pages but an erase can impact (for example) 64 such pages at a time.
- the wear leveling block 410 can manage the distribution of the writes of the user data 306 , to keep the charge distribution in the non-volatile memory array 124 in balance, and prevent exceeding the lifetime write limit of cells in any area of the non-volatile memory array 124 .
- the wear leveling block 410 can configure over-provisioned portions of the non-volatile memory array 124 for use when replacing worn or faulty blocks, as part of the wear leveling process. Since the write amplification factor is reduced for the key value storage device 102 , the wear leveling block 410 can be invoked less frequently than in products based on logical block address (LBA) access.
- LBA logical block address
- the write operation monitor 412 can configure the interface circuit 120 to communicate directly to the key value device driver 114 to provide a low latency path to the volatile memory 206 , the non-volatile memory array 124 , or a combination thereof.
- the write operation monitor 412 can receive a key value pair for the user data 306 associated with the write command.
- the key value pair can be stored in the key value index list 204 of FIG. 2 in order to access the user data 306 at a later time.
- the write operation monitor 412 can configure the status to be returned by the interface circuit 120 .
- the read operation monitor 412 can configure the interface circuit 120 after verifying the key value, received for the read command, is valid.
- the read operation monitor can configure the interface circuit 120 to communicate directly to the key value device driver 114 to provide a low latency path for the user data 306 read from the volatile memory 206 , the non-volatile memory array 124 , or a combination thereof.
- the read operation monitor 414 can configure the status to be returned by the interface circuit 120 .
- the device monitor 416 can detect error or warning conditions associated with the general operation of the key value storage device 102 , such as data errors, timeout of operations, hardware warnings, or command exceptions.
- the device monitor 416 can provide recovery strategies for recoverable errors or warnings, provide status to the interface circuit 120 for unrecoverable errors, and can transfer the contents of the volatile memory 206 to a reserved portion of the non-volatile memory array 124 upon notification of power loss or shut down.
- the recovery strategies can include retries, data error correction, retransmission, or emergency back-up to the reserved portion of the non-volatile memory array 124 .
- the key value storage device 102 can include a charge storage device or small battery in order to maintain control during the power-down processing.
- the key value translation block 418 can receive key value pairs from the write operation monitor 412 and add the key value to the key value index list 204 of FIG. 2 with the physical location in the volatile memory 206 or the non-volatile memory array 124 .
- the “key” portion of the key value pair is used to index the physical location of the “value” portion, which can be the user data 306 or a pointer to the physical location of the user data 306 .
- the key value translation block 418 can verify the key value received with the read command is located in the key value index list 204 and provide the physical address of the user data 306 to the read operation monitor 414 for configuration of the interface circuit 120 and transfer the user data 306 .
- the bad block management module 420 can tag all blocks within the non-volatile memory array 124 that contains a page with an unrecoverable error.
- the bad block management module 420 can receive error indications from the garbage collection block 408 , the wear leveling block 410 , the write operation monitor 412 , the read operation monitor 414 , the device monitor 416 , or a combination thereof.
- the device RTOS 406 can enable the device processor 122 to provide system level supervision and timing prompts for the support functions 126 .
- the device RTOS 406 can provide system level timing functions, diagnostic capabilities, system recovery processes and reset management.
- the write operation monitor 412 , the read operation monitor 414 , the device monitor 416 , and the key value translation module 418 could be combined in a single module or further partitioned to provide other more detailed module functions.
- the support functions 126 of the key value storage device 102 can provide low latency access to the user data 306 stored in the volatile memory 206 or the non-volatile memory array 124 with a simplified command interface structure without using logical block address architecture. This can reduce the use of the system memory 108 of FIG. 1 while increasing performance due to bypassing the system calls 312 of FIG. 3 and the storage stack 314 of FIG. 3 , which are used to support the LBA architecture.
- FIG. 5 therein is shown a flow chart of a MVCC operation 501 in an embodiment.
- the flow chart of the MVCC operation 501 depicts receiving an update command at step 502 , such as a write command, updating an existing file, received through the interface circuit 120 of FIG. 2 .
- the device processor 122 of FIG. 2 can identify the key value from the key value transfer command 316 of FIG. 3 .
- the device processor 122 then proceeds to identify the key at step 504 in the key value index list 204 of FIG. 2 to determine whether the user data 306 of FIG. 3 already exists in the electronic system 100 and to identify that the update command is valid. Upon identifying that the user data 306 does exist, the device processor 122 can allocate a location for the new key value pair at step 506 , which can be added to the key value index list 204 .
- the device processor 122 can update metadata in the new key value pair location at step 508 .
- the key value index list 204 can include, for each key value entry, a link pointer to identify previous versions of the key value in order to provide a history of the value associated with the key. Since the non-volatile memory array 124 of FIG. 2 cannot perform an overwrite of the existing location, the older version of user data 306 associated with the key remains available. 1 during background support operations. In multi-host environments, this feature can allow access to multiple versions of the original data set and allow each host to perform updates.
- the key can identify a data set and the value can indicate a version number of the data set.
- the metadata can include the version number and pointers to the physical location of the indicated version.
- the device processor can assign limits to the number of different values associated with a single key. It has been discovered that the electronic system 100 can provide a back-up process to restore a data set to a previous version with a single key value read command.
- the device processor 122 can update the key value index list 204 to include the new key value pair at step 510 .
- the key value index list 204 can include the link pointer to the previous key value pair location, the key provided from the command, and metadata, which includes the value associated with the key and physical pointers to the location of the user data 306 associated with the data set.
- the electronic system 100 can support multiple versions of the same data set concurrently. This feature is further supported by the use of the volatile memory 206 of FIG. 2 with the garbage collection block 408 and the wear leveling block 410 , which can substantially reduce the number of times the same data set is replicated. By transferring any active data in an erase block to the volatile memory 206 , it is not necessary to consume more of the non-volatile memory array 124 for the support functions 126 .
- the method 600 includes: transferring user data between the key value device driver 114 of FIG. 1 and the interface circuit 120 of FIG. 1 , without experiencing the normal delay caused by the I/O storage stack 314 of FIG. 3 , using a key value storage device 102 of FIG. 1 .
- the method 600 includes: receiving a key value transfer command 316 of FIG. 3 through the interface circuit 120 includes connecting a device coupling structure 104 of FIG. 1 to the interface circuit 120 in a block 602 , selecting a non-volatile memory array 124 of FIG. 1 , a volatile memory 206 of FIG.
- the resulting method, process, apparatus, device, product, and/or system is straightforward, cost-effective, uncomplicated, highly versatile, accurate, sensitive, and effective, and can be implemented by adapting known components for ready, efficient, and economical manufacturing, application, and utilization.
- Another important aspect of an embodiment of the present invention is that it valuably supports and services the historical trend of reducing costs, simplifying systems, and increasing performance.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Techniques For Improving Reliability Of Storages (AREA)
Abstract
Description
- This application is a Continuation of co-pending U.S. patent application Ser. No. 16/397,543 filed Apr. 29, 2019. U.S. application Ser. No. 16/397,543 filed Apr. 29, 2019 is a Continuation of U.S. application Ser. No. 14/926,674 filed Oct. 29, 2015, now Pat. No. 10,346,048, issued on Jul. 9, 2019, and the subject matter thereof is incorporated herein by reference thereto. U.S. patent application Ser. No. 14/926,674 filed Oct. 29, 2015, now Pat. No. 10,346,048, issued on Jul. 9, 2019, further claims the benefit of U.S. Provisional Patent Application Ser. No. 62/145,631 filed Apr. 10, 2015, and the subject matter thereof is incorporated herein by reference thereto.
- An embodiment of the present invention relates generally to an electronic system, and more particularly to a system for data storage functions for electronic systems.
- Modern electronic systems rely on rapid execution of programs and manipulation of data. The growing importance of “Big Data” databases is stressing the established data storage structures and processes. The term Big Data refers to huge volumes of data having a diverse and complex nature that is captured very quickly. Data is exploding at an astounding rate: up until 2003 approximately 5 exabytes of data was created, 90% of the data was structured data that was gathered through rational database management systems (RDBMS), and the rest was unstructured data. Currently, unstructured data is growing faster than structured data. Predictions are made that data will grow to 5 zettabytes by 2015. Unstructured data is growing faster than structured data. According to an International Data Corporation (IDC) study, 90% of all data created in the next decade will be unstructured data.
- The logistics of storing Big Data databases exposes weakness in the technology of the SSD. Storage cell wear issues can include a limited number of reads allowed between each write, a limited number of writes before the cell becomes inoperative, erasures count as a write, erase required prior to writing, erasure can only be performed in larger capacity than that of write, and Flash memory can include imperfect cells, which need to be managed as bad blocks. The storage cell wear issues require constant maintenance as well as over-provisioning, which reserves additional capacity of the storage cells that is not available in the rated capacity available to the user. In order to accommodate the requirement of the Big Data databases and their uses, large capacity storage devices must be used. The typical uses of the Big Data can very quickly cause an SSD to experience degraded performance due to the background maintenance required to protect the data. Multiple reads of the data can reduce the retention reliability and multiple writes can physically damage the storage cells of the SSD. Normal operations of Big Data databases under block-level storage architecture can cause the block-level SSD to spend more time maintaining the reliability of the data than being able to accept new read/write commands.
- Thus, a need still remains for electronic system with storage management mechanism to improve execution reliability and performance in clustered computing environments. In view of the ever-increasing commercial competitive pressures, along with growing consumer expectations and the diminishing opportunities for meaningful product differentiation in the marketplace, it is increasingly critical that answers be found to these problems. Additionally, the need to reduce costs, improve efficiencies and performance, and meet competitive pressures adds an even greater urgency to the critical necessity for finding answers to these problems.
- Solutions to these problems have been long sought but prior developments have not taught or suggested any solutions and, thus, solutions to these problems have long eluded those skilled in the art.
- An embodiment of the present invention provides an electronic system includes: a key value storage device, configured to transfer user data, the key value storage device including: a non-volatile memory array accessed by a key value address, an interface circuit, coupled to the non-volatile memory array, configured to receive a key value transfer command, a volatile memory, coupled to the interface circuit and the non-volatile memory array, configured to reduce a number of copies of the user data in the non-volatile memory array, and a device processor, coupled to the interface circuit, configured to manage the non-volatile memory array, the volatile memory, and the interface circuit by a key value index tree, including a key value translation block, to access the user data.
- An embodiment of the present invention provides a method of operation of an electronic system including: transferring user data by a key value storage device including: receiving a key value transfer command through an interface circuit includes connecting a device coupling structure to the interface circuit, selecting a non-volatile memory array, a volatile memory, or a combination thereof for reducing a number of copies of the user data in the non-volatile memory array accessed by a key value address, and managing a key value index tree, by a device processor, for transferring the user data between the interface circuit, the volatile memory, the non-volatile memory array, or a combination thereof includes accessing a key value translation block.
- Certain embodiments of the invention have other steps or elements in addition to or in place of those mentioned above. The steps or elements will become apparent to those skilled in the art from a reading of the following detailed description when taken with reference to the accompanying drawings.
-
FIG. 1 is an architectural block diagram of an electronic system with storage management mechanism in an embodiment. -
FIG. 2 is a hardware block diagram of a flash defined key value store device in an embodiment. -
FIG. 3 is a software functional block diagram of an electronic system with storage management mechanism in an embodiment. -
FIG. 4 is a functional block diagram of an operating software of a key value storage device in an embodiment. -
FIG. 5 is a flow chart of a MVCC operation in an embodiment. -
FIG. 6 is a flow chart of a method of operation of an electronic system in a further embodiment of the present invention. - Various embodiments provide a data management mechanism for Big Data databases that can simplify the distributed input/output (I/O) interface and maximize the execution efficiency of the electronic system by establishing a less burdensome I/O architecture. The data management mechanism can be configured to process the data used in a user program without utilizing the I/O storage stack structure provided in the operating system. The execution of the data management mechanism can be executed at a user program level to provide efficient access to remote storage, local storage, and memory and provide the maximum program execution efficiency.
- Various embodiments provide a data management mechanism for Big Data databases by omitting the customary I/O storage stacks required in operating system architectures. The data management mechanism also allows manipulation of the data based on a key value assigned to the data rather than a logical block address. The data management mechanism can operate in a host central processing unit (CPU) system memory, without making I/O calls through the operating system. The execution of the data management mechanism from the user defined space of the system memory can provide higher priority operation, while managing the operational characteristics of a key value storage device. This combination can increase the efficiency and operational performance of the data management mechanism.
- The following embodiments are described in sufficient detail to enable those skilled in the art to make and use the invention. It is to be understood that other embodiments would be evident based on the present disclosure, and that system, process, or mechanical changes may be made without departing from the scope of an embodiment of the present invention.
- In the following description, numerous specific details are given to provide a thorough understanding of the invention. However, it will be apparent that the invention may be practiced without these specific details. In order to avoid obscuring an embodiment of the present invention, some well-known circuits, system configurations, and process steps are not disclosed in detail.
- The drawings showing embodiments of the system are semi-diagrammatic, and not to scale and, particularly, some of the dimensions are for the clarity of presentation and are shown exaggerated in the drawing figures. Similarly, although the views in the drawings for ease of description generally show similar orientations, this depiction in the figures is arbitrary for the most part. Generally, the invention can be operated in any orientation. The embodiments have been numbered first embodiment, second embodiment, etc. as a matter of descriptive convenience and are not intended to have any other significance or provide limitations for an embodiment of the present invention.
- The term “module” referred to herein can include software, hardware, or a combination thereof in an embodiment of the present invention in accordance with the context in which the term is used. For example, the software can be machine code, firmware, embedded code, and application software. Also for example, the hardware can be circuitry, processor, computer, integrated circuit, integrated circuit cores, a pressure sensor, an inertial sensor, a microelectromechanical system (MEMS), passive devices, or a combination thereof. Further, if a module is written in the apparatus claims section below, the modules are deemed to include hardware circuitry for the purposes and the scope of apparatus claims.
- The term “unit” referred to herein is a circuit formed of hardware components or hardware state machines used for specific functions. The “unit” can be for timing critical functions and does not necessarily include software functions or support. The term “customary I/O storage stacks” referred to herein are memory segments within the host system memory that are used to hold address pointers and interrupt flags for input/output (I/O) operations performed by the operating system. The customary I/O storage stacks insert a latency to the I/O in order to load the memory with interrupt pointers to service the transfer of data to or from the I/O device based on servicing the interrupt associated with the transfer.
- Referring now to
FIG. 1 , therein is shown an architectural block diagram of anelectronic system 100 with a data management mechanism in an embodiment. Theelectronic system 100 includes a keyvalue storage device 102, examples of such devices can include solid state storage devices (SSSD) based on Flash memory, magnetic random access memory (MRAM), Resistive Random Access Memory (RRAM), and Phase Change Memory (PCM), as well as hybrid disk storage devices, and key value storage adapters, which are addressed by way of a key value for identifying file structures. The keyvalue storage device 102 can be a non-volatile data storage mechanism that stores and accesses user data by way of key value addressing. The keyvalue storage device 102 can be used for data processing in a computing environment for managing big data databases. - A
device coupling structure 104 can link the keyvalue storage device 102 to ahost processor 106. Thedevice coupling structure 104 can be an input/output interface structure connected between thehost processor 106 and the keyvalue storage device 102. Thedevice coupling structure 104 can include a peripheral computer interface express (PCI-e), fiber channel (FC), small computer system interface (SCSI), or serial attached SCSI (SAS), and the host's memory channel interface. Thedevice coupling structure 104 can be implemented as a memory bus for host internal application of theelectronic system 100. Thedevice coupling structure 104 can provide local or remote connectivity between thehost processor 106 and the keyvalue storage device 102. The connection of thedevice coupling structure 104 between the keyvalue storage device 102 and thehost processor 106 is performed in a manner that meets the specification of thedevice coupling structure 104. - The
device coupling structure 104 can provide direct coupling and communication to the keyvalue storage device 102 as well as standard storage devices (not shown). It is understood that the communication with standard storage devices requires the I/O calls provided by the operating system and may utilize a logical block address (LBA) architecture. Since the keyvalue storage device 102 does not use the LBA architecture, the communication from thehost processor 106 must be structured in a key value based linear storage architecture. - The
host processor 106 can be a general purpose computer, a computer network, a server, a storage processor, GPU, ASIC, FPGA, PLD, or the like. Thehost processor 106 can include asystem memory 108, ahardware interface controller 110, and a control processor (not shown). Thehost processor 106 can execute a key value device application program interface (API) 112 and a keyvalue device driver 114. It is understood that thesystem memory 108 can be outside the physical bounds of thehost processor 106. Thehardware interface controller 110 is a hardware adapter that controls the electrical and timing parameters associated with thedevice coupling structure 104. - The key
value device API 112 provides an application interface that allows user applications to access the keyvalue storage device 102 without having a description of the operational characteristics of the keyvalue storage device 102. The keyvalue device API 112 is a program that translates standard Application instructions into a form that is recognized by the keyvalue device driver 114. The keyvalue device API 112 passes the application requirements to the keyvalue device driver 114 for conversion to interface commands that are recognized by the keyvalue storage device 102. - The key
value device driver 114 can communicate with a physicalinterface conditioning block 116, such as Non-Volatile Memory Express (NVMe) or the Linux kernel virtual file system (sysfs). The physicalinterface conditioning block 116 can provide formatted commands to a universal Input/Output (UIO)kernel driver 118, such as an NVMe UIO kernel driver or a UIO framework with internal kernel functions provided by another block. The physicalinterface conditioning block 116 is a program that can add required information used by thehardware interface controller 110 to access thedevice coupling structure 104. TheUIO kernel driver 118 is a program that can manage and regulate the interrupt timing for thehardware interface controller 110. - The
UIO kernel driver 118 can communicate through thehardware interface controller 110 to thedevice coupling structure 104 and on to the keyvalue storage device 102. The physicalinterface conditioning block 116, theUIO kernel driver 118, thehardware interface controller 110, and thedevice coupling structure 104 provide a communication transfer mechanism between the keyvalue device driver 114 and the keyvalue storage device 102. - The key
value storage device 102 can include ahardware interface 120 for coupling to thedevice coupling structure 104. Thehardware interface 120 can be hardware structure that manages the communication and timing between the keyvalue storage device 102 and thedevice coupling structure 104. Thehardware interface 120 can be controlled by adevice processor 122, which can implement a key value flash translation layer (KVFTL), for managing the geometry of anon-volatile memory array 124, and support functions 126, which can include garbage collection, wear leveling, bad block management, and command execution. The keyvalue storage device 102 can provide the support functions 126 for expiry, multi-version concurrency control (MVCC), and available storage collection and management. The support functions 126 can include hardware and microcode for supporting the execution of the KVFTL. The hardware associated with the support functions 126 can include memory, register files, sequential state machines, multiplexers, error checking devices, and power monitoring devices. Thedevice processor 122 can include a microprocessor, an embedded processor, a sequential state machine, or a microcomputer. Thenon-volatile memory array 124 can be an array of non-volatile memory devices including flash memory, magnetoresistive random access memory, Ferroelectric random access memory, or phase-change random access memory. - The key
value storage device 102 can provide increased performance by not utilizing the standard block level architecture. The user applications that access the keyvalue storage device 102 are not required to maintain indexing structures in order to search for specific data. A key value that is unique to the data is stored by the user application to locate specific data structures and all indexing functions are performed by the keyvalue storage device 102. - The support functions 126, of the key
value storage device 102, can manage thenon-volatile memory array 124 as a linear addressable space with bad blocks masked out. Thedevice processor 122 can maintain a key value index list to facilitate the indexing of user data without support from thehost processor 106. - It has been discovered that the
electronic system 100 can enhance performance of the storage management by reducing the I/O overhead in thehost processor 106 by removing the concept of a logical block address in favor of a key value that can define the user data while leaving the identification of the location and extent of the user data to the keyvalue storage device 102. The device processor of the keyvalue storage device 102 can maintain a key value index list that can accommodate expiry, MVCC, and read/write access. - Referring now to
FIG. 2 , therein is shown a hardware block diagram of a flash defined key value store device in an embodiment. The hardware block diagram of the flash defined keyvalue store device 102 depicts theinterface circuit 120, such as a peripheral computer interface express (PCI-e), fiber channel (FC), small computer system interface (SCSI), serial attached SCSI (SAS), or the like, that can deliver the key value storage commands and data through thedevice coupling structure 104 ofFIG. 1 . Thedevice processor 122, such as a bit slice processor, a microprocessor, a hardware sequential state machine, an embedded microprocessor, an FPGA, an ASIC, a microcomputer, or the like, can be coupled to theinterface circuit 120 in order to receive key value commands, prepare theinterface circuit 120 for routing user data, manage stored data, or providing status to thehost processor 106 ofFIG. 1 . - The
device processor 122 can be coupled to adevice memory 202, which can include volatile memory, non-volatile memory, or a combination thereof. It is understood that theinterface circuit 120, thedevice processor 122, and thedevice memory 202 can be integrated into a single component, such as a system-on-a-chip (SOC), a gate array, a custom logic chip, a field programmable gate array, or an application specific integrated circuit (ASIC). Thedevice memory 202 can store the key value associated with the user data received through theinterface circuit 120. A keyvalue index list 204 can be stored in thedevice memory 202 and maintained by thedevice processor 122 for all of the support functions 126 ofFIG. 1 . The keyvalue index list 204 will link the key value with a physical location of the user data. - A
volatile memory 206 can be coupled to theinterface circuit 120. Thevolatile memory 206 can be used as a data cache and can be accessed by thedevice processor 122 through theinterface circuit 120. Anon-volatile memory array 124 can be coupled to thevolatile memory 206 and theinterface circuit 120 in order to provide reduced latency access to user data written to thevolatile memory 206, thenon-volatile memory array 124, or both. Theinterface circuit 120 can be configured to concurrently write both thevolatile memory 206 and thenon-volatile memory array 124 in order to support the MVCC function. Thedevice processor 122 can manipulate the user data in thevolatile memory 206 and thenon-volatile memory array 124 through theinterface circuit 120. Theinterface circuit 120 can actively transfer user data to or from one of thevolatile memory 206 and thenon-volatile memory array 124 while thedevice processor 122 can concurrently access the other of thevolatile memory 206 and thenon-volatile memory array 124. - The
device processor 122 can execute the support functions 126 ofFIG. 1 while theinterface circuit 120 is actively transferring user data to thevolatile memory 206 or thenon-volatile memory array 124 without adding latency to the transfer of the user data. Thenon-volatile memory array 124 can include a number of flash memory integrated circuits coupled in the array format on a key valuedevice circuit board 208. - It has been discovered that an embodiment of the key
value storage device 102 can reduce the write amplification factor as compared to a logical block address (LBA) device. The support functions 126 can produce fewer copies of the user data during garbage collection and wear leveling. The support functions 126 can take advantage of the device architecture by utilizing thevolatile memory 206 as a bridge device to hold temporary copies of the key value data blocks, rather than making multiple copies within thenon-volatile memory array 124. This technique can reduce the number of write operations applied to thenon-volatile memory array 124 and reduce the burden of the garbage collection because thevolatile memory 206 can be used without generating additional cell utilization within thenon-volatile memory array 124. - Referring now to
FIG. 3 , therein is shown an architectural block diagram of anelectronic system 301 with storage management mechanism in an embodiment. The architectural block diagram of theelectronic system 301 can depict auser space 302 and an operatingsystem kernel space 304 within thesystem memory 108 ofFIG. 1 . - The
user space 302 can be utilized by user applications for preparinguser data 306 for transfer to the keyvalue storage device 102. The keyvalue device API 112 can pass theuser data 306 to the keyvalue device driver 114 for configuration to an I/O compatible format. The keyvalue device driver 114 can access an I/O description 308 and an I/O configuration 310 in order to correctly format theuser data 306 for theUIO kernel driver 118 for direct transfer to the keyvalue storage device 102. - It has been discovered that the
electronic system 301 can reduce the latency of the transfer of theuser data 306 by bypassing the system calls 312 and the I/O storage stack 314 that are normally associated with I/O operations. The system calls 312 and the I/O storage stack 314 operate in the operatingsystem kernel space 304 of thesystem memory 108. The operatingsystem kernel space 304 is normally utilized during system interrupts, which can add operating system latency to the I/O operation. By executing theUIO kernel driver 118 directly from the keyvalue device driver 114, theuser data 306 can be transferred as a keyvalue transfer command 316 from theuser space 302 to the keyvalue storage device 102 without waiting for a system interrupt to occur. The keyvalue storage device 102 can monitor the indexing of theuser data 306 based on the unique key value associated with theuser data 306, which can reduce the amount of thesystem memory 108 and latency needed to save the information from the I/O storage stack 314. By reducing both the latency and the amount of thesystem memory 108 utilized by the I/O storage stack 314, theelectronic system 301 can enhance performance and reduce the power required for the transfer of theuser data 306. - The
UIO kernel driver 118 can convert the keyvalue transfer command 316 to akey value transfer 318. Thekey value transfer 318 can be a key value command for the keyvalue storage device 102 for a transfer of theuser data 306 to satisfy the keyvalue transfer command 316. Thedevice processor 122 ofFIG. 2 can interpret thekey value transfer 318 in order to manage the transaction of theuser data 306 with thevolatile memory 206 ofFIG. 2 , thenon-volatile memory array 124 ofFIG. 1 , or both concurrently. Once thedevice processor 122 can receive the keyvalue transfer command 316, theinterface circuit 120 can be configured to send or receive theuser data 306. Theuser data 306 can be transferred between theinterface circuit 120 and thevolatile memory 206, thenon-volatile memory array 124, or both concurrently, in the case of a multi-version concurrency control (MVCC) operation. - Referring now to
FIG. 4 , therein is shown an operational block diagram of a keyvalue storage device 401 in an embodiment. The operational block diagram of the keyvalue storage device 401 depicts the support functions 126 executed by thedevice processor 122 ofFIG. 2 including acommand processing queue 402, a keyvalue index tree 404, and a device real-time operating system (RTOS) 406. - The
command processing queue 402 can manage queued commands from the keyvalue device driver 114, while coordinating access of the keyvalue index tree 404 by thedevice processor 122 ofFIG. 1 . Thecommand processing queue 402 can configure thedevice coupling structure 104 for direct communication to the keyvalue device driver 114 by addressing the requests and responses to the address of the keyvalue device driver 114 inuser space 302 ofFIG. 2 . Thecommand processing queue 402 can set-up theinterface circuit 120 ofFIG. 2 to optimize the access for thedevice processor 122 without adding latency to the transfer of theuser data 306 ofFIG. 3 by implementing a direct communication to the keyvalue device driver 114 ofFIG. 3 and bypassing the system calls 312 ofFIG. 3 and thestorage stack 314 ofFIG. 3 theelectronic system 100 ofFIG. 1 . Thus thecommand processing queue 402 can significantly reduces the latency between thekey value transfers 318 because the transfers occur as operations inuser space 302 that do not wait for system interrupt processing to occur. - The key
value index tree 404 can include agarbage collection block 408, awear leveling block 410, awrite operation monitor 412, a read operation monitor 414, adevice monitor 416, a keyvalue translation block 418, and a badblock management module 420 for supporting thenon-volatile memory array 124 ofFIG. 1 . The keyvalue index tree 404 can maintain the operational readiness of the keyvalue storage device 102 ofFIG. 1 without adding latency to the transfer of theuser data 306. The keyvalue index tree 404 can be managed by thedevice processor 122 in order to maintain the current location of theuser data 306 contained in thevolatile memory 206 ofFIG. 2 and thenon-volatile memory array 124. Thedevice processor 122 can establish rules governing the execution of thegarbage collection block 408 and thewear leveling block 410 in order to preserve multiple versions of theuser data 306 associated with the same key. - The
garbage collection block 408 can reduce the write amplification factor by utilizing a portion of thevolatile memory 206 ofFIG. 2 to store still valid data while performing an erase of expired or invalid blocks of theuser data 306. The reduction in the write amplification factor can extend the operable life of the keyvalue storage device 102, well beyond that of a solid state drive (SSD), while maintaining the high performance and reduced latency that makes the keyvalue storage device 102 desirable. Any of the valid data displaced due to thegarbage collection block 408 is readily available for transfer to theinterface circuit 120 if required. This process is necessary because theuser data 306 can be written in 4KB to 8KB pages but an erase can impact (for example) 64 such pages at a time. - The
wear leveling block 410 can manage the distribution of the writes of theuser data 306, to keep the charge distribution in thenon-volatile memory array 124 in balance, and prevent exceeding the lifetime write limit of cells in any area of thenon-volatile memory array 124. Thewear leveling block 410 can configure over-provisioned portions of thenon-volatile memory array 124 for use when replacing worn or faulty blocks, as part of the wear leveling process. Since the write amplification factor is reduced for the keyvalue storage device 102, thewear leveling block 410 can be invoked less frequently than in products based on logical block address (LBA) access. - The write operation monitor 412 can configure the
interface circuit 120 to communicate directly to the keyvalue device driver 114 to provide a low latency path to thevolatile memory 206, thenon-volatile memory array 124, or a combination thereof. The write operation monitor 412 can receive a key value pair for theuser data 306 associated with the write command. The key value pair can be stored in the keyvalue index list 204 ofFIG. 2 in order to access theuser data 306 at a later time. Upon completion of the write operation, the write operation monitor 412 can configure the status to be returned by theinterface circuit 120. - The read operation monitor 412 can configure the
interface circuit 120 after verifying the key value, received for the read command, is valid. The read operation monitor can configure theinterface circuit 120 to communicate directly to the keyvalue device driver 114 to provide a low latency path for theuser data 306 read from thevolatile memory 206, thenon-volatile memory array 124, or a combination thereof. At the completion of the read operation, the read operation monitor 414 can configure the status to be returned by theinterface circuit 120. - The device monitor 416 can detect error or warning conditions associated with the general operation of the key
value storage device 102, such as data errors, timeout of operations, hardware warnings, or command exceptions. The device monitor 416 can provide recovery strategies for recoverable errors or warnings, provide status to theinterface circuit 120 for unrecoverable errors, and can transfer the contents of thevolatile memory 206 to a reserved portion of thenon-volatile memory array 124 upon notification of power loss or shut down. The recovery strategies can include retries, data error correction, retransmission, or emergency back-up to the reserved portion of thenon-volatile memory array 124. The keyvalue storage device 102 can include a charge storage device or small battery in order to maintain control during the power-down processing. - The key
value translation block 418 can receive key value pairs from the write operation monitor 412 and add the key value to the keyvalue index list 204 ofFIG. 2 with the physical location in thevolatile memory 206 or thenon-volatile memory array 124. The “key” portion of the key value pair is used to index the physical location of the “value” portion, which can be theuser data 306 or a pointer to the physical location of theuser data 306. For read operations the keyvalue translation block 418 can verify the key value received with the read command is located in the keyvalue index list 204 and provide the physical address of theuser data 306 to the read operation monitor 414 for configuration of theinterface circuit 120 and transfer theuser data 306. - The bad
block management module 420 can tag all blocks within thenon-volatile memory array 124 that contains a page with an unrecoverable error. The badblock management module 420 can receive error indications from thegarbage collection block 408, thewear leveling block 410, thewrite operation monitor 412, the read operation monitor 414, thedevice monitor 416, or a combination thereof. - The
device RTOS 406 can enable thedevice processor 122 to provide system level supervision and timing prompts for the support functions 126. Thedevice RTOS 406 can provide system level timing functions, diagnostic capabilities, system recovery processes and reset management. - It is understood that other partitions of the support functions are possible without changing the operation of the key
value storage device 102. By way of an example, thewrite operation monitor 412, the read operation monitor 414, thedevice monitor 416, and the keyvalue translation module 418 could be combined in a single module or further partitioned to provide other more detailed module functions. - It has been discovered that the support functions 126 of the key
value storage device 102 can provide low latency access to theuser data 306 stored in thevolatile memory 206 or thenon-volatile memory array 124 with a simplified command interface structure without using logical block address architecture. This can reduce the use of thesystem memory 108 ofFIG. 1 while increasing performance due to bypassing the system calls 312 ofFIG. 3 and thestorage stack 314 ofFIG. 3 , which are used to support the LBA architecture. - Referring now to
FIG. 5 , therein is shown a flow chart of aMVCC operation 501 in an embodiment. The flow chart of theMVCC operation 501 depicts receiving an update command atstep 502, such as a write command, updating an existing file, received through theinterface circuit 120 ofFIG. 2 . Thedevice processor 122 ofFIG. 2 can identify the key value from the keyvalue transfer command 316 ofFIG. 3 . - The
device processor 122 then proceeds to identify the key atstep 504 in the keyvalue index list 204 ofFIG. 2 to determine whether theuser data 306 ofFIG. 3 already exists in theelectronic system 100 and to identify that the update command is valid. Upon identifying that theuser data 306 does exist, thedevice processor 122 can allocate a location for the new key value pair atstep 506, which can be added to the keyvalue index list 204. - The
device processor 122 can update metadata in the new key value pair location atstep 508. The keyvalue index list 204 can include, for each key value entry, a link pointer to identify previous versions of the key value in order to provide a history of the value associated with the key. Since thenon-volatile memory array 124 ofFIG. 2 cannot perform an overwrite of the existing location, the older version ofuser data 306 associated with the key remains available. 1 during background support operations. In multi-host environments, this feature can allow access to multiple versions of the original data set and allow each host to perform updates. - By way of an example, the key can identify a data set and the value can indicate a version number of the data set. The metadata can include the version number and pointers to the physical location of the indicated version. In an embodiment, the device processor can assign limits to the number of different values associated with a single key. It has been discovered that the
electronic system 100 can provide a back-up process to restore a data set to a previous version with a single key value read command. - The
device processor 122 can update the keyvalue index list 204 to include the new key value pair atstep 510. The keyvalue index list 204 can include the link pointer to the previous key value pair location, the key provided from the command, and metadata, which includes the value associated with the key and physical pointers to the location of theuser data 306 associated with the data set. - It has been discovered that the
electronic system 100 can support multiple versions of the same data set concurrently. This feature is further supported by the use of thevolatile memory 206 ofFIG. 2 with thegarbage collection block 408 and thewear leveling block 410, which can substantially reduce the number of times the same data set is replicated. By transferring any active data in an erase block to thevolatile memory 206, it is not necessary to consume more of thenon-volatile memory array 124 for the support functions 126. - Referring now to
FIG. 6 , therein is shown a flow chart of amethod 600 of operation of anelectronic system 100 in a further embodiment of the present invention. Themethod 600 includes: transferring user data between the keyvalue device driver 114 ofFIG. 1 and theinterface circuit 120 ofFIG. 1 , without experiencing the normal delay caused by the I/O storage stack 314 ofFIG. 3 , using a keyvalue storage device 102 ofFIG. 1 . Themethod 600 includes: receiving a keyvalue transfer command 316 ofFIG. 3 through theinterface circuit 120 includes connecting adevice coupling structure 104 ofFIG. 1 to theinterface circuit 120 in ablock 602, selecting anon-volatile memory array 124 ofFIG. 1 , avolatile memory 206 ofFIG. 2 , or a combination thereof for reducing a number of copies of theuser data 306 ofFIG. 3 in thenon-volatile memory array 124 ofFIG. 1 accessed by a key value address in ablock 604, and managing a keyvalue index tree 404 ofFIG. 4 , by thedevice controller 122 ofFIG. 1 , for transferring theuser data 306 between theinterface circuit 120, thevolatile memory 206, thenon-volatile memory array 124, or a combination thereof includes accessing a key value translation block in a block 606. - The resulting method, process, apparatus, device, product, and/or system is straightforward, cost-effective, uncomplicated, highly versatile, accurate, sensitive, and effective, and can be implemented by adapting known components for ready, efficient, and economical manufacturing, application, and utilization. Another important aspect of an embodiment of the present invention is that it valuably supports and services the historical trend of reducing costs, simplifying systems, and increasing performance.
- These and other valuable aspects of an embodiment of the present invention consequently further the state of the technology to at least the next level.
- While the invention has been described in conjunction with a specific best mode, it is to be understood that many alternatives, modifications, and variations will be apparent to those skilled in the art in light of the aforegoing description. Accordingly, it is intended to embrace all such alternatives, modifications, and variations that fall within the scope of the included claims. All matters set forth herein or shown in the accompanying drawings are to be interpreted in an illustrative and non-limiting sense.
Claims (20)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US16/887,341 US11630578B2 (en) | 2015-04-10 | 2020-05-29 | Electronic system with storage management mechanism and method of operation thereof |
Applications Claiming Priority (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US201562145631P | 2015-04-10 | 2015-04-10 | |
| US14/926,674 US10346048B2 (en) | 2015-04-10 | 2015-10-29 | Electronic system with storage management mechanism and method of operation thereof |
| US16/397,543 US10747443B2 (en) | 2015-04-10 | 2019-04-29 | Electronic system with storage management mechanism and method of operation thereof |
| US16/887,341 US11630578B2 (en) | 2015-04-10 | 2020-05-29 | Electronic system with storage management mechanism and method of operation thereof |
Related Parent Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/397,543 Continuation US10747443B2 (en) | 2015-04-10 | 2019-04-29 | Electronic system with storage management mechanism and method of operation thereof |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20200293191A1 true US20200293191A1 (en) | 2020-09-17 |
| US11630578B2 US11630578B2 (en) | 2023-04-18 |
Family
ID=57112637
Family Applications (3)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/926,674 Active 2036-04-14 US10346048B2 (en) | 2015-04-10 | 2015-10-29 | Electronic system with storage management mechanism and method of operation thereof |
| US16/397,543 Active US10747443B2 (en) | 2015-04-10 | 2019-04-29 | Electronic system with storage management mechanism and method of operation thereof |
| US16/887,341 Active US11630578B2 (en) | 2015-04-10 | 2020-05-29 | Electronic system with storage management mechanism and method of operation thereof |
Family Applications Before (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/926,674 Active 2036-04-14 US10346048B2 (en) | 2015-04-10 | 2015-10-29 | Electronic system with storage management mechanism and method of operation thereof |
| US16/397,543 Active US10747443B2 (en) | 2015-04-10 | 2019-04-29 | Electronic system with storage management mechanism and method of operation thereof |
Country Status (1)
| Country | Link |
|---|---|
| US (3) | US10346048B2 (en) |
Families Citing this family (12)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10346048B2 (en) * | 2015-04-10 | 2019-07-09 | Samsung Electronics Co., Ltd. | Electronic system with storage management mechanism and method of operation thereof |
| EP3279810B1 (en) * | 2016-06-14 | 2019-09-04 | Huawei Technologies Co., Ltd. | Data access method and related apparatus and system |
| US10318185B2 (en) * | 2016-07-01 | 2019-06-11 | Intel Corporation | Method and apparatus to provide both storage mode and memory mode access to non-volatile memory within a solid state drive |
| US10649969B2 (en) * | 2017-06-13 | 2020-05-12 | Western Digital Technologies, Inc. | Memory efficient persistent key-value store for non-volatile memories |
| US10949341B2 (en) * | 2018-08-27 | 2021-03-16 | Samsung Electronics Co., Ltd. | Implementing snapshot and other functionality in KVSSD through garbage collection and FTL |
| KR20210092361A (en) | 2020-01-15 | 2021-07-26 | 삼성전자주식회사 | Storage device and operation method thereof |
| EP3851950B1 (en) | 2020-01-15 | 2024-09-04 | Samsung Electronics Co., Ltd. | Storage device and operation method thereof |
| CN112579003B (en) * | 2020-12-15 | 2022-06-14 | 浙江大华技术股份有限公司 | Key value pair adjustment method, key value pair adjustment device, storage medium and electronic device |
| US11625189B2 (en) * | 2021-06-28 | 2023-04-11 | Western Digital Technologies, Inc. | Systems and methods for fragmentation management in host buffers |
| US11853605B2 (en) * | 2021-09-17 | 2023-12-26 | Micron Technology, Inc. | Database persistence |
| JP2024000843A (en) * | 2022-06-21 | 2024-01-09 | キオクシア株式会社 | memory system |
| CN115756312A (en) * | 2022-11-02 | 2023-03-07 | 平头哥(上海)半导体技术有限公司 | Data access system, data access method, and storage medium |
Citations (22)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20010052093A1 (en) * | 2000-05-02 | 2001-12-13 | Japan Aviation Electronics Industry Limited | Memory testing method and memory testing apparatus |
| US7149832B2 (en) * | 2004-11-10 | 2006-12-12 | Microsoft Corporation | System and method for interrupt handling |
| US20080065670A1 (en) * | 2001-06-08 | 2008-03-13 | Sap Ag. | Cache-conscious concurrency control scheme for database systems |
| US20080215849A1 (en) * | 2007-02-27 | 2008-09-04 | Thomas Scott | Hash table operations with improved cache utilization |
| US7756817B2 (en) * | 2007-07-05 | 2010-07-13 | Yahoo! Inc. | System and method for enabling parallel access to serially compressed files |
| US20110010347A1 (en) * | 2007-01-26 | 2011-01-13 | Hicamp Systems, Inc. | Iterator register for structured memory |
| US20120197900A1 (en) * | 2010-12-13 | 2012-08-02 | Unisys Corporation | Systems and methods for search time tree indexes |
| US8248861B2 (en) * | 2007-01-22 | 2012-08-21 | Round Rock Research, Llc | Memory system and method having volatile and non-volatile memory devices at same hierarchical level |
| US8375012B1 (en) * | 2011-08-10 | 2013-02-12 | Hewlett-Packard Development Company, L.P. | Computer indexes with multiple representations |
| US20130060981A1 (en) * | 2011-09-06 | 2013-03-07 | Western Digital Technologies, Inc. | Systems and methods for an enhanced controller architecture in data storage systems |
| US20130275656A1 (en) * | 2012-04-17 | 2013-10-17 | Fusion-Io, Inc. | Apparatus, system, and method for key-value pool identifier encoding |
| US20140059271A1 (en) * | 2012-08-27 | 2014-02-27 | Apple Inc. | Fast execution of flush commands using adaptive compaction ratio |
| US20140310574A1 (en) * | 2012-12-28 | 2014-10-16 | Super Talent Technology, Corp. | Green eMMC Device (GeD) Controller with DRAM Data Persistence, Data-Type Splitting, Meta-Page Grouping, and Diversion of Temp Files for Enhanced Flash Endurance |
| US20150142733A1 (en) * | 2013-11-18 | 2015-05-21 | Ori Software Development Ltd. | System and method for efficient management of big data in a database using streaming tables |
| US9201825B1 (en) * | 2011-11-02 | 2015-12-01 | Marvell International Ltd. | Data storage methods and apparatus |
| US20150370720A1 (en) * | 2014-06-18 | 2015-12-24 | Netspeed Systems | Using cuckoo movement for improved cache coherency |
| US20160099810A1 (en) * | 2014-10-03 | 2016-04-07 | Seagate Technology Llc | Key-Value Data Storage Device with Hybrid Architecture |
| US9455914B2 (en) * | 2013-09-10 | 2016-09-27 | Robin Systems, Inc. | File-system requests supported in user space for enhanced efficiency |
| US20170103031A1 (en) * | 2015-10-08 | 2017-04-13 | Samsung Electronics Co., Ltd. | Method for application-aware interrupts management |
| US9648102B1 (en) * | 2012-12-27 | 2017-05-09 | Iii Holdings 2, Llc | Memcached server functionality in a cluster of data processing nodes |
| US20170277451A1 (en) * | 2016-03-22 | 2017-09-28 | Kabushiki Kaisha Toshiba | Method to limit impact of partial media failure of disk drive and detect/report the loss of data for objects due to partial failure of media |
| US10346048B2 (en) * | 2015-04-10 | 2019-07-09 | Samsung Electronics Co., Ltd. | Electronic system with storage management mechanism and method of operation thereof |
Family Cites Families (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6591338B1 (en) | 1999-05-03 | 2003-07-08 | 3Ware, Inc. | Methods and systems for mirrored disk arrays |
| JP2006235960A (en) | 2005-02-24 | 2006-09-07 | Fujitsu Ltd | Garbage collection acceleration method |
| US8683456B2 (en) | 2009-07-13 | 2014-03-25 | Apple Inc. | Test partitioning for a non-volatile memory |
| JPWO2012020544A1 (en) | 2010-08-11 | 2013-10-28 | 日本電気株式会社 | Data processing system, data processing method, and program |
| US8694754B2 (en) | 2011-09-09 | 2014-04-08 | Ocz Technology Group, Inc. | Non-volatile memory-based mass storage devices and methods for writing data thereto |
| US8782344B2 (en) | 2012-01-12 | 2014-07-15 | Fusion-Io, Inc. | Systems and methods for managing cache admission |
| CN103902632B (en) | 2012-12-31 | 2018-01-02 | 华为技术有限公司 | The method, apparatus and electronic equipment of file system are built in key assignments storage system |
| US9323798B2 (en) | 2013-01-14 | 2016-04-26 | International Business Machines Corporation | Storing a key value to a deleted row based on key range density |
| US9329991B2 (en) | 2013-01-22 | 2016-05-03 | Seagate Technology Llc | Translation layer partitioned between host and controller |
-
2015
- 2015-10-29 US US14/926,674 patent/US10346048B2/en active Active
-
2019
- 2019-04-29 US US16/397,543 patent/US10747443B2/en active Active
-
2020
- 2020-05-29 US US16/887,341 patent/US11630578B2/en active Active
Patent Citations (23)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20010052093A1 (en) * | 2000-05-02 | 2001-12-13 | Japan Aviation Electronics Industry Limited | Memory testing method and memory testing apparatus |
| US20080065670A1 (en) * | 2001-06-08 | 2008-03-13 | Sap Ag. | Cache-conscious concurrency control scheme for database systems |
| US7149832B2 (en) * | 2004-11-10 | 2006-12-12 | Microsoft Corporation | System and method for interrupt handling |
| US8248861B2 (en) * | 2007-01-22 | 2012-08-21 | Round Rock Research, Llc | Memory system and method having volatile and non-volatile memory devices at same hierarchical level |
| US20110010347A1 (en) * | 2007-01-26 | 2011-01-13 | Hicamp Systems, Inc. | Iterator register for structured memory |
| US20080215849A1 (en) * | 2007-02-27 | 2008-09-04 | Thomas Scott | Hash table operations with improved cache utilization |
| US7756817B2 (en) * | 2007-07-05 | 2010-07-13 | Yahoo! Inc. | System and method for enabling parallel access to serially compressed files |
| US20120197900A1 (en) * | 2010-12-13 | 2012-08-02 | Unisys Corporation | Systems and methods for search time tree indexes |
| US8375012B1 (en) * | 2011-08-10 | 2013-02-12 | Hewlett-Packard Development Company, L.P. | Computer indexes with multiple representations |
| US20130060981A1 (en) * | 2011-09-06 | 2013-03-07 | Western Digital Technologies, Inc. | Systems and methods for an enhanced controller architecture in data storage systems |
| US9201825B1 (en) * | 2011-11-02 | 2015-12-01 | Marvell International Ltd. | Data storage methods and apparatus |
| US20130275656A1 (en) * | 2012-04-17 | 2013-10-17 | Fusion-Io, Inc. | Apparatus, system, and method for key-value pool identifier encoding |
| US20140059271A1 (en) * | 2012-08-27 | 2014-02-27 | Apple Inc. | Fast execution of flush commands using adaptive compaction ratio |
| US9648102B1 (en) * | 2012-12-27 | 2017-05-09 | Iii Holdings 2, Llc | Memcached server functionality in a cluster of data processing nodes |
| US20140310574A1 (en) * | 2012-12-28 | 2014-10-16 | Super Talent Technology, Corp. | Green eMMC Device (GeD) Controller with DRAM Data Persistence, Data-Type Splitting, Meta-Page Grouping, and Diversion of Temp Files for Enhanced Flash Endurance |
| US9455914B2 (en) * | 2013-09-10 | 2016-09-27 | Robin Systems, Inc. | File-system requests supported in user space for enhanced efficiency |
| US20150142733A1 (en) * | 2013-11-18 | 2015-05-21 | Ori Software Development Ltd. | System and method for efficient management of big data in a database using streaming tables |
| US20150370720A1 (en) * | 2014-06-18 | 2015-12-24 | Netspeed Systems | Using cuckoo movement for improved cache coherency |
| US20160099810A1 (en) * | 2014-10-03 | 2016-04-07 | Seagate Technology Llc | Key-Value Data Storage Device with Hybrid Architecture |
| US10346048B2 (en) * | 2015-04-10 | 2019-07-09 | Samsung Electronics Co., Ltd. | Electronic system with storage management mechanism and method of operation thereof |
| US20170103031A1 (en) * | 2015-10-08 | 2017-04-13 | Samsung Electronics Co., Ltd. | Method for application-aware interrupts management |
| US9965412B2 (en) * | 2015-10-08 | 2018-05-08 | Samsung Electronics Co., Ltd. | Method for application-aware interrupts management |
| US20170277451A1 (en) * | 2016-03-22 | 2017-09-28 | Kabushiki Kaisha Toshiba | Method to limit impact of partial media failure of disk drive and detect/report the loss of data for objects due to partial failure of media |
Also Published As
| Publication number | Publication date |
|---|---|
| US10747443B2 (en) | 2020-08-18 |
| US11630578B2 (en) | 2023-04-18 |
| US10346048B2 (en) | 2019-07-09 |
| US20160299688A1 (en) | 2016-10-13 |
| US20190250821A1 (en) | 2019-08-15 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11630578B2 (en) | Electronic system with storage management mechanism and method of operation thereof | |
| CN115495389B (en) | Memory controller, calculation memory device, and operation method of calculation memory device | |
| US10157142B2 (en) | Offload data transfer engine for a block data transfer interface | |
| US9772802B2 (en) | Solid-state device management | |
| US9927984B2 (en) | Electronic system with interface control mechanism and method of operation thereof | |
| KR102329762B1 (en) | Electronic system with memory data protection mechanism and method of operation thereof | |
| US20150347325A1 (en) | Obtaining diagnostic information through host interfaces | |
| US20060085626A1 (en) | Updating system configuration information | |
| US20130111103A1 (en) | High-speed synchronous writes to persistent storage | |
| CN110806837A (en) | Data processing system and method of operation thereof | |
| KR20190066466A (en) | Storage method and apparatus for reducing write latency | |
| US12536083B2 (en) | Out-of-band management method for storage apparatus, baseboard management controller and storage apparatus | |
| CN113722131A (en) | Method and system for facilitating fast crash recovery in a storage device | |
| CN111435334B (en) | Apparatus and method for checking valid data in memory system | |
| US20230376238A1 (en) | Computing system for managing distributed storage devices, and method of operating the same | |
| US12340120B2 (en) | Storage controller, computational storage device, and operational method of computational storage device | |
| EP4148572A1 (en) | Computational storage device and storage system including the computational storage device | |
| US20190042355A1 (en) | Raid write request handling without prior storage to journaling drive | |
| US11789654B2 (en) | Data storage device and method for file-based interrupt coalescing | |
| CN118140205A (en) | Storage system and method for early command cancellation | |
| US12204795B2 (en) | Error detection and correction in a controller | |
| EP2924576A1 (en) | Storage control apparatus, control program, and control method | |
| WO2018140036A1 (en) | Read operation redirect | |
| KR101887663B1 (en) | Method and apparatus for processing data using dummy page | |
| KR20250138622A (en) | Data storage device and method for multi-tenant dynamic bandwidth reallocation |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| CC | Certificate of correction |