[go: up one dir, main page]

US20250095683A1 - Information processing device, magnetic recording/reproducing device, and magnetic recording/reproducing system - Google Patents

Information processing device, magnetic recording/reproducing device, and magnetic recording/reproducing system Download PDF

Info

Publication number
US20250095683A1
US20250095683A1 US18/740,842 US202418740842A US2025095683A1 US 20250095683 A1 US20250095683 A1 US 20250095683A1 US 202418740842 A US202418740842 A US 202418740842A US 2025095683 A1 US2025095683 A1 US 2025095683A1
Authority
US
United States
Prior art keywords
input data
feature value
processor
neural network
attention layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US18/740,842
Inventor
Susumu Takeda
Kenichiro Yamada
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Toshiba Corp
Original Assignee
Toshiba Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Toshiba Corp filed Critical Toshiba Corp
Assigned to KABUSHIKI KAISHA TOSHIBA reassignment KABUSHIKI KAISHA TOSHIBA ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: TAKEDA, SUSUMU, YAMADA, KENICHIRO
Publication of US20250095683A1 publication Critical patent/US20250095683A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • G11B20/10009Improvement or modification of read or write signals
    • G11B20/10018Improvement or modification of read or write signals analog processing for digital recording or reproduction
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • G11B20/10009Improvement or modification of read or write signals
    • G11B20/10481Improvement or modification of read or write signals optimisation methods
    • G11B20/10518Improvement or modification of read or write signals optimisation methods using neural networks

Definitions

  • Embodiments described herein relate generally to an information processing device, a magnetic recording/reproducing device, and a magnetic recording/reproducing system.
  • information recorded on a magnetic recording medium is reproduced by an information processing device.
  • magnetic recording density can be improved.
  • FIG. 1 is a schematic diagram illustrating an information processing device and an information recording/reproducing device according to a first embodiment
  • FIG. 2 is a schematic diagram illustrating a part of the information processing device according to the first embodiment
  • FIG. 3 is a schematic diagram illustrating a part of the information processing device according to the first embodiment.
  • FIG. 4 is a schematic diagram illustrating an information processing device and an information recording/reproducing device according to a second embodiment.
  • an information processing device includes an acquisitor and a processor.
  • the acquisitor is configured to acquire a reproduced signal based on a signal obtained from a magnetic recording medium.
  • the reproduced signal includes first input data, second input data, and third input data.
  • the processor is configured to perform a first process and a second process.
  • the processor is configured to derive a first derived feature value from a first inner product of a first feature value based on the first input data and a second feature value based on the second input data.
  • the processor is configured to derive a second derived feature value from a second inner product of a third feature value based on the third input data and the first derived feature value.
  • FIG. 1 is a schematic diagram illustrating an information processing device and an information recording/reproducing device according to a first embodiment.
  • a magnetic recording/reproducing device 210 includes an information processing device 70 , a magnetic recording medium 80 , and a reproducing portion 80 R.
  • the magnetic recording medium 80 and the reproducing portion 80 R may be included in a recording section 80 D, and the recording section 80 D may be included in, for example, a magnetic disk (HDD: Hard Disk Drive).
  • the magnetic recording/reproducing device 210 may further include, for example, an SSD (Solid State Drive).
  • the reproducing portion 80 R is configured to reproduce a signal 80 S corresponding to the information recorded on the magnetic recording medium 80 .
  • the reproducing portion 80 R includes, for example, a magnetic head 80 H.
  • the signal 80 S corresponding to the reproduced information is output from the reproducing portion 80 R.
  • the signal 80 S is an electrical signal.
  • Information may be recorded on the magnetic recording medium 80 by the magnetic head 80 H.
  • the information processing device 70 includes an acquisitor 70 i and a processor 70 p.
  • the acquisitor 70 i is configured to acquire a reproduced signal RS based on the signal 80 S obtained from the magnetic recording medium 80 .
  • the reproduced signal RS may be based on a signal obtained by amplifying the signal 80 S.
  • the reproduced signal RS may be a digital signal obtained by AD (Analog to Digital) conversion of the signal 80 S.
  • the acquisitor 70 i may be, for example, an interface.
  • the reproduced signal RS acquired by the acquisitor 70 i is processed by the processor 70 p.
  • the processor 70 p includes, for example, an input portion 70 pi and an output portion 70 po.
  • the reproduced signal RS is supplied to the input portion 70 pi.
  • the processing result in the processor 70 p is output from the output portion 70 po.
  • the reproduced signal RS includes first input data 11 , second input data 12 , and third input data 13 .
  • the processor 70 p is configured to perform a first process PR 1 and a second process PR 2 .
  • the first input data 11 includes data regarding a decoding target bit.
  • the second input data 12 and the third input data 13 include data regarding other bits.
  • the processor 70 p derives a first derived feature value FD 1 from a first inner product 31 of a first feature value F 11 based on the first input data 11 and a second feature value F 12 based on the second input data 12 .
  • the processor 70 p derives a second derived feature value FD 2 from a second inner product 32 of a third feature value F 13 based on the third input data 13 and the first derived feature value FD 1 .
  • the reproduced signal RS corresponding to information recorded on the magnetic recording medium 80 can be decoded with high accuracy.
  • the processor 70 p is configured to further perform a third process PR 3 .
  • the processor 70 p outputs a bit decoding result BD 1 regarding the decoding target bits included in the first input data 11 based on at least a part of the second derived feature value FD 2 .
  • the bit decoding result BD 1 being obtained is output from the output portion 70 po.
  • bit decoding result BD 1 corresponds to information related to 0/1 determination.
  • the information related to the 0/1 determination may be 0 or 1, for example.
  • the bit decoding result BD 1 may be information that can be used for 0/1 determination.
  • the information that can be used for the 0/1 determination may include, for example, the probability of being 0 or 1.
  • the information that can be used for the 0/1 determination may be, for example, the likelihood regarding 0 or 1.
  • the processor 70 p is configured to perform neural network processing (NN processing). As shown in FIG. 1 , the processor 70 p may include an attention layer 71 and a classification layer 72 .
  • the attention layer 71 is configured to perform the first process PR 1 and the second process PR 2 .
  • the classification layer 72 is configured to perform the third process PR 3 .
  • the classification layer 72 includes a fully connected neural network (FC-NN).
  • the neural network may be a component that includes parameters whose values can be adjusted (e.g., learned) based on information such as data.
  • the neural network may include a perceptron.
  • the reproduced signal RS obtained from the magnetic recording medium 80 is processed by attention processing including two inner product calculations.
  • attention processing including two inner product calculations.
  • feature value extraction of the waveform data of the reproduced signal RS is performed (first process PR 1 and second process PR 2 ).
  • the classification processing is performed using a fully connected NN.
  • the classification process corresponds to, for example, a process of converting information extracted by feature value extraction into information related to 0/1 determination.
  • highly accurate decoding can be performed.
  • a decoding at high speed and with high accuracy can be performed.
  • An information processing device that can improve processing accuracy can be provided.
  • the embodiment by calculation based on the inner product, it is possible to appropriately extract the feature value that takes into account the relationship of input data.
  • Parallel processing is easy in the processing based on inner products.
  • high-accuracy decoding can be performed at high speed due to high speed through parallel calculation and appropriate feature values.
  • an RNN Recurrent Neural Network
  • LSTM Long Short Term Memory
  • the first input data 11 , the second input data 12 , the third input data 13 , etc. correspond to information recorded on the magnetic recording medium 80 .
  • the first input data 11 includes data regarding a decoding target bit.
  • the second input data 12 includes data regarding a first other bit.
  • the third input data 13 includes data regarding a second other bit.
  • the decoding target bit is between the first other bit and the second other bit.
  • the decoding target bit may be next to the first other bit.
  • the decoding target bit may be next to the second other bit.
  • Another bit may be provided between the decoding target bit and the first other bit.
  • Another bit may be provided between the decoding target bit and the second other bit.
  • the first other bit may include one or more bits.
  • the second other bit may include one or more bits.
  • the first other bit may be provided between the decoding target bit and the second other bit.
  • At least one of the first input data 11 , the second input data 12 , or the third input data 13 may be determined by, for example, a method of the attention processing.
  • a method of the attention processing For example, self-attention may be applied, as described below.
  • the first feature value F 11 includes “query” regarding the first input data 11 .
  • the second feature value F 12 includes a “key” regarding the second input data 12 .
  • the attention layer 71 may include a first neural network 11 N and a second neural network 12 N.
  • the attention layer 71 is configured to derive the “query” (first feature value F 11 ) by processing the first input data 11 with the first neural network 11 N.
  • the attention layer 71 is configured to derive the “key” (second feature value F 12 ) with processing the second input data 12 by the second neural network 12 N.
  • the first derived feature value FD 1 includes, for example, a first weight indicating the relationship between the first input data 11 and the second input data 12 .
  • the third feature value F 13 includes “value” regarding the third input data 13 .
  • the attention layer 71 further includes a third neural network 13 N.
  • the attention layer 71 is configured to derive the “value” (third feature value F 13 ) by processing the third input data 13 with the third neural network 13 N.
  • the attention layer 71 may further include a fourth neural network 14 N.
  • the attention layer 71 is configured to derive the second derived feature value FD 2 by processing the second inner product 32 with the fourth neural network 14 N.
  • interference between bits is taken into consideration, for example, in HDD signal reproduction.
  • the attention processing according to the embodiment it is possible to extract feature values by considering the relationship of input data.
  • attention processing for HDD signal reproduced signal reproduction can be performed in consideration of interference (relationship) between bits.
  • the reproduced signal RS obtained from the magnetic recording medium 80 is processed by attention processing including two inner product calculations. Thereby, highly accurate decoding can be performed. An information processing device that can improve processing accuracy can be provided.
  • the number of attention layers 71 may be one, or two or more.
  • the processor 70 p may be configured to perform embedding processing 71 e.
  • the processor 70 p performs embedding processing 71 e of the reproduced signal RS (first input data 11 , second input data 12 , and third input data 13 ).
  • the embedding process may include, for example, embedding location information.
  • the embedding process may include, for example, a class token used when classifying input data to perform attention processing.
  • FIG. 2 is a schematic diagram illustrating a part of the information processing device according to the first embodiment.
  • FIG. 2 shows one example of the configuration of the attention layer 71 .
  • the processor 70 p performs the first process PR 1 after the embedding process 71 e.
  • the processor 70 p may be configured to perform multi-head attention processing 71 m.
  • the attention layer 71 may divide at least one of the first input data 11 , the second input data 12 , or the third input data 13 into a plurality of vectors.
  • the attention layer 71 may be configured to perform processing regarding at least two of the plurality of vectors in parallel. For example, processing results can be obtained at high speed.
  • the processor 70 p may include a normalization layer (a first normalization layer 71 a, a second normalization layer 71 b, etc.).
  • the normalization layer may be configured to perform normalization processing in at least a part of at least one of the first process PR 1 or the second process PR 2 .
  • processing regarding at least two of the plurality of vectors may be performed.
  • processing by the second normalization layer 71 b may be performed.
  • the processor 70 p may include a first skip connection portion 71 p and a second skip connection portion 71 q.
  • the first skip connection portion 71 p connects, for example, the result of processing regarding the at least two of the plurality of vectors (multi-head attention processing 71 m ) and the first value group 71 u input to the first normalization layer 71 a.
  • the connections may include, for example, residual connections. In the case of residual connection, for example, the result of adding the two is supplied to the second normalization layer 71 b.
  • the second skip connection portion 71 q adds the result of the processing by the second normalization layer 71 b processed by the neural network 71 n and the second value group 71 v input to the second normalization layer 71 b.
  • the second derived feature value FD 2 (see FIG. 1 ) is derived.
  • FIG. 3 is a schematic diagram illustrating a part of the information processing device according to the first embodiment.
  • FIG. 3 shows one example of the configuration of the attention layer 71 .
  • the processor 70 p may be configured to perform the embedding process 71 e described with reference to FIG. 2 .
  • the attention layer 71 at least one of the first input data 11 , the second input data 12 , or the third input data 13 is divided into a plurality of vectors. As shown in FIG. 3 , for example, after the embedding process 71 e, a process regarding the at least two of the plurality of vectors (multi-head attention process 71 m ) is performed.
  • the process regarding the at least two of the plurality of vectors is normalized, for example, in the first normalization layer 71 a.
  • processing by the first normalization layer 71 a is performed at least after the first process PR 1 .
  • the first skip connection portion 71 p connects the result of the processing in the first normalization layer 71 a and the first value group 71 u input to the multi-head attention processing 71 m.
  • the results of the processing by the first skip connection portion 71 p are processed by the neural network 71 n.
  • the result processed by the neural network 71 n is normalized by the second normalization layer 71 b.
  • the second skip connection portion 71 q adds the result of the processing in the second normalization layer 71 b and the second value group 71 v input to the neural network 71 n .
  • the second derived feature value FD 2 (see FIG. 1 ) is derived.
  • FIG. 4 is a schematic diagram illustrating an information processing device and an information recording/reproducing device according to a second embodiment.
  • an information processing device 70 A also includes the acquisitor 70 i and the processor 70 p. Self-attention is performed in the operation of the information processing device 70 A.
  • the configuration and operation of the information processing device 70 A except for this may be the same as the configuration and operation of the information processing device 70 .
  • the reproduced signal RS includes the first input data 11 .
  • the processor 70 p is perform the first process PR 1 and the second process PR 2 .
  • the processor 70 p calculates a first derived feature value FD 1 from the first inner product 31 of the first feature value F 11 based on the first input data 11 and the second feature value F 12 based on the first input data 11 .
  • the processor 70 p derives the second derived feature value FD 2 from the second inner product 32 of the third feature value F 13 based on the first input data 11 and the first derived feature value FD 1 .
  • the processor 70 p may be configured to further perform the third process PR 3 .
  • the processor 70 p outputs the bit decoding result BD 1 regarding the first input data 11 based on at least a part of the second derived feature value FD 2 .
  • the information processing device 70 A feature value extraction of the waveform data of the reproduced signal RS is performed based on the calculation of the inner product (first process PR 1 and second process PR 2 ). After the feature value extraction, the classification processing is performed using the fully connected NN. Thereby, for example, highly accurate decoding can be performed. For example, decoding at high speed and with high accuracy is possible. An information processing device that can improve processing accuracy can be provided.
  • the first input data 11 includes data regarding the decoding target bit and other bits.
  • the other bits include a first other bit and a second other bit.
  • the decoding target bit is between the first other bit and the second other bit.
  • the decoding target bit may be next to the first other bit.
  • the decoding target bit may be next to the second other bit.
  • Another bit may be provided between the decoding target bit and the first other bit.
  • Another bit may be provided between the decoding target bit and the second other bit.
  • the first other bit may include one or more bits.
  • the second other bit may include one or more bits.
  • the first other bit may be provided between the decoding target bit and the second other bit.
  • the processor 70 p is configured to perform neural network processing.
  • the processor 70 p may include the attention layer 71 and the classification layer 72 .
  • the attention layer 71 is configured to perform the first process PR 1 and the second process PR 2 .
  • the classification layer 72 is configured to perform the third process PR 3 .
  • the first feature value F 11 includes the “query” regarding the first input data 11 .
  • the second feature value F 12 includes the “key” regarding the first input data 11 .
  • the attention layer 71 includes the first neural network 11 N and the second neural network 12 N.
  • the attention layer 71 is configured to derive the “query” by processing the first input data 11 with the first neural network 11 N.
  • the attention layer 71 is configured to derive the “key” by processing the first input data 11 with the second neural network 12 N.
  • the third feature value F 13 includes the “value” regarding the first input data 11 .
  • the attention layer 71 may further include the third neural network 13 N.
  • the attention layer 71 may be configured to derive the “value” by processing the first input data 11 with the third neural network 13 N.
  • the attention layer 71 may further include the fourth neural network 14 N.
  • the attention layer 71 is configured to derive the second derived feature value FD 2 by processing the second inner product 32 with the fourth neural network 14 N.
  • the processing described with reference to FIG. 2 or FIG. 3 may be performed.
  • the information processing device 70 A uses one type of input data (first input data 11 ) as input for processing.
  • the processor 70 p may perform embedding processing 71 e of the reproduced signal RS (first input data 11 ).
  • the embedding process may include, for example, embedding location information.
  • the embedding process may include, for example, a class token used when classifying input data to perform attention processing.
  • the processor 70 p may be configured to perform multi-head attention processing 71 m.
  • the attention layer 71 divides the first input data 11 into a plurality of vectors.
  • the third embodiment relates to a magnetic recording/reproducing device 210 (see FIG. 1 ).
  • the magnetic recording/reproducing device 210 includes the information processing device (information processing device 70 or information processing device 70 A) according to the embodiment, the magnetic recording medium 80 , and the reproducing portion 80 R.
  • the reproducing portion 80 R is configured to reproduce the signal 80 S corresponding to information recorded on the magnetic recording medium 80 .
  • the fourth embodiment relates to a magnetic recording/reproducing system 310 (see FIG. 1 ).
  • the magnetic recording/reproducing system 310 includes the information processing device (the information processing device 70 or the information processing device 70 A) according to the embodiment.
  • the information processing device 70 may be provided at a location away from the magnetic recording medium 80 .
  • a signal obtained by AD converting the signal 80 S obtained from the magnetic recording medium 80 is supplied to the processor 70 p.
  • the processor 70 p may include a “Channel Estimation”, a waveform equalizer, and a maximum likelihood decoder.
  • the result obtained by the processor 70 p may be supplied to an error correction decoder (for example, a low-density parity-check code (LDPC)).
  • the processing result of the error correction decoder may be supplied to the processor 70 p . For example, an iterative process may be performed.
  • An information processing device comprising:
  • An information processing device comprising:
  • a magnetic recording/reproducing device comprising;
  • a magnetic recording/reproducing system comprising:
  • an information processing device a magnetic recording/reproducing device, and a magnetic recording/reproducing system that can improve processing accuracy can be provided.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Signal Processing For Digital Recording And Reproducing (AREA)

Abstract

According to one embodiment, an information processing device includes an acquisitor and a processor. The acquisitor is configured to acquire a reproduced signal based on a signal obtained from a magnetic recording medium. The reproduced signal includes first input data, second input data, and third input data. The processor is configured to perform a first process and a second process. In the first process, the processor is configured to derive a first derived feature value from a first inner product of a first feature value based on the first input data and a second feature value based on the second input data. In the second process, the processor is configured to derive a second derived feature value from a second inner product of a third feature value based on the third input data and the first derived feature value.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is based upon and claims the benefit of priority from Japanese Patent Application No. 2023-149545, filed on Sep. 14, 2023; the entire contents of which are incorporated herein by reference.
  • FIELD
  • Embodiments described herein relate generally to an information processing device, a magnetic recording/reproducing device, and a magnetic recording/reproducing system.
  • BACKGROUND
  • For example, information recorded on a magnetic recording medium is reproduced by an information processing device. For example, by improving the processing accuracy of an information processing device, magnetic recording density can be improved.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a schematic diagram illustrating an information processing device and an information recording/reproducing device according to a first embodiment;
  • FIG. 2 is a schematic diagram illustrating a part of the information processing device according to the first embodiment;
  • FIG. 3 is a schematic diagram illustrating a part of the information processing device according to the first embodiment; and
  • FIG. 4 is a schematic diagram illustrating an information processing device and an information recording/reproducing device according to a second embodiment.
  • DETAILED DESCRIPTION
  • According to one embodiment, an information processing device includes an acquisitor and a processor. The acquisitor is configured to acquire a reproduced signal based on a signal obtained from a magnetic recording medium. The reproduced signal includes first input data, second input data, and third input data. The processor is configured to perform a first process and a second process. In the first process, the processor is configured to derive a first derived feature value from a first inner product of a first feature value based on the first input data and a second feature value based on the second input data. In the second process, the processor is configured to derive a second derived feature value from a second inner product of a third feature value based on the third input data and the first derived feature value.
  • Various embodiments are described below with reference to the accompanying drawings.
  • In the specification and drawings, components similar to those described previously or illustrated in an antecedent drawing are marked with like reference numerals, and a detailed description is omitted as appropriate.
  • First Embodiment
  • FIG. 1 is a schematic diagram illustrating an information processing device and an information recording/reproducing device according to a first embodiment.
  • As shown in FIG. 1 , a magnetic recording/reproducing device 210 according to the embodiment includes an information processing device 70, a magnetic recording medium 80, and a reproducing portion 80R. The magnetic recording medium 80 and the reproducing portion 80R may be included in a recording section 80D, and the recording section 80D may be included in, for example, a magnetic disk (HDD: Hard Disk Drive). The magnetic recording/reproducing device 210 may further include, for example, an SSD (Solid State Drive).
  • The reproducing portion 80R is configured to reproduce a signal 80S corresponding to the information recorded on the magnetic recording medium 80. The reproducing portion 80R includes, for example, a magnetic head 80H. The signal 80S corresponding to the reproduced information is output from the reproducing portion 80R. The signal 80S is an electrical signal. Information may be recorded on the magnetic recording medium 80 by the magnetic head 80H.
  • The information processing device 70 includes an acquisitor 70 i and a processor 70 p. The acquisitor 70 i is configured to acquire a reproduced signal RS based on the signal 80S obtained from the magnetic recording medium 80. The reproduced signal RS may be based on a signal obtained by amplifying the signal 80S. The reproduced signal RS may be a digital signal obtained by AD (Analog to Digital) conversion of the signal 80S. The acquisitor 70 i may be, for example, an interface.
  • The reproduced signal RS acquired by the acquisitor 70 i is processed by the processor 70 p. The processor 70 p includes, for example, an input portion 70 pi and an output portion 70 po. The reproduced signal RS is supplied to the input portion 70 pi. The processing result in the processor 70 p is output from the output portion 70 po.
  • As shown in FIG. 1 , the reproduced signal RS includes first input data 11, second input data 12, and third input data 13. The processor 70 p is configured to perform a first process PR1 and a second process PR2. In one example, the first input data 11 includes data regarding a decoding target bit. For example, the second input data 12 and the third input data 13 include data regarding other bits.
  • In the first process PR1, the processor 70 p derives a first derived feature value FD1 from a first inner product 31 of a first feature value F11 based on the first input data 11 and a second feature value F12 based on the second input data 12.
  • In the second process PR2, the processor 70 p derives a second derived feature value FD2 from a second inner product 32 of a third feature value F13 based on the third input data 13 and the first derived feature value FD1.
  • By such first process PR1 and second process PR2, the reproduced signal RS corresponding to information recorded on the magnetic recording medium 80 can be decoded with high accuracy.
  • As shown in FIG. 1 , the processor 70 p is configured to further perform a third process PR3. In the third process PR3, the processor 70 p outputs a bit decoding result BD1 regarding the decoding target bits included in the first input data 11 based on at least a part of the second derived feature value FD2. The bit decoding result BD1 being obtained is output from the output portion 70 po. In one example, bit decoding result BD1 corresponds to information related to 0/1 determination. The information related to the 0/1 determination may be 0 or 1, for example. In one example, the bit decoding result BD1 may be information that can be used for 0/1 determination. The information that can be used for the 0/1 determination may include, for example, the probability of being 0 or 1. The information that can be used for the 0/1 determination may be, for example, the likelihood regarding 0 or 1.
  • The processor 70 p is configured to perform neural network processing (NN processing). As shown in FIG. 1 , the processor 70 p may include an attention layer 71 and a classification layer 72. The attention layer 71 is configured to perform the first process PR1 and the second process PR2. The classification layer 72 is configured to perform the third process PR3. The classification layer 72 includes a fully connected neural network (FC-NN).
  • In the embodiments, the neural network may be a component that includes parameters whose values can be adjusted (e.g., learned) based on information such as data. In the embodiment, the neural network may include a perceptron.
  • In the embodiment, the reproduced signal RS obtained from the magnetic recording medium 80 is processed by attention processing including two inner product calculations. Through such processing, feature value extraction of the waveform data of the reproduced signal RS is performed (first process PR1 and second process PR2). After the feature value extraction, the classification processing is performed using a fully connected NN. The classification process corresponds to, for example, a process of converting information extracted by feature value extraction into information related to 0/1 determination. Thereby, for example, highly accurate decoding can be performed. For example, a decoding at high speed and with high accuracy can be performed. An information processing device that can improve processing accuracy can be provided.
  • In the embodiment, by calculation based on the inner product, it is possible to appropriately extract the feature value that takes into account the relationship of input data. Parallel processing is easy in the processing based on inner products. In the embodiment, high-accuracy decoding can be performed at high speed due to high speed through parallel calculation and appropriate feature values. For example, in a reference example NN model that takes into account the relationship between input data, an RNN (Recurrent Neural Network) (for example, LSTM: Long Short Term Memory) is applied. In such a reference example, since parallel calculation is difficult, the calculation time is long. In contrast, in the embodiment, high-speed processing is possible.
  • In the embodiment, the first input data 11, the second input data 12, the third input data 13, etc. correspond to information recorded on the magnetic recording medium 80. For example, the first input data 11 includes data regarding a decoding target bit. The second input data 12 includes data regarding a first other bit. The third input data 13 includes data regarding a second other bit. In one example, the decoding target bit is between the first other bit and the second other bit. In one example, the decoding target bit may be next to the first other bit. The decoding target bit may be next to the second other bit. Another bit may be provided between the decoding target bit and the first other bit. Another bit may be provided between the decoding target bit and the second other bit. The first other bit may include one or more bits. The second other bit may include one or more bits. In another example, the first other bit may be provided between the decoding target bit and the second other bit.
  • At least one of the first input data 11, the second input data 12, or the third input data 13 may be determined by, for example, a method of the attention processing. For example, self-attention may be applied, as described below.
  • As shown in FIG. 1 , for example, the first feature value F11 includes “query” regarding the first input data 11. For example, the second feature value F12 includes a “key” regarding the second input data 12.
  • As shown in FIG. 1 , the attention layer 71 may include a first neural network 11N and a second neural network 12N. The attention layer 71 is configured to derive the “query” (first feature value F11) by processing the first input data 11 with the first neural network 11N. The attention layer 71 is configured to derive the “key” (second feature value F12) with processing the second input data 12 by the second neural network 12N.
  • The first derived feature value FD1 includes, for example, a first weight indicating the relationship between the first input data 11 and the second input data 12.
  • For example, the third feature value F13 includes “value” regarding the third input data 13. For example, the attention layer 71 further includes a third neural network 13N. The attention layer 71 is configured to derive the “value” (third feature value F13) by processing the third input data 13 with the third neural network 13N.
  • As shown in FIG. 1 , the attention layer 71 may further include a fourth neural network 14N. The attention layer 71 is configured to derive the second derived feature value FD2 by processing the second inner product 32 with the fourth neural network 14N.
  • In the embodiment, interference between bits is taken into consideration, for example, in HDD signal reproduction. In the attention processing according to the embodiment, it is possible to extract feature values by considering the relationship of input data. By using attention processing for HDD signal reproduced signal reproduction can be performed in consideration of interference (relationship) between bits.
  • As described above, in the embodiment, the reproduced signal RS obtained from the magnetic recording medium 80 is processed by attention processing including two inner product calculations. Thereby, highly accurate decoding can be performed. An information processing device that can improve processing accuracy can be provided.
  • The number of attention layers 71 may be one, or two or more.
  • As shown in FIG. 1 , the processor 70 p may be configured to perform embedding processing 71 e. For example, the processor 70 p performs embedding processing 71 e of the reproduced signal RS (first input data 11, second input data 12, and third input data 13). The embedding process may include, for example, embedding location information. The embedding process may include, for example, a class token used when classifying input data to perform attention processing.
  • FIG. 2 is a schematic diagram illustrating a part of the information processing device according to the first embodiment.
  • FIG. 2 shows one example of the configuration of the attention layer 71. The processor 70 p performs the first process PR1 after the embedding process 71 e.
  • As shown in FIG. 2 , the processor 70 p may be configured to perform multi-head attention processing 71 m. For example, the attention layer 71 may divide at least one of the first input data 11, the second input data 12, or the third input data 13 into a plurality of vectors. The attention layer 71 may be configured to perform processing regarding at least two of the plurality of vectors in parallel. For example, processing results can be obtained at high speed.
  • As shown in FIG. 2 , the processor 70 p may include a normalization layer (a first normalization layer 71 a, a second normalization layer 71 b, etc.). The normalization layer may be configured to perform normalization processing in at least a part of at least one of the first process PR1 or the second process PR2.
  • As shown in FIG. 2 , after the processing by the first normalization layer 71 a, processing regarding at least two of the plurality of vectors (multi-head attention processing 71 m) may be performed. After processing on the at least two of the plurality of vectors, processing by the second normalization layer 71 b may be performed.
  • As shown in FIG. 2 , the processor 70 p may include a first skip connection portion 71 p and a second skip connection portion 71 q. The first skip connection portion 71 p connects, for example, the result of processing regarding the at least two of the plurality of vectors (multi-head attention processing 71 m) and the first value group 71 u input to the first normalization layer 71 a. The connections may include, for example, residual connections. In the case of residual connection, for example, the result of adding the two is supplied to the second normalization layer 71 b.
  • The second skip connection portion 71 q, for example, adds the result of the processing by the second normalization layer 71 b processed by the neural network 71 n and the second value group 71 v input to the second normalization layer 71 b. By this adding process, the second derived feature value FD2 (see FIG. 1 ) is derived.
  • FIG. 3 is a schematic diagram illustrating a part of the information processing device according to the first embodiment.
  • FIG. 3 shows one example of the configuration of the attention layer 71. As shown in FIG. 3 , the processor 70 p may be configured to perform the embedding process 71 e described with reference to FIG. 2 .
  • For example, in the attention layer 71, at least one of the first input data 11, the second input data 12, or the third input data 13 is divided into a plurality of vectors. As shown in FIG. 3 , for example, after the embedding process 71 e, a process regarding the at least two of the plurality of vectors (multi-head attention process 71 m) is performed.
  • The process regarding the at least two of the plurality of vectors (multi-head attention processing 71 m) is normalized, for example, in the first normalization layer 71 a. In this example, processing by the first normalization layer 71 a is performed at least after the first process PR1.
  • For example, the first skip connection portion 71 p connects the result of the processing in the first normalization layer 71 a and the first value group 71 u input to the multi-head attention processing 71 m.
  • As shown in FIG. 3 , the results of the processing by the first skip connection portion 71 p are processed by the neural network 71 n. The result processed by the neural network 71 n is normalized by the second normalization layer 71 b.
  • The second skip connection portion 71 q, for example, adds the result of the processing in the second normalization layer 71 b and the second value group 71 v input to the neural network 71 n. By this adding process, the second derived feature value FD2 (see FIG. 1 ) is derived.
  • Second Embodiment
  • FIG. 4 is a schematic diagram illustrating an information processing device and an information recording/reproducing device according to a second embodiment.
  • As shown in FIG. 4 , an information processing device 70A according to the embodiment also includes the acquisitor 70 i and the processor 70 p. Self-attention is performed in the operation of the information processing device 70A. The configuration and operation of the information processing device 70A except for this may be the same as the configuration and operation of the information processing device 70.
  • As shown in FIG. 4 , the reproduced signal RS includes the first input data 11. The processor 70 p is perform the first process PR1 and the second process PR2.
  • In the first process PR1, the processor 70 p calculates a first derived feature value FD1 from the first inner product 31 of the first feature value F11 based on the first input data 11 and the second feature value F12 based on the first input data 11.
  • In the second process PR2, the processor 70 p derives the second derived feature value FD2 from the second inner product 32 of the third feature value F13 based on the first input data 11 and the first derived feature value FD1.
  • The processor 70 p may be configured to further perform the third process PR3. In the third process PR3, the processor 70 p outputs the bit decoding result BD1 regarding the first input data 11 based on at least a part of the second derived feature value FD2.
  • In the information processing device 70A, feature value extraction of the waveform data of the reproduced signal RS is performed based on the calculation of the inner product (first process PR1 and second process PR2). After the feature value extraction, the classification processing is performed using the fully connected NN. Thereby, for example, highly accurate decoding can be performed. For example, decoding at high speed and with high accuracy is possible. An information processing device that can improve processing accuracy can be provided.
  • For example, the first input data 11 includes data regarding the decoding target bit and other bits. In one example, the other bits include a first other bit and a second other bit. In one example, the decoding target bit is between the first other bit and the second other bit. In one example, the decoding target bit may be next to the first other bit. The decoding target bit may be next to the second other bit. Another bit may be provided between the decoding target bit and the first other bit. Another bit may be provided between the decoding target bit and the second other bit. The first other bit may include one or more bits. The second other bit may include one or more bits. In another example, the first other bit may be provided between the decoding target bit and the second other bit.
  • As shown in FIG. 4 , also in the information processing device 70A, the processor 70 p is configured to perform neural network processing. The processor 70 p may include the attention layer 71 and the classification layer 72. The attention layer 71 is configured to perform the first process PR1 and the second process PR2. The classification layer 72 is configured to perform the third process PR3.
  • For example, the first feature value F11 includes the “query” regarding the first input data 11. For example, the second feature value F12 includes the “key” regarding the first input data 11.
  • For example, the attention layer 71 includes the first neural network 11N and the second neural network 12N. The attention layer 71 is configured to derive the “query” by processing the first input data 11 with the first neural network 11N. The attention layer 71 is configured to derive the “key” by processing the first input data 11 with the second neural network 12N.
  • For example, the third feature value F13 includes the “value” regarding the first input data 11. For example, the attention layer 71 may further include the third neural network 13N. The attention layer 71 may be configured to derive the “value” by processing the first input data 11 with the third neural network 13N.
  • The attention layer 71 may further include the fourth neural network 14N. The attention layer 71 is configured to derive the second derived feature value FD2 by processing the second inner product 32 with the fourth neural network 14N.
  • In the information processing device 70A, the processing described with reference to FIG. 2 or FIG. 3 may be performed. In this case, the information processing device 70A uses one type of input data (first input data 11) as input for processing.
  • In the information processing device 70A, the processor 70 p may perform embedding processing 71 e of the reproduced signal RS (first input data 11). The embedding process may include, for example, embedding location information. The embedding process may include, for example, a class token used when classifying input data to perform attention processing. For example, the processor 70 p may be configured to perform multi-head attention processing 71 m. For example, the attention layer 71 divides the first input data 11 into a plurality of vectors.
  • Third Embodiment
  • The third embodiment relates to a magnetic recording/reproducing device 210 (see FIG. 1 ). The magnetic recording/reproducing device 210 includes the information processing device (information processing device 70 or information processing device 70A) according to the embodiment, the magnetic recording medium 80, and the reproducing portion 80R. The reproducing portion 80R is configured to reproduce the signal 80S corresponding to information recorded on the magnetic recording medium 80. In the magnetic recording/reproducing device 210, it is possible to improve processing accuracy.
  • Fourth Embodiment
  • The fourth embodiment relates to a magnetic recording/reproducing system 310 (see FIG. 1 ). The magnetic recording/reproducing system 310 includes the information processing device (the information processing device 70 or the information processing device 70A) according to the embodiment. In the magnetic recording/reproducing system 310, it is possible to processing improve accuracy. In the magnetic recording/reproducing system 310, the information processing device 70 may be provided at a location away from the magnetic recording medium 80.
  • In the embodiment, a signal obtained by AD converting the signal 80S obtained from the magnetic recording medium 80 is supplied to the processor 70 p. The processor 70 p may include a “Channel Estimation”, a waveform equalizer, and a maximum likelihood decoder. The result obtained by the processor 70 p may be supplied to an error correction decoder (for example, a low-density parity-check code (LDPC)). The processing result of the error correction decoder may be supplied to the processor 70 p. For example, an iterative process may be performed.
  • The embodiments may include the following Technical proposals:
  • Technical Proposal 1
  • An information processing device, comprising:
      • an acquisitor; and
      • a processor,
      • the acquisitor being configured to acquire a reproduced signal based on a signal obtained from a magnetic recording medium,
      • the reproduced signal including first input data, second input data, and third input data,
      • the processor being configured to perform a first process and a second process,
      • in the first process, the processor being configured to derive a first derived feature value from a first inner product of a first feature value based on the first input data and a second feature value based on the second input data, and
      • in the second process, the processor being configured to derive a second derived feature value from a second inner product of a third feature value based on the third input data and the first derived feature value.
    Technical Proposal 2
  • The information processing device according to Technical proposal 1, wherein
      • the processor is configured to further perform a third process,
      • the first input data includes data regarding a decoding target bit, and
      • in the third process, the processor is configured to output a bit decoding result regarding the decoding target bit based on at least a part of the second derived feature value.
    Technical Proposal 3
  • The information processing device according to Technical proposal 2, wherein
      • the processor is configured to perform neural network processing,
      • the processor includes an attention layer and a classification layer,
      • the attention layer is configured to perform the first process and the second process, and
      • the classification layer is configured to perform the third process.
    Technical Proposal 4
  • The information processing device according to Technical proposal 3, wherein
      • the first feature value includes a “query” regarding the first input data, and
      • the second feature value includes a “key” regarding the second input data.
    Technical Proposal 5
  • The information processing device according to Technical proposal 4, wherein
      • the attention layer includes a first neural network and a second neural network,
      • the attention layer is configured to derive the “query” by processing the first input data with the first neural network, and
      • the attention layer is configured to derive the “key” by processing the second input data with the second neural network.
    Technical Proposal 6
  • The information processing device according to Technical proposal 4 or 5, wherein
      • the third feature value includes a “value” regarding the third input data.
    Technical Proposal 7
  • The information processing device according to Technical proposal 6, wherein
      • the attention layer further includes a third neural network, and
      • the attention layer is configured to derive the “value” by processing the third input data with the third neural network.
    Technical Proposal 8
  • The information processing device according to Technical proposal 7, wherein
      • the attention layer further includes a fourth neural network, and
      • the attention layer is configured to derive the second derived feature value by processing the second inner product with the fourth neural network.
    Technical Proposal 9
  • The information processing device according to any one of Technical proposals 3-8, wherein
      • the processor is configured to perform an embedding process to embedding the first input data, the second input data, and the third input data, and
      • the processor is configured to perform the first processing after the embedding processing.
    Technical Proposal 10
  • The information processing device according to any one of Technical proposals 3-9, wherein
      • the attention layer is configured to divide at least one of the first input data, the second input data, or the third input data into a plurality of vectors, and
      • the attention layer is configured to perform processing on at least two of the plurality of vectors in parallel.
    Technical Proposal 11
  • An information processing device, comprising:
      • an acquisitor; and
      • a processor,
      • the acquisitor being configured to acquire a reproduced signal based on a signal obtained from a magnetic recording medium,
      • the reproduced signal including first input data,
      • the processor being configured to perform a first process and a second process,
      • in the first process, the processor being configured to derive a first derived feature value from a first inner product of a first feature value based on the first input data and a second feature value based on the first input data, and
      • in the second process, the processor being configured to derive a second derived feature value from a second inner product of a third feature value based on the first input data and the first derived feature value.
    Technical Proposal 12
  • The information processing device according to Technical proposal 11, wherein
      • the processor is configured to further perform a third process,
      • the first input data includes data regarding a decoding target bit and other bits, and
      • in the third process, the processor is configured to output a bit decoding result regarding the decoding target bit based on at least a part of the second derived feature value.
    Technical Proposal 13
  • The information processing device according to Technical proposal 12, wherein
      • the processor is configured to perform neural network processing,
      • the processor includes an attention layer and a classification layer,
      • the attention layer is configured to perform the first process and the second process, and
      • the classification layer is configured to perform the third process.
    Technical Proposal 14
  • The information processing device according to Technical proposal 13, wherein
      • the first feature value includes a “query” regarding the first input data, and
      • the second feature value includes a “key” regarding the first input data.
    Technical Proposal 15
  • The information processing device according to Technical proposal 14, wherein
      • the attention layer includes a first neural network and a second neural network,
      • the attention layer is configured to derive the “query” by processing the first input data with the first neural network, and
      • the attention layer is configured to derive the “key” by processing the first input data with the second neural network.
    Technical Proposal 16
  • The information processing device according to Technical proposal 14 or 15, wherein
      • the third feature value includes a “value” regarding the first input data.
    Technical Proposal 17
  • The information processing device according to Technical proposal 16, wherein
      • the attention layer further includes a third neural network, and
      • the attention layer is configured to derive the “value” by processing the first input data with the third neural network.
    Technical Proposal 18
  • The information processing device according to Technical proposal 17, wherein
      • the attention layer further includes a fourth neural network, and
      • the attention layer is configured to derive the second derived feature value by processing the second inner product with the fourth neural network.
    Technical Proposal 19
  • A magnetic recording/reproducing device, comprising;
      • the information processing device according to any one of Technical proposals 1-18;
      • the magnetic recording medium; and
      • a reproducing portion configured to reproduce the signal corresponding to information recorded on the magnetic recording medium.
    Technical Proposal 20
  • A magnetic recording/reproducing system, comprising:
      • the information processing device according to any one of Technical proposals 1-18.
  • According to the embodiment, an information processing device, a magnetic recording/reproducing device, and a magnetic recording/reproducing system that can improve processing accuracy can be provided.
  • Hereinabove, exemplary embodiments of the invention are described with reference to specific examples. However, the embodiments of the invention are not limited to these specific examples. For example, one skilled in the art may similarly practice the invention by appropriately selecting specific configurations of components included in the information processing devices, the magnetic recording/reproducing devices, and the magnetic recording/reproducing systems such as acquisitors, processors, magnetic recording mediums, etc., from known art. Such practice is included in the scope of the invention to the extent that similar effects thereto are obtained.
  • Further, any two or more components of the specific examples may be combined within the extent of technical feasibility and are included in the scope of the invention to the extent that the purport of the invention is included.
  • Moreover, all information processing devices, all magnetic and recording/reproducing devices, all magnetic recording/reproducing systems practicable by an appropriate design modification by one skilled in the art based on the information processing devices, the magnetic recording/reproducing devices, and the magnetic recording/reproducing systems described above as embodiments of the invention also are within the scope of the invention to the extent that the purport of the invention is included.
  • Various other variations and modifications can be conceived by those skilled in the art within the spirit of the invention, and it is understood that such variations and modifications are also encompassed within the scope of the invention.
  • While certain embodiments have been described, these embodiments have been presented by way of example only, and are not intended to limit the scope of the inventions. Indeed, the novel embodiments described herein may be embodied in a variety of other forms; furthermore, various omissions, substitutions and changes in the form of the embodiments described herein may be made without departing from the spirit of the inventions. The accompanying claims and their equivalents are intended to cover such forms or modifications as would fall within the scope and spirit of the invention.

Claims (20)

What is claimed is:
1. An information processing device, comprising:
an acquisitor; and
a processor,
the acquisitor being configured to acquire a reproduced signal based on a signal obtained from a magnetic recording medium,
the reproduced signal including first input data, second input data, and third input data,
the processor being configured to perform a first process and a second process,
in the first process, the processor being configured to derive a first derived feature value from a first inner product of a first feature value based on the first input data and a second feature value based on the second input data, and
in the second process, the processor being configured to derive a second derived feature value from a second inner product of a third feature value based on the third input data and the first derived feature value.
2. The device according to claim 1, wherein
the processor is configured to further perform a third process,
the first input data includes data regarding a decoding target bit, and
in the third process, the processor is configured to output a bit decoding result regarding the decoding target bit based on at least a part of the second derived feature value.
3. The device according to claim 2, wherein
the processor is configured to perform neural network processing,
the processor includes an attention layer and a classification layer,
the attention layer is configured to perform the first process and the second process, and
the classification layer is configured to perform the third process.
4. The device according to claim 3, wherein
the first feature value includes a “query” regarding the first input data, and
the second feature value includes a “key” regarding the second input data.
5. The device according to claim 4, wherein
the attention layer includes a first neural network and a second neural network,
the attention layer is configured to derive the “query” by processing the first input data with the first neural network, and
the attention layer is configured to derive the “key” by processing the second input data with the second neural network.
6. The device according to claim 4, wherein
the third feature value includes a “value” regarding the third input data.
7. The device according to claim 6, wherein
the attention layer further includes a third neural network, and
the attention layer is configured to derive the “value” by processing the third input data with the third neural network.
8. The device according to claim 7, wherein
the attention layer further includes a fourth neural network, and
the attention layer is configured to derive the second derived feature value by processing the second inner product with the fourth neural network.
9. The device according to claim 3, wherein
the processor is configured to perform an embedding process to embedding the first input data, the second input data, and the third input data, and
the processor is configured to perform the first processing after the embedding processing.
10. The device according to claim 3, wherein
the attention layer is configured to divide at least one of the first input data, the second input data, or the third input data into a plurality of vectors, and
the attention layer is configured to perform processing on at least two of the plurality of vectors in parallel.
11. An information processing device, comprising:
an acquisitor; and
a processor,
the acquisitor being configured to acquire a reproduced signal based on a signal obtained from a magnetic recording medium,
the reproduced signal including first input data,
the processor being configured to perform a first process and a second process,
in the first process, the processor being configured to derive a first derived feature value from a first inner product of a first feature value based on the first input data and a second feature value based on the first input data, and
in the second process, the processor being configured to derive a second derived feature value from a second inner product of a third feature value based on the first input data and the first derived feature value.
12. The device according to claim 11, wherein
the processor is configured to further perform a third process,
the first input data includes data regarding a decoding target bit and other bits, and
in the third process, the processor is configured to output a bit decoding result regarding the decoding target bit based on at least a part of the second derived feature value.
13. The device according to claim 12, wherein
the processor is configured to perform neural network processing,
the processor includes an attention layer and a classification layer,
the attention layer is configured to perform the first process and the second process, and
the classification layer is configured to perform the third process.
14. The device according to claim 13, wherein
the first feature value includes a “query” regarding the first input data, and
the second feature value includes a “key” regarding the first input data.
15. The device according to claim 14, wherein
the attention layer includes a first neural network and a second neural network,
the attention layer is configured to derive the “query” by processing the first input data with the first neural network, and
the attention layer is configured to derive the “key” by processing the first input data with the second neural network.
16. The device according to claim 14, wherein
the third feature value includes a “value” regarding the first input data.
17. The device according to claim 16, wherein
the attention layer further includes a third neural network, and
the attention layer is configured to derive the “value” by processing the first input data with the third neural network.
18. The device according to claim 17, wherein
the attention layer further includes a fourth neural network, and
the attention layer is configured to derive the second derived feature value by processing the second inner product with the fourth neural network.
19. A magnetic recording/reproducing device, comprising;
the information processing device according to claim 1;
the magnetic recording medium; and
a reproducing portion configured to reproduce the signal corresponding to information recorded on the magnetic recording medium.
20. A magnetic recording/reproducing system, comprising:
the information processing device according to claim 1.
US18/740,842 2023-09-14 2024-06-12 Information processing device, magnetic recording/reproducing device, and magnetic recording/reproducing system Pending US20250095683A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2023149545A JP2025042491A (en) 2023-09-14 2023-09-14 Information processing device, magnetic recording and reproducing device, and magnetic recording and reproducing system
JP2023-149545 2023-09-14

Publications (1)

Publication Number Publication Date
US20250095683A1 true US20250095683A1 (en) 2025-03-20

Family

ID=94975730

Family Applications (1)

Application Number Title Priority Date Filing Date
US18/740,842 Pending US20250095683A1 (en) 2023-09-14 2024-06-12 Information processing device, magnetic recording/reproducing device, and magnetic recording/reproducing system

Country Status (2)

Country Link
US (1) US20250095683A1 (en)
JP (1) JP2025042491A (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE112020006001T5 (en) 2020-03-02 2022-11-03 Hitachi High-Tech Corporation CHARGED DETECTOR, CHARGED RADIATION DEVICE, RADIATION DETECTOR AND RADIATION DETECTION DEVICE

Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5898531A (en) * 1994-09-01 1999-04-27 Cannon Kabushiki Kaisha Signal processing apparatus for digitally modulating input data
US6163423A (en) * 1996-12-16 2000-12-19 Lg Electronics Inc. Synchronizing signal detector for magnetic recording/reproducing apparatus and synchronizing signal detecting method thereof
US20030099052A1 (en) * 2001-10-24 2003-05-29 Viswanath Annampedu Servo data detection in presence of radial incoherence using multiple data detectors
US6819724B2 (en) * 1999-12-24 2004-11-16 Victor Company Of Japan, Limited Viterbi decoder and Viterbi decoding method
US20060152845A1 (en) * 2005-01-12 2006-07-13 Kabushiki Kaisha Toshiba Magnetic recording medium and magnetic recording/reproducing apparatus
US7787534B2 (en) * 2006-05-30 2010-08-31 Fujitsu Limited System and method for adjusting offset compensation applied to a signal
US20100232071A1 (en) * 2009-03-16 2010-09-16 Toshiba Storage Device Corporation Perpendicular magnetic recording medium, method of manufacturing the same, and magnetic recording/playback apparatus
US20130083873A1 (en) * 2011-09-30 2013-04-04 Hajime Ishihara Information reproduction apparatus and information reproduction method
US20200099453A1 (en) * 2016-03-04 2020-03-26 Inphi Corporation Pam4 transceivers for high-speed communication
US20210049411A1 (en) * 2019-08-15 2021-02-18 Canon Kabushiki Kaisha Information processing apparatus, information processing method, and non-transitory computer-readable storage medium
US20210312279A1 (en) * 2020-04-07 2021-10-07 Fujitsu Limited Information processing apparatus and information processing method
US20220182438A1 (en) * 2020-12-04 2022-06-09 Kabushiki Kaisha Toshiba Information processing system
US20230298619A1 (en) * 2022-03-17 2023-09-21 Kabushiki Kaisha Toshiba Information processing device, magnetic recording and reproducing device, and magnetic recording and reproducing system
US20240054404A1 (en) * 2022-08-12 2024-02-15 Canon Kabushiki Kaisha Information processing system, information processing apparatus, method for training inference model, and non-transitory storage medium
US20240071421A1 (en) * 2022-08-31 2024-02-29 Kabushiki Kaisha Toshiba Magnetic reproduction processing device, magnetic recording/reproducing device, and magnetic reproducing method
US20240170011A1 (en) * 2022-11-17 2024-05-23 Kabushiki Kaisha Toshiba Data processor, magnetic recording/reproducing device, and magnetic recording/reproducing system
US20250021436A1 (en) * 2023-07-12 2025-01-16 Kabushiki Kaisha Toshiba Magnetic reproducing processing device, magnetic recording/reproducing device and magnetic reproducing method
US20250272457A1 (en) * 2024-02-28 2025-08-28 Canon Kabushiki Kaisha Information processing system, information processing method, recording medium

Patent Citations (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5898531A (en) * 1994-09-01 1999-04-27 Cannon Kabushiki Kaisha Signal processing apparatus for digitally modulating input data
US6163423A (en) * 1996-12-16 2000-12-19 Lg Electronics Inc. Synchronizing signal detector for magnetic recording/reproducing apparatus and synchronizing signal detecting method thereof
US6819724B2 (en) * 1999-12-24 2004-11-16 Victor Company Of Japan, Limited Viterbi decoder and Viterbi decoding method
US20030099052A1 (en) * 2001-10-24 2003-05-29 Viswanath Annampedu Servo data detection in presence of radial incoherence using multiple data detectors
US20060152845A1 (en) * 2005-01-12 2006-07-13 Kabushiki Kaisha Toshiba Magnetic recording medium and magnetic recording/reproducing apparatus
US7787534B2 (en) * 2006-05-30 2010-08-31 Fujitsu Limited System and method for adjusting offset compensation applied to a signal
US20100232071A1 (en) * 2009-03-16 2010-09-16 Toshiba Storage Device Corporation Perpendicular magnetic recording medium, method of manufacturing the same, and magnetic recording/playback apparatus
US20130083873A1 (en) * 2011-09-30 2013-04-04 Hajime Ishihara Information reproduction apparatus and information reproduction method
US20200099453A1 (en) * 2016-03-04 2020-03-26 Inphi Corporation Pam4 transceivers for high-speed communication
US20210049411A1 (en) * 2019-08-15 2021-02-18 Canon Kabushiki Kaisha Information processing apparatus, information processing method, and non-transitory computer-readable storage medium
US20210312279A1 (en) * 2020-04-07 2021-10-07 Fujitsu Limited Information processing apparatus and information processing method
US20220182438A1 (en) * 2020-12-04 2022-06-09 Kabushiki Kaisha Toshiba Information processing system
US11765223B2 (en) * 2020-12-04 2023-09-19 Kabushiki Kaisha Toshiba Information processing system
US20230298619A1 (en) * 2022-03-17 2023-09-21 Kabushiki Kaisha Toshiba Information processing device, magnetic recording and reproducing device, and magnetic recording and reproducing system
US20240054404A1 (en) * 2022-08-12 2024-02-15 Canon Kabushiki Kaisha Information processing system, information processing apparatus, method for training inference model, and non-transitory storage medium
US20240071421A1 (en) * 2022-08-31 2024-02-29 Kabushiki Kaisha Toshiba Magnetic reproduction processing device, magnetic recording/reproducing device, and magnetic reproducing method
US20240170011A1 (en) * 2022-11-17 2024-05-23 Kabushiki Kaisha Toshiba Data processor, magnetic recording/reproducing device, and magnetic recording/reproducing system
US20250021436A1 (en) * 2023-07-12 2025-01-16 Kabushiki Kaisha Toshiba Magnetic reproducing processing device, magnetic recording/reproducing device and magnetic reproducing method
US12367100B2 (en) * 2023-07-12 2025-07-22 Kabushiki Kaisha Toshiba Magnetic reproducing processing device, magnetic recording/reproducing device and magnetic reproducing method
US20250272457A1 (en) * 2024-02-28 2025-08-28 Canon Kabushiki Kaisha Information processing system, information processing method, recording medium

Also Published As

Publication number Publication date
JP2025042491A (en) 2025-03-27

Similar Documents

Publication Publication Date Title
JP5251000B2 (en) Error correction circuit and medium storage device
US20250095683A1 (en) Information processing device, magnetic recording/reproducing device, and magnetic recording/reproducing system
Wang et al. Improved multitrack detection with hybrid 2-D equalizer and modified Viterbi detector
US3320598A (en) Self-clocking complementary redundant recording system
US7830284B2 (en) Entropy encoding apparatus, entropy encoding method, and computer program
CN101523488B (en) Apparatus and method for processing optical information using low density parity check code
US11942120B2 (en) Information processing device, magnetic recording and reproducing device, and magnetic recording and reproducing system
US10951234B2 (en) Non-linear LLR look-up tables
US20080123783A1 (en) Run length limiting apparatus and run length limiting method
JP2015049923A (en) Systems and methods for multi-level encoding and decoding
TW536872B (en) Viterbi decoding device and method for multi-input data and multi-output data
US20080025178A1 (en) Data recording apparatus, recording medium and error detection method
JP2002057584A (en) Data reproducer and method for reproduction of data
US8618963B2 (en) Bit converter and bit conversion method
US7924522B2 (en) Apparatus and method of detecting error symbol, and disk drive apparatus using the same
EP2214173B1 (en) Method of evaluating reproduced signals and optical disc drive
JP3081047B2 (en) Coding method for reducing DC voltage component in data stream of digital signal
JP4916728B2 (en) Encoding device, signal processing device, and storage system
US11791841B2 (en) Bit flipping device and method and computer readable program for the same
US9824724B2 (en) Magnetic recording device capable of implementing read after write
JP3271187B2 (en) Soft decision decoding circuit
KR101848199B1 (en) data storage systems and method
JP2008198239A (en) Encoding device, decoding device, and recorded information reader
US20140344642A1 (en) Modulation/demodulation method, demodulation apparatus, and code modulation method
US20100149003A1 (en) Run length limiter and run length limiting method

Legal Events

Date Code Title Description
AS Assignment

Owner name: KABUSHIKI KAISHA TOSHIBA, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TAKEDA, SUSUMU;YAMADA, KENICHIRO;REEL/FRAME:067703/0319

Effective date: 20240521

Owner name: KABUSHIKI KAISHA TOSHIBA, JAPAN

Free format text: ASSIGNMENT OF ASSIGNOR'S INTEREST;ASSIGNORS:TAKEDA, SUSUMU;YAMADA, KENICHIRO;REEL/FRAME:067703/0319

Effective date: 20240521

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: ALLOWED -- NOTICE OF ALLOWANCE NOT YET MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS