Redundant memory access for rows or columns containing faulty memory cells in analog neural memory in deep learning artificial neural network
Numerous embodiments are disclosed for accessing redundant non-volatile memory cells in place of one or more rows or columns containing one or more faulty non-volatile memory cells during a program, erase, read, or neural read operation in an analog neural memory system used in a deep learning artificial neural network.
Latest SILICON STORAGE TECHNOLOGY, INC. Patents:
- Memory Device Formed On Silicon-On-Insulator Substrate, And Method Of Making Same
- Method of forming a semiconductor device with memory cells, high voltage devices and logic devices on a substrate using a dummy area
- Memory cell array with row direction gap between erase gate lines and dummy floating gates
- Precise data tuning method and apparatus for analog neural memory in an artificial neural network
- Output circuitry for non-volatile memory array in neural network
This application claims priority to U.S. Provisional Patent Application No. 62/696,778, filed on Jul. 11, 2018, and titled, “Redundant Memory Access for Rows or Columns Containing Faulty Memory Cells in Analog Neuro Memory in Deep Learning Artificial Neural Network,” which is incorporated by reference herein.
FIELD OF THE INVENTIONNumerous embodiments are disclosed for accessing redundant non-volatile memory cells in place of one or more rows or columns containing one or more faulty non-volatile memory cells during a program, erase, read, or neural read operation in an analog neural memory system used in a deep learning artificial neural network.
BACKGROUND OF THE INVENTIONArtificial neural networks mimic biological neural networks (e.g., the central nervous systems of animals, in particular the brain) which are used to estimate or approximate functions that can depend on a large number of inputs and are generally unknown. Artificial neural networks generally include layers of interconnected “neurons” which exchange messages between each other.
One of the major challenges in the development of artificial neural networks for high-performance information processing is a lack of adequate hardware technology. Indeed, practical neural networks rely on a very large number of synapses, enabling high connectivity between neurons, i.e. a very high computational parallelism. In principle, such complexity can be achieved with digital supercomputers or specialized graphics processing unit clusters. However, in addition to high cost, these approaches also suffer from mediocre energy efficiency as compared to biological networks, which consume much less energy primarily because they perform low-precision analog computation. CMOS analog circuits have been used for artificial neural networks, but most CMOS-implemented synapses have been too bulky given the high number of neurons and synapses.
Applicant previously disclosed an artificial (analog) neural network that utilizes one or more non-volatile memory arrays as the synapses in U.S. patent application Ser. No. 15/594,439, which is incorporated by reference. The non-volatile memory arrays operate as analog neuromorphic memory. The neural network device includes a first plurality of synapses configured to receive a first plurality of inputs and to generate therefrom a first plurality of outputs, and a first plurality of neurons configured to receive the first plurality of outputs. The first plurality of synapses includes a plurality of memory cells, wherein each of the memory cells includes spaced apart source and drain regions formed in a semiconductor substrate with a channel region extending there between, a floating gate disposed over and insulated from a first portion of the channel region and a non-floating gate disposed over and insulated from a second portion of the channel region. Each of the plurality of memory cells is configured to store a weight value corresponding to a number of electrons on the floating gate. The plurality of memory cells is configured to multiply the first plurality of inputs by the stored weight values to generate the first plurality of outputs.
Each non-volatile memory cells used in the analog neuromorphic memory system must be erased and programmed to hold a very specific and precise amount of charge in the floating gate. For example, each floating gate must hold one of N different values, where N is the number of different weights that can be indicated by each cell. Examples of N include 16, 32, and 64.
One unique characteristic of analog neuromorphic memory systems is that the system must support two different types of read operations. In a normal read operation, an individual memory cell is read as in conventional memory systems. However, in a neural read operation, the entire array of memory cells is read at one time, where each bit line will output a current that is the sum of all currents from the memory cells connected to that bit line.
Supporting both types of read operations leads to several challenges. One challenge is how to provide redundancy for the system. Where multiple redundant rows or columns are being used (due to the occurrence of multiple faulty rows or columns), the system must be able to activate all of the redundant rows or columns at one time during a neural read operation. However, in conventional systems, a read or program operation will operate on only one row or a sector of rows at any given time—and not all of them—and therefore only some of the redundant rows or columns will need to be asserted at any given time. Thus, prior art decoding systems do not support a neural read operation.
What is needed is an improved decoding system to be used with analog neuromorphic memory to provide redundancy during programming, erase, read, and neural read operations.
SUMMARY OF THE INVENTIONNumerous embodiments are disclosed for accessing redundant non-volatile memory cells in place of one or more rows or columns containing one or more faulty non-volatile memory cells during a program, erase, read, or neural read operation in an analog neural memory system used in a deep learning artificial neural network.
The artificial neural networks of the present invention utilize a combination of CMOS technology and non-volatile memory arrays.
Non-Volatile Memory Cells
Digital non-volatile memories are well known. For example, U.S. Pat. No. 5,029,130 (“the '130 patent”) discloses an array of split gate non-volatile memory cells, and is incorporated herein by reference for all purposes. Such a memory cell is shown in
Memory cell 210 is erased (where electrons are removed from the floating gate) by placing a high positive voltage on the word line terminal 22, which causes electrons on the floating gate 20 to tunnel through the intermediate insulation from the floating gate 20 to the word line terminal 22 via Fowler-Nordheim tunneling.
Memory cell 210 is programmed (where electrons are placed on the floating gate) by placing a positive voltage on the word line terminal 22, and a positive voltage on the source 16. Electron current will flow from the source 16 towards the drain 14. The electrons will accelerate and become heated when they reach the gap between the word line terminal 22 and the floating gate 20. Some of the heated electrons will be injected through the gate oxide 26 onto the floating gate 20 due to the attractive electrostatic force from the floating gate 20.
Memory cell 210 is read by placing positive read voltages on the drain 14 and word line terminal 22 (which turns on the channel region under the word line terminal). If the floating gate 20 is positively charged (i.e. erased of electrons and positively coupled to the drain 16), then the portion of the channel region under the floating gate 20 is turned on as well, and current will flow across the channel region 18, which is sensed as the erased or “1” state. If the floating gate 20 is negatively charged (i.e. programmed with electrons), then the portion of the channel region under the floating gate 20 is mostly or entirely turned off, and current will not flow (or there will be little flow) across the channel region 18, which is sensed as the programmed or “0” state.
Table No. 1 depicts typical voltage ranges that can be applied to the terminals of memory cell 210 for performing read, erase, and program operations:
Other split gate memory cell configurations are known. For example,
Table No. 2 depicts typical voltage ranges that can be applied to the terminals of memory cell 310 for performing read, erase, and program operations:
Table No. 3 depicts typical voltage ranges that can be applied to the terminals of memory cell 410 for performing read, erase, and program operations:
Table No. 4 depicts typical voltage ranges that can be applied to the terminals of memory cell 510 for performing read, erase, and program operations:
In order to utilize the memory arrays comprising one of the types of non-volatile memory cells described above in an artificial neural network, two modifications are made. First, the lines are configured so that each memory cell can be individually programmed, erased, and read without adversely affecting the memory state of other memory cells in the array, as further explained below. Second, continuous (analog) programming of the memory cells is provided.
Specifically, the memory state (i.e. charge on the floating gate) of each memory cells in the array can be continuously changed from a fully erased state to a fully programmed state, independently and with minimal disturbance of other memory cells. In another embodiment, the memory state (i.e., charge on the floating gate) of each memory cell in the array can be continuously changed from a fully programmed state to a fully erased state, and vice-versa, independently and with minimal disturbance of other memory cells. This means the cell storage is analog or at the very least can store one of many discrete values (such as 16 or 64 different values), which allows for very precise and individual tuning of all the cells in the memory array, and which makes the memory array ideal for storing and making fine tuning adjustments to the synapsis weights of the neural network.
Neural Networks Employing Non-Volatile Memory Cell Arrays
S0 is the input, which for this example is a 32×32 pixel RGB image with 5 bit precision (i.e. three 32×32 pixel arrays, one for each color R, G and B, each pixel being 5 bit precision). The synapses CB1 going from S0 to C1 have both different sets of weights and shared weights, and scan the input image with 3×3 pixel overlapping filters (kernel), shifting the filter by 1 pixel (or more than 1 pixel as dictated by the model). Specifically, values for 9 pixels in a 3×3 portion of the image (i.e., referred to as a filter or kernel) are provided to the synapses CB1, whereby these 9 input values are multiplied by the appropriate weights and, after summing the outputs of that multiplication, a single output value is determined and provided by a first neuron of CB1 for generating a pixel of one of the layers of feature map C1. The 3×3 filter is then shifted one pixel to the right (i.e., adding the column of three pixels on the right, and dropping the column of three pixels on the left), whereby the 9 pixel values in this newly positioned filter are provided to the synapses CB1, whereby they are multiplied by the same weights and a second single output value is determined by the associated neuron. This process is continued until the 3×3 filter scans across the entire 32×32 pixel image, for all three colors and for all bits (precision values). The process is then repeated using different sets of weights to generate a different feature map of C1, until all the features maps of layer C1 have been calculated.
At C1, in the present example, there are 16 feature maps, with 30×30 pixels each. Each pixel is a new feature pixel extracted from multiplying the inputs and kernel, and therefore each feature map is a two dimensional array, and thus in this example the synapses CB1 constitutes 16 layers of two dimensional arrays (keeping in mind that the neuron layers and arrays referenced herein are logical relationships, not necessarily physical relationships—i.e., the arrays are not necessarily oriented in physical two dimensional arrays). Each of the 16 feature maps is generated by one of sixteen different sets of synapse weights applied to the filter scans. The C1 feature maps could all be directed to different aspects of the same image feature, such as boundary identification. For example, the first map (generated using a first weight set, shared for all scans used to generate this first map) could identify circular edges, the second map (generated using a second weight set different from the first weight set) could identify rectangular edges, or the aspect ratio of certain features, and so on.
An activation function P1 (pooling) is applied before going from C1 to S1, which pools values from consecutive, non-overlapping 2×2 regions in each feature map. The purpose of the pooling stage is to average out the nearby location (or a max function can also be used), to reduce the dependence of the edge location for example and to reduce the data size before going to the next stage. At S1, there are 16 15×15 feature maps (i.e., sixteen different arrays of 15×15 pixels each). The synapses and associated neurons in CB2 going from S1 to C2 scan maps in S1 with 4×4 filters, with a filter shift of 1 pixel. At C2, there are 22 12×12 feature maps. An activation function P2 (pooling) is applied before going from C2 to S2, which pools values from consecutive non-overlapping 2×2 regions in each feature map. At S2, there are 22 6×6 feature maps. An activation function is applied at the synapses CB3 going from S2 to C3, where every neuron in C3 connects to every map in S2. At C3, there are 64 neurons. The synapses CB4 going from C3 to the output S3 fully connects S3 to C3. The output at S3 includes 10 neurons, where the highest output neuron determines the class. This output could, for example, be indicative of an identification or classification of the contents of the original image.
Each level of synapses is implemented using an array, or a portion of an array, of non-volatile memory cells.
The output of the memory array is supplied to a differential summer (such as summing op-amp or summing current mirror) 38, which sums up the outputs of the memory cell array to create a single value for that convolution. The differential summer is such as to realize summation of positive weight and negative weight with positive input. The summed up output values are then supplied to the activation function circuit 39, which rectifies the output. The activation function may include sigmoid, tanh, or ReLU functions. The rectified output values become an element of a feature map as the next layer (C1 in the description above for example), and are then applied to the next synapse to produce next feature map layer or final layer. Therefore, in this example, the memory array constitutes a plurality of synapses (which receive their inputs from the prior layer of neurons or from an input layer such as an image database), and summing op-amp 38 and activation function circuit 39 constitute a plurality of neurons.
Vector-by-Matrix Multiplication (VMM) Arrays
As described herein for neural networks, the flash cells are preferably configured to operate in sub-threshold region.
The memory cells described herein are biased in weak inversion:
Ids=Io*e(Vg−Vth)/kVt=w*Io*e(Vg)/kVt
w=e(−Vth)/kVt
For an I-to-V log converter using a memory cell to convert input current into an input voltage:
Vg=k*Vt*log[Ids/wp*Io]
For a memory array used as a vector matrix multiplier VMM, the output current is:
Iout=wa*Io*e(Vg)/kVt, namely
Iout=(wa/wp)*Iin=W*Iin
W=e(Vthp−Vtha)/kVt
A wordline or control gate can be used as the input for the memory cell for the input voltage.
Alternatively, the flash memory cells can be configured to operate in the linear region:
Ids=beta*(Vgs−Vth)*Vds; beta=u*Cox*W/L
Wα(Vgs−Vth)
For an I-to-V linear converter, a memory cell operating in the linear region can be used to convert linearly an input/output current into an input/output voltage.
Other embodiments for the ESF vector matrix multiplier are as described in U.S. patent application Ser. No. 15/826,345, which is incorporated by reference herein. A sourceline or a bitline can be used as the neuron output (current summation output).
VMM 1400 implements uni-directional tuning for memory cells in memory array 1403. That is, each cell is erased and then partially programmed until the desired charge on the floating gate is reached. If too much charge is placed on the floating gate (such that the wrong value is stored in the cell), the cell must be erased and the sequence of partial programming operations must start over. As shown, two rows sharing the same erase gate need to be erased together (to be known as a page erase), and thereafter, each cell is partially programmed until the desired charge on the floating gate is reached,
The prior art includes a concept known as long short-term memory (LSTM). LSTM units often are used in neural networks. LSTM allows a neural network to remember information over arbitrary time intervals and to use that information in subsequent operations. A conventional LSTM unit comprises a cell, an input gate, an output gate, and a forget gate. The three gates regulate the flow of information into and out of the cell. VMMs are particularly useful in LSTM units.
LSTM cell 2900 comprises sigmoid function devices 2901, 2902, and 2903, each of which applies a number between 0 and 1 to control how much of each component in the input vector is allowed through to the output vector. LSTM cell 2900 also comprises tank devices 2904 and 2905 to apply a hyperbolic tangent function to an input vector, multiplier devices 2906, 2907, and 2908 to multiply two vectors together, and addition device 2909 to add two vectors together.
It can be further appreciated that LSTM systems will typically comprise multiple VMM arrays, each of which requires functionality provided by certain circuit blocks outside of the VMM arrays, such as a summer and activation circuit block and high voltage generation blocks. Providing separate circuit blocks for each VMM array would require a significant amount of space within the semiconductor device and would be somewhat inefficient. The embodiments described below therefore attempt to minimize the circuitry required outside of the VMM arrays themselves.
Similarly, an analog VMM implementation can be used for a GRU (gated recurrent unit) system. GRUs are a gating mechanism in recurrent neural networks. GRUs are similar to LSTMs, with one notable difference being that GRUs lack an output gate.
It can be further appreciated that GRU systems will typically comprise multiple VMM arrays, each of which requires functionality provided by certain circuit blocks outside of the VMM arrays, such as a summer and activation circuit block and high voltage generation blocks. Providing separate circuit blocks for each VMM array would require a significant amount of space within the semiconductor device and would be somewhat inefficient. The embodiments described below therefore attempt to minimize the circuitry equired outside of the VMM arrays themselves.
Low voltage row decoder 1803 provides a bias voltage for read and program operations and provides a decoding signal for high voltage row decoder 1805. High voltage row decoder 1805 provides a high voltage bias signal for program and erase operations. Bit line PE driver 1801 provides controlling function for bit line in program, verify, and erase. Bias circuit 1801 is a shared bias block that provides the multiple voltages needed for the various program, erase, program verify, and read operations.
VMM system 1800 further comprises redundancy array 1810 and/or redundancy array 1813. Redundancy arrays 1810 and 1813 each provide array redundancy for replacing a defective portion in array 1807, in accordance with the redundancy embodiments described in greater detail below.
VMM system 1800 further comprises NVR (non-volatile register, aka info sector) sectors 1811, which are array sectors used to store user info, device ID, password, security key, trimbits, configuration bits, manufacturing info, etc.
VMM system 1800 optionally comprises reference sector 1812 and/or reference system 1850. Reference system 1850 comprises reference array 1852, reference array low voltage row decoder 1851, reference array high voltage row decoder 1853, and reference array low voltage column decoder 1854. The reference system can be shared across multiple VMM systems.
Reference array low voltage row decoder 1851 provides a bias voltage for read and programming operations involving reference array 1852 and also provides a decoding signal for reference array high voltage row decoder 1853. Reference array high voltage row decoder 1853 provides a high voltage bias for program and operations involving reference array 1852. Reference array low voltage column decoder 1854 provides a decoding function for reference array 1852. Reference array 1852 is such as to provide reference target for program verify or cell margining (searching for marginal cells).
During a testing or configuration phase, each row of memory cells in array 1906 is tested and verified. Any memory cells deemed to be faulty are identified, and the address for each row that contains one or more faulty memory cells is stored in non-volatile memory (not shown). Thereafter, during operation of memory system 1900, each address 1901 for a read or write operation is compared by address comparator 1903 against each address in the set of stored addresses corresponding to rows containing one or more faulty memory cells.
If a match is found by address comparator 1903 with any of the stored addresses, enable signal 1904 is asserted, which signifies that the received address is for a faulty row. Enable signal 1904 is received by redundant array row decoder 1909, which then selects a row that has been assigned to the row containing the faulty memory cell. Thus, the program, erase, or read operation is directed to the redundant row instead of the row containing the faulty memory cell.
If a match is not found by address comparator 1903, then enable signal 1904 is de-asserted, and row decoder 1908 is enabled by the output of logic circuit 1905 (here shown as an inverter). In this situation, the received address 1901 is used to access a row in array 1906 for the operation.
By design, only one of the assigned redundant rows can be asserted at any given time. This prior art system therefore could not be used to perform a neural read operation, where all non-faulty rows and assigned redundant rows are asserted.
Row decoder and redundancy latch block 2101 further comprises redundancy sub-block 2113 coupled to a pair of word lines in redundancy array 2103 (here, word lines 2117 and 2118). Additional redundancy sub-blocks similar to redundancy sub-block 2113 can be included in row decoder and redundancy latch block 2101.
Sub-block 2105 comprises NAND gates 2106 and 2107, inverters 2108 and 2109, NAND gates 2110 and 2111, inverters 2112 and 2113, and latch 2114. Latch 2114 is programmed (or loaded with configuration data at power-up or in response to a redundancy load command) during a testing or configuration phase. If word lines 2115 or 2116 are coupled to a row containing one or more faulty memory cells, then a “1” will be programmed into latch 2114. Otherwise, latch 2114 will store a “0”. During normal operation, when NAND gates 2106 or 2107 receive an address corresponding to word lines 2115 or 2116, respectively, latch 2114 will cause that word line to be de-asserted instead of asserted. Thus, the word line containing the faulty memory cell will not be selected.
Redundancy sub-block 2113 contains similar components as sub-block 2105. Here, latch 2119 is programmed during a testing or configuration phase. If word lines 2117 or 2118 are to be used, then latch 2119 is programmed with a “0”. Otherwise, latch 2119 is programmed with a “1”. During normal operation, when the receiving NAND gate receives an address corresponding to word lines 2117 or 2118, latch 2119 will cause that word line to be asserted. Thus, the word line containing the redundant memory cells will be selected. Notably, multiple redundant rows can be selected at any given time (such as during a neural read operation) by configuring latches such as latch 2119 with a “0”.
Row decoder and redundancy latch block 2201 further comprises redundancy block 2213 coupled to a redundant word line in redundancy array 2203 (here, word line 2212).
Sub-block 2205 comprises NAND gate 2206, inverter 2207, latch 2208, and switches 2209 and 2210. Here, redundancy latch 2208 is programmed during a configuration stage of memory system 2200. If latch 2208 contains a “1”, then the corresponding row coupled to word line 2211 in array 2202 is not faulty. During normal operation, switch 2209 will be closed and switch 2210 will be open, and word line 2211 in array 2202 will be accessed when the appropriate address is received. If latch 2208 contains a “0”, then the corresponding row in array 2202 is faulty. During normal operation, switch 2209 will be open and switch 2109 will be closed, and word line 2211 in array 2202 will not be accessed when the appropriate address is received. Instead, redundant word line 2212 in array 2202 will be accessed through a vertical line 2215.
Latch 2305 is programmed during a testing or configuration stage of memory system 2300. If latch 2305 contains a “1”, then the corresponding column coupled to bit line 2308 is not faulty. During normal operation, switch 2306 will be closed and switch 2307 will be open, and bit line 2308 will be accessed when the appropriate address is received. If latch 2305 contains a “0”, then the corresponding column in array 2302 is faulty. During normal operation, switch 2306 will be open and switch 2307 will be closed, and bit line 2308 in array 2202 will not be accessed when the appropriate address is received. Instead, redundant bit line 2309 in redundancy array 2303 will be accessed through a horizontal line 2315.
If a match is found (step 2404), that means a bad address exists. The system will disable the bad address (step 2405) and enable a corresponding redundant address (step 2406). The program, program verify, or erase command is then executed using the redundant address (step 2407). The process is then complete (step 2408).
If a match is not found (step 2404), the received address is enabled (step 2409). The program, program verify, or erase command is then executed using the received address (step 2410). The process is then complete (step 2408).
It should be noted that, as used herein, the terms “over” and “on” both inclusively include “directly on” (no intermediate materials, elements or space disposed therebetween) and “indirectly on” (intermediate materials, elements or space disposed therebetween). Likewise, the term “adjacent” includes “directly adjacent” (no intermediate materials, elements or space disposed therebetween) and “indirectly adjacent” (intermediate materials, elements or space disposed there between), “mounted to” includes “directly mounted to” (no intermediate materials, elements or space disposed there between) and “indirectly mounted to” (intermediate materials, elements or spaced disposed there between), and “electrically coupled” includes “directly electrically coupled to” (no intermediate materials or elements there between that electrically connect the elements together) and “indirectly electrically coupled to” (intermediate materials or elements there between that electrically connect the elements together). For example, forming an element “over a substrate” can include forming the element directly on the substrate with no intermediate materials/elements therebetween, as well as forming the element indirectly on the substrate with one or more intermediate materials/elements there between.
Claims
1. A method of performing a neural read operation in a memory system comprising a memory array and a redundant memory array, the method comprising:
- loading data into one or more latches;
- disabling a plurality of rows of memory cells in the memory array in response to the one or more latches;
- enabling a plurality of rows of memory cells in the redundant memory array; and
- performing a concurrent read operation of all memory cells in non-disabled rows in the memory array and all memory cells in the plurality of enabled rows in the redundant memory array.
2. The method of claim 1, further comprising:
- receiving a current on each output line in the memory array and redundant memory array, wherein the current on each output line comprises current drawn during the concurrent read operation by each memory cell in a non-disabled row in the memory array coupled to the output line and each memory cell in the plurality of enabled rows in the redundant memory array coupled to the output line.
3. The method in claim 2, wherein the output line is a bitline.
4. The method in claim 2, wherein the output line is a source line.
5. The method of claim 1, wherein the disabling utilizes discrete logic.
6. The method of claim 1, wherein the disabling utilizes one or more switches.
7. The method of claim 1, wherein each of the one or more latches is coupled to word lines for a sector of memory cells in the memory array.
8. The method of claim 1, wherein each of the memory cells in the memory array and each of the memory cells in the redundant memory array is a split-gate flash memory cell.
9. The method of claim 1, further comprising performing an address comparison during a programming or erasing operation to determine if an address corresponds to faulty memory.
10. The method of claim 9, wherein the redundant memory array is enabled for a program or erase operation if an address comparison identifies a match.
11. A method of performing a neural read operation in a memory system comprising a memory array and a redundant memory array, the method comprising:
- loading data into one or more latches;
- disabling a plurality of columns of memory cells in the memory array in response to the one or more latches;
- enabling a plurality of columns of memory cells in the redundant memory array; and
- performing a concurrent read operation of all memory cells in non-disabled columns in the memory array and all memory cells in the plurality of enabled columns in the redundant memory array.
12. The method of claim 11, further comprising:
- receiving a current on each output line in the memory array and redundant memory array, wherein the current on each output line comprises current drawn during the read operation by each memory cell in a non-disabled column in the memory array coupled to the output line and each memory cell in an enabled column in the redundant memory array coupled to the output line.
13. The method of claim 12, wherein the output line is a bitline.
14. The method of claim 12, wherein the output line is a source line.
15. The method of claim 11, wherein the disabling is performed by one or more switches.
16. The method of claim 11, wherein each of the memory cells in the memory array and each of the memory cells in the redundant memory array is a split-gate flash memory cell.
17. The method of claim 11, further comprising performing an address comparison during a programming or erasing operation to determine if an address corresponds to faulty memory.
18. The method of claim 17, wherein the redundant memory array is enabled for a program or erase operation if an address comparison identifies a match.
19. A memory system comprising:
- a memory array comprising a plurality of sectors, each sector comprising a plurality of rows of memory cells;
- a redundant memory array comprising a plurality of redundant sectors, each redundant sector comprising a plurality of rows of memory cells;
- for each sector in the memory array, a control block comprising a latch, wherein the latch can be programmed to disable one or more rows in the sector in the memory array; and
- for each sector in the redundant memory array, a control block comprising a redundancy latch, wherein the redundancy latch can be programmed to enable one or more rows in a redundant sector in the redundant memory array;
- wherein, during a neural read operation, a concurrent read operation is performed of all memory cells in non-disabled rows in the memory array and all memory cells in the enabled rows in the redundant memory array.
20. The system of claim 19, further comprising:
- circuitry for receiving a current on each bit line in the memory array and redundant memory array, wherein the current on each bit line comprises current drawn during the read operation by each memory cell in a non-disabled row in the memory array coupled to the bit line and each memory cell in an enabled row in the redundant memory cell coupled to the bit line.
21. The system of claim 19, further comprising:
- for each sector in the memory array, discrete logic between the latch and the memory array.
22. The system of claim 19, further comprising:
- for each sector in the memory array, one or more switches between the latch and the memory array.
23. The system of claim 19, wherein each of the memory cells in the memory array and each of the memory cells in the redundant memory array is a split-gate flash memory cell.
24. The system of claim 19, wherein the memory array is a vector-by-matrix multiplication array in a long short term memory system.
25. The system of claim 19, wherein the memory array is a vector-by-matrix multiplication array in a gated recurrent unit system.
26. A memory system comprising:
- a memory array comprising memory cells arranged in rows and columns, wherein each column of memory cells is coupled to a bit line;
- a redundant memory array comprising redundant memory cells arranged in rows and columns, wherein each column of redundant memory cells is coupled to a bit line;
- for each bit line in the memory array, a control block comprising a latch, wherein the latch can be programmed to disable the column of memory cells coupled to the bit line in the memory array; and
- for each bit line in the redundant memory array, a control block comprising a redundancy latch, wherein the redundancy latch can be programmed to enable a column of memory cells coupled to the bit line in the redundant memory array;
- wherein, during a neural read operation, a concurrent read operation is performed of all memory cells in non-disabled columns in the memory array and all memory cells in the enabled columns in the redundant memory array.
27. The system of claim 26, further comprising:
- circuitry for receiving a current on each bit line coupled to a non-disabled column in the memory array and each bit line coupled to an enabled column in the redundant memory array during a neural read operation.
28. The system of claim 26, further comprising:
- for each bit line in the memory array, one or more switches between the latch and the memory array.
29. The system of claim 26, wherein each of the memory cells in the memory array and each of the memory cells in the redundant memory array is a split-gate flash memory cell.
30. The system of claim 26, wherein the memory array is a vector-by-matrix multiplication array in a long short term memory system.
31. The system of claim 26, wherein the memory array is a vector-by-matrix multiplication array in a gated recurrent unit system.
5729152 | March 17, 1998 | Leung |
6747310 | June 8, 2004 | Fan |
7643354 | January 5, 2010 | Landfield |
7911872 | March 22, 2011 | Lakhani |
8971124 | March 3, 2015 | Manning |
20060198209 | September 7, 2006 | Tran |
20080137446 | June 12, 2008 | Kohara |
20110176377 | July 21, 2011 | Koyama |
20170337466 | November 23, 2017 | Bayat et al. |
201742075 | December 2017 | TW |
2017/200888 | November 2017 | WO |
Type: Grant
Filed: Oct 3, 2018
Date of Patent: Jan 31, 2023
Patent Publication Number: 20200019849
Assignee: SILICON STORAGE TECHNOLOGY, INC. (San Jose, CA)
Inventors: Hieu Van Tran (San Jose, CA), Stanley Hong (San Jose, CA), Thuan Vu (San Jose, CA), Anh Ly (San Jose, CA), Hien Pham (Ho Chi Minh), Kha Nguyen (Ho Chi Minh), Han Tran (Ho Chi Minh)
Primary Examiner: Justin R Knapp
Application Number: 16/151,259
International Classification: G06N 3/06 (20060101); G06F 3/06 (20060101); G06F 11/16 (20060101); G06N 3/063 (20060101);