Wave pipeline including synchronous stage
11544208 ยท 2023-01-03
Assignee
Inventors
Cpc classification
G11C7/1039
PHYSICS
Y02D10/00
GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
G11C7/22
PHYSICS
G11C16/0483
PHYSICS
International classification
G06F1/08
PHYSICS
Abstract
A wave pipeline includes a data path and a clock path. The data path includes a plurality of wave pipeline data stages and a synchronous data stage. The synchronous data stage includes a first data latch to latch the data from the synchronous data stage. The synchronous data stage is between a first wave pipeline data stage of the plurality of wave pipeline data stages and a second wave pipeline data stage of the plurality of wave pipeline data stages. The clock path corresponds to the plurality of wave pipeline data stages. The first data latch latches the data from the synchronous data stage in response to a clock signal on the clock path.
Claims
1. A wave pipeline comprising: a data path comprising a plurality of wave pipeline data stages and a synchronous data stage, the synchronous data stage comprising a first data latch to latch the data from the synchronous data stage, and the synchronous data stage between a first wave pipeline data stage of the plurality of wave pipeline data stages and a second wave pipeline data stage of the plurality of wave pipeline data stages; and a clock path corresponding to the plurality of wave pipeline data stages, wherein the first data latch latches the data from the synchronous data stage in response to a clock signal on the clock path.
2. The wave pipeline of claim 1, wherein the clock path comprises a plurality of clock stages corresponding to the plurality of wave pipeline data stages, each clock stage of the plurality of clock stages having a delay configured to be equal to a delay of the corresponding wave pipeline data stage of the plurality of wave pipeline data stages.
3. The wave pipeline of claim 2, further comprising: a second data latch, wherein the data path comprises a data input node and a data output node, wherein the clock path comprises an input clock node and a return clock node, and wherein the second data latch is to latch the data on the data output node in response to a return clock signal on the return clock node.
4. The wave pipeline of claim 3, wherein the second data latch comprises a FIFO.
5. The wave pipeline of claim 1, wherein a delay of each wave pipeline data stage of the plurality of wave pipeline data stages is less than one cycle of the clock signal.
6. The wave pipeline of claim 1, wherein a delay of the synchronous data stage is less than one cycle of the clock signal.
7. The wave pipeline of claim 1, wherein a delay of the synchronous data stage is greater than one cycle of the clock signal, and wherein the clock path further comprises a delay stage such that the first data latch latches the data from the synchronous data stage in response to a delayed clock signal from the delay stage.
8. The wave pipeline of claim 7, wherein a delay of the delay stage is less than a delay of the synchronous data stage.
9. The wave pipeline of claim 1, wherein the data path comprises a plurality of synchronous data stages, each synchronous data stage of the plurality of synchronous data stages comprising a first data latch to latch the data from the corresponding synchronous data stage of the plurality of synchronous data stages, and wherein each first data latch of the plurality of synchronous data stages latches the data from the corresponding synchronous data stage in response to the clock signal on the clock path.
10. A memory comprising: a memory array; a synchronous data stage to output data from the memory array in response to an address signal, the synchronous data stage comprising a first data latch to latch the output data; an address path coupled to an input of the synchronous data stage, the address path comprising a plurality of wave pipeline address stages; a data path coupled to an output of the synchronous data stage, the data path comprising a plurality of wave pipeline data stages; an input clock path corresponding to the plurality of wave pipeline address stages; and a return clock path coupled to the input clock path at the first data latch, the return clock path corresponding to the plurality of wave pipeline data stages, wherein the first data latch latches the output data in response to a clock signal on the input clock path.
11. The memory of claim 10, wherein the input clock path comprises a plurality of input clock stages corresponding to the plurality of wave pipeline address stages, each input clock stage of the plurality of input clock stages comprising a delay configured to be equal to a delay of the corresponding wave pipeline address stage of the plurality of wave pipeline address stages.
12. The memory of claim 10, wherein the return clock path comprises a plurality of return clock stages corresponding to the plurality of wave pipeline data stages, a delay of each return clock stage of the plurality of return clock stages equal to a delay of the corresponding wave pipeline data stage of the plurality of wave pipeline data stages.
13. The memory of claim 10, further comprising: a latch to latch the data from the data path in response to a return clock signal from the return clock path.
14. The memory of claim 13, wherein the latch comprises a FIFO.
15. The memory of claim 10, wherein a delay of each wave pipeline address stage of the plurality of wave pipeline address stages is less than one cycle of the clock signal, and wherein a delay of each wave pipeline data stage of the plurality of wave pipeline data stages is less than one cycle of the clock signal.
16. The memory of claim 10, wherein a delay of the synchronous data stage is less than one cycle of the clock signal.
17. The memory of claim 10, wherein the memory array comprises a NAND memory array.
18. A method for processing data through a wave pipeline, the method comprising: asynchronously processing data through a first wave pipeline data stage; aligning a clock signal with the data from the first wave pipeline data stage; processing the data from the first wave pipeline data stage through a synchronous data stage; latching the data from the synchronous data stage in response to the clock signal; processing the latched data through a second wave pipeline data stage; aligning the clock signal with the data from the second wave pipeline data stage; and latching the data from the second wave pipeline data stage in response to the clock signal.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
DETAILED DESCRIPTION
(9) In the following detailed description, reference is made to the accompanying drawings that form a part hereof, and in which is shown, by way of illustration, specific embodiments. In the drawings, like reference numerals describe substantially similar components throughout the several views. Other embodiments may be utilized and structural, logical and electrical changes may be made without departing from the scope of the present disclosure. The following detailed description is, therefore, not to be taken in a limiting sense.
(10) Disclosed herein are apparatus and methods for sending or receiving data between different parts of an integrated circuit, such as a memory device, using a wave pipeline including at least one synchronous stage. By using a synchronous stage or multiple synchronous stages in the wave pipeline, the time it takes for the data signal and the clock signal to reach the output of the wave pipeline may be less than if no synchronous stages are used. By reducing the time it takes for the data signal and the clock signal to reach the output of the wave pipeline, the number of latch (e.g., FIFO) stages needed to latch the data at the output may be reduced. In addition, a delay circuit to match the delay of each synchronous stage is not needed, thereby reducing power use.
(11)
(12) Memory device 100 includes a wave pipeline including a synchronous stage or multiple synchronous stages for reading data out of memory device 100. A clock signal path 126 may be routed along with a data bus 128. A return clock signal path 127 also may be routed along with the data bus 128. A clock signal on the clock signal path 126 may be used to trigger data out of the sensing devices 106 (e.g., sense amplifiers). A return clock signal on the return clock signal path 127 may be used to latch the data from the sensing devices 106 into a data latch (e.g., FIFO) of input/output (I/O) control circuitry 112 just prior to outputting the data to processor 130. By routing the clock signal and the return clock signal along with the data, they may be subjected to the same logic circuitry and process, voltage, and temperature (PVT) variations as the data, and the setup and hold time margin at the data latch may be improved. It will be recognized that process variations typically experienced in fabrication will generally lead to variations in performance of circuits, even where those circuits are intended to be of the same design or otherwise provide the same functionality. Similarly, even small separations of circuits may expose those circuits to differing voltage and temperature values if measured to sufficient precision. Thus, while this disclosure seeks to mitigate the effects of such variations between clock signal paths and data paths, there is no expectation that such variations are necessarily eliminated.
(13) Memory device 100 includes an array of memory cells 104 logically arranged in rows and columns. Memory cells of a logical row are typically coupled to the same access line (commonly referred to as a word line) while memory cells of a logical column are typically selectively coupled to the same data line (commonly referred to as a bit line). A single access line may be associated with more than one logical row of memory cells and a single data line may be associated with more than one logical column. Memory cells (not shown in
(14) A row decode circuitry 108 and a column decode circuitry 110 are provided to decode address signals. Address signals are received and decoded to access the array of memory cells 104. Memory device 100 also includes I/O control circuitry 112 to manage input of commands, addresses and data to the memory device 100 as well as output of data and status information from the memory device 100. An address register 114 is in communication with I/O control circuitry 112 and row decode circuitry 108 and column decode circuitry 110 to latch the address signals prior to decoding. A command register 124 is in communication with I/O control circuitry 112 and control logic 116 to latch incoming commands.
(15) An internal controller (e.g., control logic 116) controls access to the array of memory cells 104 in response to the commands and generates status information for the external processor 130, i.e., control logic 116 is configured to perform access operations in accordance with embodiments described herein. The control logic 116 is in communication with row decode circuitry 108 and column decode circuitry 110 to control the row decode circuitry 108 and column decode circuitry 110 in response to the addresses.
(16) Control logic 116 is also in communication with a cache register 118. Cache register 118 latches data, either incoming or outgoing, as directed by control logic 116 to temporarily store data while the array of memory cells 104 is busy writing or reading, respectively, other data. During a program operation (e.g., write operation), data is passed from sensing devices 106 to the cache register 118. The data is then passed from the cache register 118 to data register 120 for transfer to the array of memory cells 104; then new data is latched in the cache register 118 from sensing devices 106, which receive the new data from the I/O control circuitry 112. During a read operation, data is passed from the cache register 118 to sensing devices 106, which pass the data to the I/O control circuitry 112 for output to the external processor 130; then new data is passed from the data register 120 to the cache register 118. A status register 122 is in communication with I/O control circuitry 112 and control logic 116 to latch the status information for output to the processor 130.
(17) Memory device 100 receives control signals at control logic 116 from processor 130 over a control link 132. The control signals may include at least a chip enable CE#, a command latch enable CLE, an address latch enable ALE, a write enable WE#, and a read enable RE#. Additional control signals (not shown) may be further received over control link 132 depending upon the nature of the memory device 100. Memory device 100 receives command signals (which represent commands), address signals (which represent addresses), and data signals (which represent data) from processor 130 over a multiplexed input/output (I/O) bus 134 and outputs data to processor 130 over I/O bus 134.
(18) For example, the commands are received over input/output (I/O) pins [7:0] of I/O bus 134 at I/O control circuitry 112 and are written into command register 124. The addresses are received over input/output (I/O) pins [7:0] of bus 134 at I/O control circuitry 112 and are written into address register 114. The data are received over input/output (I/O) pins [7:0] for an 8-bit device or input/output (I/O) pins [15:0] for a 16-bit device at I/O control circuitry 112 and are written into cache register 118 through sensing devices 106. The data are subsequently written into data register 120 for programming the array of memory cells 104. For another embodiment, cache register 118 may be omitted, and the data are written directly into data register 120 through sensing devices 106. Data are also output over input/output (I/O) pins [7:0] for an 8-bit device or input/output (I/O) pins [15:0] for a 16-bit device.
(19) It will be appreciated by those skilled in the art that additional circuitry and signals can be provided, and that the memory device of
(20) Additionally, while specific I/O pins are described in accordance with popular conventions for receipt and output of the various signals, it is noted that other combinations or numbers of I/O pins may be used in the various embodiments.
(21)
(22) Memory array 200A might be arranged in rows (each corresponding to a word line 202) and columns (each corresponding to a bit line 204). Each column may include a string of series-coupled memory cells, such as one of NAND strings 206.sub.0 to 206.sub.M. Each NAND string 206 might be coupled to a common source 216 and might include memory cells 208.sub.0 to 208.sub.N. The memory cells 208 represent non-volatile memory cells for storage of data. The memory cells 208 of each NAND string 206 might be connected in series between a select transistor 210 (e.g., a field-effect transistor), such as one of the select transistors 210.sub.0 to 210.sub.M (e.g., that may be source select transistors, commonly referred to as select gate source), and a select transistor 212 (e.g., a field-effect transistor), such as one of the select transistors 212.sub.0 to 212.sub.M (e.g., that may be drain select transistors, commonly referred to as select gate drain). Select transistors 210.sub.0 to 210.sub.M might be commonly coupled to a select line 214, such as a source select line, and select transistors 212.sub.0 to 212.sub.M might be commonly coupled to a select line 215, such as a drain select line.
(23) A source of each select transistor 210 might be connected to common source 216. The drain of each select transistor 210 might be connected to the source of a memory cell 208.sub.0 of the corresponding NAND string 206. For example, the drain of select transistor 210.sub.0 might be connected to the source of memory cell 208.sub.0 of the corresponding NAND string 206.sub.0. Therefore, each select transistor 210 might be configured to selectively couple a corresponding NAND string 206 to common source 216. A control gate of each select transistor 210 might be connected to select line 214.
(24) The drain of each select transistor 212 might be connected to the bit line 204 for the corresponding NAND string 206. For example, the drain of select transistor 212.sub.0 might be connected to the bit line 204.sub.0 for the corresponding NAND string 206.sub.0. The source of each select transistor 212 might be connected to the drain of a memory cell 208.sub.N of the corresponding NAND string 206. For example, the source of select transistor 212.sub.0 might be connected to the drain of memory cell 208.sub.N of the corresponding NAND string 206.sub.0. Therefore, each select transistor 212 might be configured to selectively couple a corresponding NAND string 206 to a corresponding bit line 204. A control gate of each select transistor 212 might be connected to select line 215.
(25) The memory array in
(26) Typical construction of memory cells 208 includes a data-storage structure 234 (e.g., a floating gate, charge trap, etc.) that can determine a data value of the cell (e.g., through changes in threshold voltage), and a control gate 236, as shown in
(27) A column of the memory cells 208 is a NAND string 206 or a plurality of NAND strings 206 coupled to a given bit line 204. A row of the memory cells 208 are memory cells 208 commonly coupled to a given word line 202. A row of memory cells 208 can, but need not include all memory cells 208 commonly coupled to a given word line 202. Rows of memory cells 208 may often be divided into one or more groups of physical pages of memory cells 208, and physical pages of memory cells 208 often include every other memory cell 208 commonly coupled to a given word line 202. For example, memory cells 208 commonly coupled to word line 202.sub.N and selectively coupled to even bit lines 204 (e.g., bit lines 204.sub.0, 204.sub.2, 204.sub.4, etc.) may be one physical page of memory cells 208 (e.g., even memory cells) while memory cells 208 commonly coupled to word line 202.sub.N and selectively coupled to odd bit lines 204 (e.g., bit lines 204.sub.1, 204.sub.3, 204.sub.5, etc.) may be another physical page of memory cells 208 (e.g., odd memory cells). Although bit lines 204.sub.3, 204.sub.5 are not expressly depicted in
(28)
(29) Although the examples of
(30)
(31) Wave pipeline 300 also includes a clock path including a plurality of clock stages 314.sub.0 to 314.sub.4 corresponding to the plurality of wave pipeline data stages 302.sub.0 to 302.sub.4 between an input clock node 316 and a return clock node 318. In this example, the input of clock stage 314.sub.0 is coupled to the clock input node 316. The output of clock stage 314.sub.0 is coupled to the input of clock stage 314.sub.1, and the output of clock stage 314.sub.1 is coupled to the clock input of synchronous data stage 304 through a clock input (CLK_IN) node 303. The clock output of synchronous data stage 304 is coupled to the input of clock stage 314.sub.2 through a clock output (CLK_OUT) node 305. The output of clock stage 314.sub.2 is coupled to the input of clock stage 314.sub.3. The output of clock stage 314.sub.3 is coupled to the input of clock stage 314.sub.4, and the output of clock stage 314.sub.4 is coupled to the return clock node 318. Each clock stage 314.sub.0 to 314.sub.4 has a delay configured to be equal to a delay of the corresponding wave pipeline data stage 302.sub.0 to 302.sub.4, respectively, such that the clock signal and the data move together (e.g., are aligned). In one example, a delay of each wave pipeline data stage 302.sub.0 to 302.sub.4 is less than one cycle of the clock signal.
(32) While wave pipeline 300 includes five wave pipeline data stages and a corresponding five clock stages, in other embodiments wave pipeline 300 may include less than five wave pipeline data stages and corresponding clock stages or more than five wave pipeline data stages and corresponding clock stages.
(33) The synchronous data stage 304 includes a data processing portion 310 and a first data latch 312 to latch the data from the synchronous data stage 304 (e.g., from data processing portion 310). Data is input to synchronous data stage 304 (e.g., to data processing portion 310) through the data input node 309. Data is output from the data processing portion 310 and input to the first data latch 312 through a data output (DATA_OUT) node 311. A clock signal is input to synchronous data stage 304 (e.g., to first data latch 312) through the clock input node 303. The first data latch 312 latches the data on the data output node 311 in response to the clock signal. First data latch 312 outputs the latched data to latched data output node 313. The clock signal on the clock input node 303 becomes the clock signal on the clock output node 305 without a delay. The time to process the data through data processing portion 310 between data input node 309 and data output node 311 is indicated by a delay TD1. In this example, TD1 is less than one cycle of the clock signal.
(34) The synchronous data stage 304 is between a first wave pipeline data stage (e.g., wave pipeline data stage 302.sub.1 in this example) and a second wave pipeline data stage (e.g., wave pipeline data stage 302.sub.2 in this example) of the plurality of wave pipeline data stages 302.sub.0 to 302.sub.4. While synchronous data stage 304 is illustrated as being arranged between wave pipeline data stage 302.sub.1 and wave pipeline data stage 302.sub.2, in other embodiments synchronous data stage 304 may be arranged between other wave pipeline data stages of the plurality of wave pipeline data stages 302.sub.0 to 302.sub.4 of wave pipeline 300.
(35) Wave pipeline 300 also includes a second data latch (e.g., FIFO) 320 to latch the data on the data output node 308 in response to a return clock signal on the return clock node 318. The data stored in FIFO 320 is output to an output data node 322 in response to an output clock signal on an output clock signal node 324. In one example, FIFO 320 includes a plurality of stages equal to the number of clock stages 314.sub.0 to 314.sub.4 (e.g., five in this example). It is noted that FIFO 320 does not include a stage for synchronous data stage 304 since the clock signal is not delayed by synchronous data stage 304. Therefore, by using synchronous data stage 304 in wave pipeline 300 in place of another wave pipeline data stage, FIFO 320 may be smaller compared to a FIFO in a wave pipeline that does not include synchronous data stage 304. In one example, FIFO 320 may be part of I/O control circuitry 112 of
(36) The time for the data on the data input node 306 to be processed through wave pipeline data stages 302.sub.0 to 302.sub.4 and synchronous data stage 304 and reach the data output node 308 (and for the clock signal on the input clock node 316 to be delayed through clock stages 314.sub.0 to 314.sub.4 and reach the return clock node 318) is indicated by a latency (e.g., address access time (TAA)) 326. By using synchronous data stage 304 in wave pipeline 300 in place of another wave pipeline data stage, the latency 326 may be reduced compared to a wave pipeline not including synchronous data stage 304. In addition, since synchronous data stage 304 does not include a clock stage to delay the clock signal, wave pipeline 300 may use less power than a wave pipeline not including synchronous data stage 304.
(37)
(38)
(39) Wave pipeline 350 also includes a clock path including a plurality of clock stages 314.sub.0 to 314.sub.3 corresponding to the plurality of wave pipeline data stages 302.sub.0 to 302.sub.3 between an input clock node 316 and a return clock node 318. In this example, the input of clock stage 314.sub.0 is coupled to the clock input node 316. The output of clock stage 314.sub.0 is coupled to the input of clock stage 314.sub.1, and the output of clock stage 314.sub.1 is coupled to the clock input of synchronous data stage 304.sub.0 through a clock input node 303.sub.0. The clock output of synchronous data stage 304.sub.0 is coupled to the input of clock stage 314.sub.2 through a clock output node 305.sub.0. The output of clock stage 314.sub.2 is coupled to the clock input of synchronous data stage 304.sub.1 through a clock input node 303.sub.1. The clock output of synchronous data stage 304.sub.1 is coupled to the input of clock stage 314.sub.3 through a clock output node 305.sub.1. The output of clock stage 314.sub.3 is coupled to the return clock node 318.
(40) Each synchronous data stage 304.sub.0 to 304.sub.1 includes a data processing portion 310.sub.0 to 310.sub.1 and a first data latch 312.sub.0 to 312.sub.1 to latch the data from the synchronous data stage 304.sub.0 to 304.sub.1 (e.g., from data processing portion 310.sub.0 to 310.sub.1), respectively. Data is input to each synchronous data stage 304.sub.0 to 304.sub.1 (e.g., to data processing portion 310.sub.0 to 310.sub.1) through the data input node 309.sub.0 to 309.sub.1, respectively. Data is output from each data processing portion 310.sub.0 to 310.sub.1 and input to the first data latch 312.sub.0 to 312.sub.1 through a data output node 311.sub.0 to 311.sub.1, respectively. A clock signal is input to each synchronous data stage 304.sub.0 to 304.sub.1 (e.g., to first data latch 312.sub.0 to 312.sub.1) through the clock input node 303.sub.0 to 303.sub.1, respectively. Each first data latch 312.sub.0 to 312.sub.1 latches the data on the data output node 311.sub.0 to 311.sub.1, respectively, in response to the clock signal. Each first data latch 312.sub.0 to 312.sub.1 outputs the latched data to latched data output node 313.sub.0 to 313.sub.1, respectively. The clock signal on each clock input node 303.sub.0 to 303.sub.1 becomes the clock signal on the clock output node 305.sub.0 to 305.sub.1, respectively, without a delay. The time to process the data through each data processing portion 310.sub.0 to 310.sub.1 between data input node 309.sub.0 to 309.sub.1 and data output node 311.sub.0 to 311.sub.1 is indicated by delays TD1 and TD2, respectively. In this example, TD1 and TD2 are each less than one cycle of the clock signal.
(41) Wave pipeline 350 also includes a second data latch (e.g., FIFO) 320 to latch the data on the data output node 308 in response to a return clock signal on the return clock node 318. The data stored in FIFO 320 is output to an output data node 322 in response to an output clock signal on an output clock signal node 324. In one example, FIFO 320 includes a plurality of stages equal to the number of clock stages 314.sub.0 to 314.sub.3 (e.g., four in this example). It is noted that FIFO 320 does not include a stage for each synchronous data stage 304.sub.0 to 304.sub.1 since the clock signal is not delayed by synchronous data stages 304.sub.0 to 304.sub.1. Therefore, by using synchronous data stages 304.sub.0 to 304.sub.1 in wave pipeline 350, FIFO 320 may be smaller compared to a FIFO in a wave pipeline that does not include synchronous data stages 304.sub.0 to 304.sub.1 and compared to a wave pipeline including fewer synchronous data stages, such as wave pipeline 300 of
(42) The time for the data on the data input node 306 to be processed through wave pipeline data stages 302.sub.0 to 302.sub.3 and synchronous data stages 304.sub.0 to 304.sub.1 and reach the data output node 308 (and for the clock signal on the input clock node 316 to be delayed through clock stages 314.sub.0 to 314.sub.3 and reach the return clock node 318) is indicated by a latency (e.g., address access time (TAA)) 326. By using multiple synchronous data stages 304.sub.0 to 304.sub.1 in wave pipeline 350, the latency 326 may be reduced compared to a wave pipeline not including synchronous data stages 304.sub.0 to 304.sub.1 and compared to a wave pipeline including fewer synchronous data stages, such as wave pipeline 300 of
(43) In this example, by including multiple synchronous data stages 304.sub.0 to 304.sub.1 in wave pipeline 350, an additional clock cycle is used to latch the data in each first data latch 312.sub.0 to 312.sub.1 such that three clock cycles are used to process the data through wave pipeline 350. While wave pipeline 350 includes two synchronous data stages 304.sub.0 to 304.sub.1, in other embodiments wave pipeline 350 may include more than two synchronous data stages. In this case, the number of clock cycles used to process the data through the wave pipeline would be equal to the number of synchronous data stages plus one.
(44)
(45)
(46) Wave pipeline 400 also includes an input clock path including a plurality of input clock stages 414.sub.0 to 414.sub.1 between a clock input node 416 and an input node 403 of synchronous data stage 404. Each clock stage 414.sub.0 to 414.sub.1 corresponds to the wave pipeline address stages 402.sub.0 to 402.sub.1 and includes a delay configured to be equal to a delay of the corresponding wave pipeline address stage 402.sub.0 to 402.sub.1, respectively, such that the clock signal and the address move together (e.g., are aligned). In one example, a delay of each wave pipeline address stage 402.sub.0 to 402.sub.1 is less than one cycle of the clock signal. Wave pipeline 400 also includes a return clock path coupled to the input clock path at the first data latch 412. The return clock path includes a plurality of return clock stages 414.sub.2 to 414.sub.3 between an output node 405 of the synchronous data stage 404 and a return clock node 418. Each return clock stage 414.sub.2 to 414.sub.3 corresponds to the plurality of wave pipeline data stages 402.sub.2 to 402.sub.3 and includes a delay configured to be equal to a delay of the corresponding wave pipeline data stage 402.sub.2 to 402.sub.3, respectively, such that the clock signal and the data move together (e.g., are aligned). In one example, a delay of each wave pipeline data stage 402.sub.2 to 402.sub.3 is less than one cycle of the clock signal.
(47) In this example, the input of wave pipeline address stage 402.sub.0 is coupled to the address input node 406. The output of wave pipeline address stage 402.sub.0 is coupled to the input of wave pipeline address stage 402.sub.1, and the output of wave pipeline address stage 402.sub.1 is coupled to the address input of synchronous data stage 404 through the input node 409. The data output of synchronous data stage 404 is coupled to the input of wave pipeline data stage 402.sub.2 through the output node 413. The output of wave pipeline data stage 402.sub.2 is coupled to the input of wave pipeline data stage 402.sub.3. The output of wave pipeline data stage 402.sub.3 is coupled to the data output node 408.
(48) The input of input clock stage 414.sub.0 is coupled to the clock input node 416. The output of input clock stage 414.sub.0 is coupled to the input of input clock stage 414.sub.1, and the output of input clock stage 414.sub.1 is coupled to the clock input of synchronous data stage 404 through input node 403. The clock output of synchronous data stage 404 is coupled to the input of return clock stage 414.sub.2 through output node 405. The output of return clock stage 414.sub.2 is coupled to the input of return clock stage 414.sub.3. The output of return clock stage 414.sub.3 is coupled to the return clock node 418.
(49) The synchronous stage 404 includes a data processing portion 410 and the first data latch 412 to latch the data from the synchronous data stage 404 (e.g., from data processing portion 410). An address is input to synchronous data stage 404 (e.g., to data processing portion 410) through the input node 409. The data within the memory array corresponding to the address is retrieved from the memory array. The retrieved data is output from the data processing portion 410 and input to the first data latch 412 through a data output node 411. A clock signal is input to synchronous data stage 404 (e.g., to first data latch 412) through the input node 403. The first data latch 412 latches the data on the data output node 411 in response to the clock signal. First data latch 412 outputs the latched data to output node 413. The clock signal on the input node 403 becomes the clock signal on the output node 405 without a delay. In one example, the time to process the data through data processing portion 410 between input node 409 and data output node 411 (e.g., a delay of synchronous data stage 404) is less than one cycle of the clock signal.
(50) Wave pipeline 400 also includes a second data latch (e.g., FIFO) 420 to latch the data on the data output node 408 in response to a return clock signal on the return clock node 418. The data stored in FIFO 420 is output to an output data node 422 in response to an output clock signal on an output clock signal node 424. In one example, FIFO 420 includes a plurality of stages equal to the number of clock stages 414.sub.0 to 414.sub.3 (e.g., four in this example). It is noted that FIFO 420 does not include a stage for synchronous data stage 404 since the clock signal is not delayed by synchronous data stage 404. Therefore, by using synchronous data stage 404 in wave pipeline 400 in place of another wave pipeline data stage, FIFO 420 may be smaller compared to a FIFO in a wave pipeline that does not include synchronous data stage 404. In one example, FIFO 420 may be part of I/O control circuitry 112 of
(51)
(52) As illustrated in
CONCLUSION
(53) Although specific embodiments have been illustrated and described herein, it will be appreciated by those of ordinary skill in the art that any arrangement that is calculated to achieve the same purpose may be substituted for the specific embodiments shown. Many adaptations of the embodiments will be apparent to those of ordinary skill in the art. Accordingly, this application is intended to cover any adaptations or variations of the embodiments.