US9529745B2 - System on chip and method of operating a system on chip - Google Patents
System on chip and method of operating a system on chip Download PDFInfo
- Publication number
- US9529745B2 US9529745B2 US14/190,374 US201414190374A US9529745B2 US 9529745 B2 US9529745 B2 US 9529745B2 US 201414190374 A US201414190374 A US 201414190374A US 9529745 B2 US9529745 B2 US 9529745B2
- Authority
- US
- United States
- Prior art keywords
- memory
- buffer
- data
- unit
- buffers
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F13/00—Interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units
- G06F13/14—Handling requests for interconnection or transfer
- G06F13/16—Handling requests for interconnection or transfer for access to memory bus
- G06F13/1668—Details of memory controller
- G06F13/1673—Details of memory controller using buffers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F13/00—Interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units
- G06F13/14—Handling requests for interconnection or transfer
- G06F13/20—Handling requests for interconnection or transfer for access to input/output bus
- G06F13/32—Handling requests for interconnection or transfer for access to input/output bus using combination of interrupt and burst mode transfer
Definitions
- This invention relates to a system on chip and to a method of operating a system on chip.
- a system on chip is a data processing device located on a single semi-conductor chip and capable of executing instructions stored in a memory unit on the chip.
- the instructions may be soft coded or hard coded.
- a system on chip may comprise multiple interconnected functional units, for example, one or more processor cores, one or more memory units, or input/output units for enabling the system on chip (SoC) to interact with its environment, e.g., via one or more antennas or via wired connections.
- SoC system on chip
- a SoC may comprise a buffer, e.g., a First-In-First-Out (FIFO) buffer or, e.g., a stack buffer, connected between a data source and a processing unit.
- the buffer may absorb an incoming flux of data, i.e. a data stream, while the processing unit is occupied.
- a data source may send out a continuous stream of data which has to be received while the processing unit services the data source only when it is available, then emptying the data that has accumulated in the buffer.
- a data source may be any hardware or software component that is capable of providing, e.g., generating or forwarding, a data stream.
- a processing unit may be any hardware or software component for processing the received data.
- the data source may be an Ethernet port on the SoC.
- the processing unit may for example be a hardware accelerator arranged to carry out various processing actions on incoming packets from the data source.
- a buffer may for example be a FIFO buffer, a data stack, or any other memory structure capable of maintaining a write pointer for indicating a next available address for writing and a read pointer for indicating a next address to read from.
- Such buffers may for example be implemented as on-chip static-random-access-memory (SRAM) units.
- SRAM static-random-access-memory
- the size of a buffer may be chosen so as to be able to absorb enough data while the processing unit is busy with previously received data. This may notably be a factor of a peak data bandwidth, which may be significantly higher than an average data bandwidth.
- the present invention is motivated by the insight that providing a buffer with a fixed capacity separately for each of multiple data streams may result in a memory array that may often be largely empty. This may be undesirable as on-chip memories can be an expensive resource, increasing die size and power consumption.
- FIG. 1 schematically shows an example of a system on chip 10 in which a data source 12 sends a stream of data to a receiver 14 .
- the receiver 14 transfers the data to a buffer 18 to be read by a processing unit 20 .
- the buffer 18 may be located in a memory unit 16 .
- the memory unit 16 may for example be an SRAM unit.
- the memory unit 16 may comprise a set of memory regions. Each memory region may comprise, for example, one or more registers.
- the memory regions may have the same capacity, e.g., a certain number of kilobytes, or they may have different capacities.
- An application e.g., a process executed by the processing unit 20 , may employ a virtualization mechanism to share a single physical resource among two or more virtual machines, effectively creating several, i.e. two or more, virtual channels on a single physical channel.
- these virtual machines may be unaware of each other and each may operate independently of the other virtual machines while a hypervisor unit or a super user unit arbitrates between the virtual machines.
- a hypervisor unit or a super user unit arbitrates between the virtual machines.
- each virtual machine since each virtual machine “assumes” that it can use the full bandwidth of the physical channel, it may require a buffer having a size adapted to the full bandwidth of the physical channel, for each virtual machine.
- the total size of the set of buffers provided for the virtual machines is then proportional to the number of virtual machines.
- FIG. 2 schematically shows another example of a SoC 10 .
- the SoC 10 comprises four data sources 12 a to 12 d sharing a single physical line for feeding their data to a processing unit 20 .
- the processing unit 20 comprises processor cores 20 a to 20 d .
- the data sources 12 a to 12 d may provide data streams simultaneously.
- An arbitrator unit 22 of the SoC 10 may select one of the data streams and route it to the receiver 14 .
- the receiver 14 may further provide it to the memory unit 16 .
- the memory unit 16 may comprise, for each of the data sources 12 a to 12 d , a separate buffer having a fixed capacity.
- the data sources 12 a to 12 d may thus be assigned buffers 18 a , 18 b , 18 c , and 18 d , respectively. Accordingly, when the arbitrator unit 22 has selected, for example, data source 12 c as the active data source among the data sources 12 a to 12 d , the data stream from the selected data source, e.g., from data source 12 c , may be routed to the processing unit 20 through the buffer, e g., buffer 18 c , that belongs to the selected data source.
- the data rate of the data received by the receiver 14 may be the same in both Figures.
- the total buffer capacity is four times as large in FIG. 2 as in FIG. 1 , assuming that each of the buffers 18 a to 18 d has the same capacity as the buffer 18 in FIG. 1 .
- the memory unit 16 receives the same amount of data per time from the receiver 14 it may be expected that the buffers 18 a to 18 d in FIG. 2 may on average be largely empty.
- a device with an Ethernet port e.g., a one-GB-per-second-Ethernet port, supporting eight virtual channels and a priority-based flow control in accordance with the prior art may require providing a static buffer to each virtual channel wherein the buffer must be large enough to absorb the incoming data while the processor is busy with previously received data.
- an emergency buffer may be required to absorb the incoming data from the moment the processor has issued a pause frame to a transmitter on the other side until the flow of data stops. Assuming that the required buffer size to absorb the incoming data at the steady state is 256 KB, the buffer required to absorb peak traffic may be two to four times larger than that, i.e., at least 512 KB.
- the total buffer size required for each virtual channel may be 532 KB.
- Each of the virtual channels may require a dedicated buffer because it is not known beforehand how the data is distributed among the different virtual channels. Therefore a system designer must assume that at any given point in time any one of the virtual channels may have the full one GB per second bandwidth and therefore a memory array of eight times 532 KB, i.e., approximately 4 MB, must be provided.
- the present invention provides a system on chip and method as described in the accompanying claims.
- FIG. 1 schematically shows an example of a system on chip.
- FIG. 2 schematically shows an example of an embodiment of a system on chip.
- FIG. 3 schematically shows an example of an embodiment of a system on chip.
- FIG. 4 schematically shows an example of an embodiment of a memory control unit.
- FIG. 5 schematically shows an example of an embodiment of a memory allocation unit.
- FIG. 6 shows a flow chart of an example of an embodiment of a method of operating a system on chip.
- FIGS. 7, 8, and 9 schematically illustrate an example of allocation and re-allocation of memory to multiple buffers for buffering multiple data streams.
- FIG. 3 schematically shows an example of an embodiment of a system on chip 10 which comprises two or more data sources 12 A to 12 H, a memory unit 16 , a memory control unit 24 , and a processing unit 20 .
- Each of the data sources 12 A to 12 H may be capable of providing a respective data stream.
- the data sources 12 A to 12 H may be capable of providing their data streams concurrently.
- the memory control unit 24 is arranged to maintain, for each of the data streams, a respective buffer inside the memory unit 16 and to route the respective data stream to the processing unit 20 via the respective buffer.
- Each of the thus maintained buffers has buffer characteristics which are variable and which comprise at least the amount of free memory of the respective buffer.
- the memory control unit 24 may be arranged to allocate and de-allocate memory to and from each of these buffers in dependence of the buffer characteristics of the respective buffer, thereby allowing memory of the memory unit 16 to be reallocated among the buffers.
- the size, i.e. the capacity, of each buffer is thus flexible and can be adapted as needed.
- the buffer characteristics of each buffer may comprise, in addition to the amount of free memory of the buffer, a data accumulation rate of the respective buffer.
- the data accumulation rate may be defined as the amount of data accumulating in the respective buffer per time, i.e. the rate at which data in the buffer is accumulating.
- the time interval with respect to which the accumulated data is measured may for example be a sliding time window.
- the sliding time window may for example comprise a fixed number of clock cycles and may advance periodically, e.g., with each new clock cycle.
- the current sliding time window may overlap a certain number of previous sliding time windows. Use of a sliding time window allows updating the data accumulation rate of each buffer periodically, e.g., with each new clock cycle.
- the operation of allocating and de-allocating memory to and from each buffer may be independent of the buffer characteristics of the respective other one or more buffers.
- Such a scheme may be simpler to implement than, e.g., an alternative scheme in which the allocation and de-allocation of memory to and from a buffer involves taking into account the buffer characteristics of the other buffers.
- the memory control unit 24 comprises a memory allocation unit 26 and a control unit 28 .
- the memory allocation unit 26 may also be referred to as the Efficient Memory Allocator (EMA).
- EMA Efficient Memory Allocator
- the memory allocation unit 26 may have an interface to the control unit 28 and to a system bus 30 , for example.
- the memory allocation unit 26 may further have an interface for each of the data sources 12 A to 12 H.
- the memory allocation unit 26 does not have a dedicated interface per data source but is arranged to receive the data streams from the data sources 12 A to 12 H via, e.g., the system bus 30 .
- the memory allocation unit 26 may be arranged to identify the particular data source of a received data stream, for example, on the basis of a destination address of the respective data stream or on the basis of a source identification tag included in the data stream.
- the memory allocation unit 26 may be arranged to monitor the amount of data written to the memory unit 16 for each of the data sources 12 A to 12 H, e.g., using a sliding time window, in order to determine an average data rate. Similarly, the memory allocation unit 26 may also be arranged to monitor the amount of data read from each of the buffers assigned to the data sources 12 A to 12 H in the memory unit 16 . When the rate of data accumulation for a certain data source exceeds a threshold, e.g., a user configurable threshold, the memory allocation unit may allocate additional memory to that data source.
- the data accumulation rate may be defined as the difference between the rate at which data is written to the respective buffer and the rate at which data is read from the respective buffer.
- a negative data accumulation rate may be referred to as a data depletion rate.
- a negative data depletion rate may be referred to as a data accumulation rate.
- the memory allocation unit does not necessarily determine or evaluate the actual rate of write accesses to the respective buffer but may determine the data accumulation rate on the basis of a hypothetical constant data retrieval rate, that is, by assuming that the processing unit reads data from the memory unit 16 at a fixed rate.
- the memory allocation unit may be arranged to allocate additional memory to a given buffer in response to detecting that an average rate of write transactions to that buffer is above a certain threshold. This threshold may be user-configurable.
- the memory allocation unit may de-allocate that memory region from the buffer.
- the de-allocated memory region may thus be reused for another data stream.
- the memory allocation unit 26 is configurable by a user so as to ignore the rate of read accesses and to de-allocate a memory region in response to an average of write accesses dropping below a certain threshold, e.g., a user-configurable threshold, when the memory region is empty.
- a certain threshold e.g., a user-configurable threshold
- the boundaries of a memory region allocated to a buffer associated with a certain data source may be user-configurable.
- each of the data sources 12 A to 12 H may have a dynamic buffer assigned to it.
- Each of these buffers may be dynamic in the sense that its size, i.e. its data capacity, may be varied, in dependence of, e.g., a rate at which data is accumulating in the respective buffer.
- the memory provided by the memory unit 16 may thus be used more efficiently as compared to, for example, the schemes described above in reference to FIGS. 1 and 2 in which each data source, or, equivalently each data stream, has a static buffer assigned to it.
- the memory unit 16 may comprise one or more memory regions that may be allocated to different buffers, depending on the data accumulation rates of these buffers. Operation of the data sources may be unaffected by any variations in the buffer size of the respective data source. For instance, each of the data sources 12 A to 12 H may operate under the assumption that it has a buffer inside the memory unit 16 at a certain constant size. This constant size may for instance be the maximum possible size of the buffer, e.g., the total size of the memory provided by the memory unit 16 .
- the memory unit 16 runs out of free memory regions so that no more memory can be allocated to the buffers.
- An alarm mechanism for signaling such an event may be implemented in, e.g., the memory allocation unit 26 .
- the memory allocation unit 26 may be arranged to initiate an interrupt when the memory unit 16 has run out of free memory.
- the system on chip 10 may be arranged so as to allow such an interrupt to be handled by, e.g., user software in a user defined manner.
- the interrupt may for example include issuing a signal to one or more of the data sources 12 A to 12 H to stop sending data.
- the memory allocation unit 26 may for example comprise a Region Request Identification (RRI) unit 32 , a memory allocation and de-fragmentation (MAD) unit 34 , and a set of control and status registers 36 .
- the RRI unit 32 may be arranged to detect requests for transactions (i.e., transaction requests) on, e.g., the system bus 30 .
- the SoC 10 may comprise dedicated data source buses and in this case the RRI unit 32 may also monitor these dedicated data source buses.
- the RRI unit 32 may further be arranged to keep statistical data of read and write accesses to the buffers in the memory unit 16 , e.g., as described above.
- the RRI unit 32 When the RRI unit 32 detects that the data accumulation rate of one of the buffers is above an allocation threshold or below a de-allocation threshold, it may instruct the MAD unit 34 to allocate or to de-allocate a memory region in the memory unit 16 to or from the respective buffer.
- the MAD unit 34 may be arranged to keep a table or list of the memory regions allocated to each data source and of all free memory regions in the memory unit 16 .
- the MAD unit 34 When the MAD unit 34 is requested by the RRI unit 32 to allocate a memory region to a certain data source, it may respond by allocating one of the free memory regions to the respective data source.
- the MAD unit 34 when the MAD unit 34 is requested by the RRI unit 32 to de-allocate a memory region from a certain buffer, it may remove it from the list of allocated memory regions and add it to the list of free memory regions.
- the RRI unit 32 may comprise, for each data source, a respective data counter which is incremented in response to data written to the respective buffer and decremented in response to data retrieved from the buffer.
- the counter may be zeroed at regular time intervals, e.g., at user defined time intervals.
- the respective counter value may serve as an indication of the data accumulation rate of the respective buffer.
- the counter value may be compared to an allocation threshold and a de-allocation threshold, and a memory region may be added or removed to the respective buffer unless the data accumulation rate is between the de-allocation threshold and the allocation threshold.
- the allocation threshold should in this example be higher than the de-allocation threshold.
- the RRI unit 32 comprises a counter for keeping track of an amount of data written during a sliding window.
- the sliding window may have a user-defined length.
- a command for allocating a new memory region to the respective buffer may be issued.
- a mask bit may be asserted to block additional requests for the same data source for a certain period, e.g., a user-defined period.
- the MAD unit 34 may further be arranged to translate an address indicated by a transaction request on the system bus or on the optional dedicated data source buses, if present, to a physical address of a memory array which corresponds to the allocated memory regions.
- the MAD unit 34 may for example comprise a control logic unit 40 arranged to receive memory allocation and memory de-allocation commands from, e.g., the RRI unit 32 and to program a monolithic memory conversion table 38 with an original address, a physical address, and a size of allocated memory regions.
- the original address may relate to a monolithic memory region.
- the physical address may relate to a fragmented memory region.
- the monolithic conversion table may serve to translate an address from a monolithic address used by a data source and by the processing unit 20 to a physical address of the memory array.
- the control logic unit 40 may be arranged to keep track of allocated and free memory regions.
- the control and status register set 36 may include registers configurable by a user in order to define the parameters mentioned in this disclosure, for example, status registers for statistical data collection, e.g., referring to a number of memory regions allocated to each data source.
- the MAD unit 34 may for example be implemented as a set of registers or, alternatively, as a static random access memory unit. In the latter case, a small prefetch buffer may be required to maintain adequate performance.
- the memory control unit may be arranged to collect statistical data of incoming traffic and may use it to allocate memory dynamically.
- the present teaching may be useful in any application in which a buffer, e.g., a FIFO buffer or a data stack, is connected between a data source and a processor, the data source supporting several virtual channels.
- a buffer e.g., a FIFO buffer or a data stack
- the data source supporting several virtual channels.
- it may find application in a device with a network port supporting virtual channels.
- the total buffer required may be 512 KB in this example, for peak aggregated traffic over all channels. If an emergency buffer of, e.g., 20 KB for each virtual channel is included, an additional 8 times 20 KB may be required resulting in a total of 672 KB.
- the memory control unit 24 may be arranged to associate, with each of the buffers, an allocation condition and a de-allocation condition and to monitor whether the respective buffer characteristics satisfy the allocation condition and whether they satisfy the de-allocation condition.
- the memory control unit 24 in response to detecting that the buffer characteristics of the buffer in question satisfy the allocation condition, may allocate memory of the memory unit 16 to the respective buffer.
- the memory control unit 24 in response to detecting that the respective buffer characteristics satisfy the de-allocation condition, the memory control unit 24 may de-allocate memory of the memory unit 16 from the respective buffer.
- the buffer characteristics of a given buffer may include a hypothetical fill time.
- the fill time of a buffer is the time it takes to fill the buffer starting from a present fill level of the buffer and assuming a certain constant data accumulation rate.
- the hypothetical fill time may, for example, be determined as the remaining free memory of the buffer divided by the current data accumulation rate of the buffer.
- the memory control unit 24 may be arranged to set an allocation threshold for each of the buffers.
- the allocation threshold may be the same for each buffer. Alternatively, the buffers may have individual allocation thresholds associated with them.
- the allocation condition associated with a given buffer may be satisfied when the hypothetical fill time of the respective buffer is shorter than the respective allocation threshold. Both the data accumulation rate and the free memory of a given buffer may be determined and updated repeatedly.
- the hypothetical fill time being shorter than the allocation threshold may be an indication that the buffer may soon be full unless additional memory is allocated to the buffer. Therefore, allocating a memory region to the buffer when the hypothetical fill time is shorter than the allocation threshold may remove this risk.
- the system on chip may comprise two or more data sources, a memory unit, a memory control unit, and a processing unit.
- the system on chip may for example be the SoC 10 described above.
- the method may comprise, for each of the data sources, providing a data stream.
- the method may comprise, for each of the data streams: maintaining a buffer within the memory unit and routing the respective data stream to the processing unit via the respective buffer.
- Each of these buffers may have buffer characteristics which are variable and which include at least the amount of free memory of the respective buffer.
- the method may further include allocating and de-allocating memory of the memory unit to and from each buffer in dependence of the buffer characteristics of the respective buffer, thereby reallocating memory among the buffers.
- the data streams may for example be Ethernet data streams.
- the flow chart in FIG. 6 refers to one buffer among the two or more buffers in the memory unit.
- a fill time of the buffer may be estimated (Box 6 . 2 ).
- a hypothetical fill time T of the buffer may thus be determined.
- the hypothetical fill time T may, for example, be determined as the amount of free memory of the buffer divided by the current data accumulation rate Q of the buffer. If the hypothetical fill time T is greater than an allocation threshold T_ALLOC, i.e., if T>T_ALLOC, no memory is allocated to or de-allocated from the buffer, i.e., the buffer size may remain unchanged, and the fill time T may be re-estimated at a later time. The process of flow may thus return to Box 6 . 1 .
- the hypothetical fill time T (determined in Box 6 . 2 ) is less than or equal to the allocation threshold T_ALLOC, it may be determined whether the memory unit comprises a free memory region ( 6 . 3 ), i.e., a non-allocated memory region. If it contains a free memory region, that memory region may be allocated to the buffer ( 6 . 4 ). Otherwise, i.e., if the memory unit does not contain any free memory region that could be allocated to the buffer, an interrupt may be triggered ( 6 . 5 ) and the process flow may return to Box 6 . 1 .
- the interrupt may involve an error handling action.
- the error handling action may be user-configurable. From Box 6 . 5 , the process flow may return to Box 6 . 1 .
- a memory region may be de-allocated from a given buffer even when the data accumulation rate Q of the buffer is positive, provided that the hypothetical fill time is sufficiently large.
- each buffer may have a de-allocation threshold T_DEALLOC associated with it.
- the de-allocation threshold may be larger than the allocation threshold, and an empty memory region may be de-allocated from the buffer in response to detecting that the hypothetical fill time is larger than the de-allocation threshold.
- FIGS. 7, 8, and 9 schematically illustrate an example of allocating and de-allocating memory of the memory unit 16 to and from a first buffer 18 A and a second buffer 18 B.
- FIG. 7 relates to a first point in time at which the buffers 18 A and 18 B each consist of a respective initial region of the memory unit 16 . A remaining large region of the memory unit is not allocated in this example.
- FIG. 8 relates to a later second point in time in which part of the previously non-allocated region has now been allocated to the buffer 18 A. Note that the buffer 18 A now comprises two non-contiguous memory regions.
- FIG. 9 relates to a later third point in time in which part of the memory of the memory unit 16 has been de-allocated from the buffer 18 A and allocated to the buffer 18 B.
- connections as discussed herein may be any type of connection suitable to transfer signals from or to the respective nodes, units or devices, for example via intermediate devices. Accordingly, unless implied or stated otherwise, the connections may for example be direct connections or indirect connections.
- the connections may be illustrated or described in reference to being a single connection, a plurality of connections, unidirectional connections, or bidirectional connections. However, different embodiments may vary the implementation of the connections. For example, separate unidirectional connections may be used rather than bidirectional connections and vice versa.
- plurality of connections may be replaced with a single connection that transfers multiple signals serially or in a time multiplexed manner. Likewise, single connections carrying multiple signals may be separated out into various different connections carrying subsets of these signals. Therefore, many options exist for transferring signals.
- control unit 28 and the memory allocation unit 26 may be implemented as one module.
- processing unit 20 may be connected to the memory allocation unit 26 via a dedicated data line and not via the system bus.
- any arrangement of components to achieve the same functionality is effectively “associated” such that the desired functionality is achieved.
- any two components herein combined to achieve a particular functionality can be seen as “associated with” each other such that the desired functionality is achieved, irrespective of architectures or intermedial components.
- any two components so associated can also be viewed as being “operably connected,” or “operably coupled,” to each other to achieve the desired functionality.
- the examples, or portions thereof may implemented as soft or code representations of physical circuitry or of logical representations convertible into physical circuitry, such as in a hardware description language of any appropriate type.
- any reference signs placed between parentheses shall not be construed as limiting the claim.
- the word ‘comprising’ does not exclude the presence of other elements or steps then those listed in a claim.
- the terms “a” or “an,” as used herein, are defined as one or more than one.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Information Transfer Systems (AREA)
Abstract
Description
Claims (15)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/190,374 US9529745B2 (en) | 2014-02-26 | 2014-02-26 | System on chip and method of operating a system on chip |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/190,374 US9529745B2 (en) | 2014-02-26 | 2014-02-26 | System on chip and method of operating a system on chip |
Publications (2)
Publication Number | Publication Date |
---|---|
US20150242343A1 US20150242343A1 (en) | 2015-08-27 |
US9529745B2 true US9529745B2 (en) | 2016-12-27 |
Family
ID=53882352
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/190,374 Active 2035-02-09 US9529745B2 (en) | 2014-02-26 | 2014-02-26 | System on chip and method of operating a system on chip |
Country Status (1)
Country | Link |
---|---|
US (1) | US9529745B2 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160321153A1 (en) * | 2015-04-30 | 2016-11-03 | Advantest Corporation | Method and system for advanced fail data transfer mechanisms |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10666754B2 (en) * | 2015-05-29 | 2020-05-26 | Dell Products, L.P. | System and method for cache reservation in an information handling system |
US10318175B2 (en) * | 2017-03-07 | 2019-06-11 | Samsung Electronics Co., Ltd. | SSD with heterogeneous NVM types |
CN110928255B (en) * | 2019-11-20 | 2021-02-05 | 珠海格力电器股份有限公司 | Data anomaly statistical alarm method and device, storage medium and electronic equipment |
US11782634B2 (en) * | 2020-09-28 | 2023-10-10 | EMC IP Holding Company LLC | Dynamic use of non-volatile ram as memory and storage on a storage system |
US20230102843A1 (en) * | 2021-09-27 | 2023-03-30 | Nvidia Corporation | User-configurable memory allocation |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070174513A1 (en) * | 2006-01-23 | 2007-07-26 | Arm Limited | Buffering data during a data transfer |
US7552306B2 (en) | 2005-11-14 | 2009-06-23 | Kabushiki Kaisha Toshiba | System and method for the sub-allocation of shared memory |
EP2219114A1 (en) | 2009-02-13 | 2010-08-18 | Huawei Technologies Co., Ltd. | Method and apparatus for allocating storage addresses |
US20110153856A1 (en) * | 2009-12-18 | 2011-06-23 | At&T Intellectual Property I, L.P. | Method and System for Selecting a Delivery Method for Media on Demand |
US20130055004A1 (en) * | 2011-08-31 | 2013-02-28 | Kleanthes Koniaris | Hardware-based automatic clock gating |
-
2014
- 2014-02-26 US US14/190,374 patent/US9529745B2/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7552306B2 (en) | 2005-11-14 | 2009-06-23 | Kabushiki Kaisha Toshiba | System and method for the sub-allocation of shared memory |
US20070174513A1 (en) * | 2006-01-23 | 2007-07-26 | Arm Limited | Buffering data during a data transfer |
EP2219114A1 (en) | 2009-02-13 | 2010-08-18 | Huawei Technologies Co., Ltd. | Method and apparatus for allocating storage addresses |
US20110153856A1 (en) * | 2009-12-18 | 2011-06-23 | At&T Intellectual Property I, L.P. | Method and System for Selecting a Delivery Method for Media on Demand |
US20130055004A1 (en) * | 2011-08-31 | 2013-02-28 | Kleanthes Koniaris | Hardware-based automatic clock gating |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160321153A1 (en) * | 2015-04-30 | 2016-11-03 | Advantest Corporation | Method and system for advanced fail data transfer mechanisms |
US9842038B2 (en) * | 2015-04-30 | 2017-12-12 | Advantest Corporation | Method and system for advanced fail data transfer mechanisms |
Also Published As
Publication number | Publication date |
---|---|
US20150242343A1 (en) | 2015-08-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9529745B2 (en) | System on chip and method of operating a system on chip | |
US9128925B2 (en) | System and method for direct memory access buffer utilization by setting DMA controller with plurality of arbitration weights associated with different DMA engines | |
US11381515B2 (en) | On-demand packet queuing in a network device | |
US9769081B2 (en) | Buffer manager and methods for managing memory | |
CN101460912A (en) | Data communication flow control device and methods thereof | |
KR20210064377A (en) | Methods for maintaining cache coherency during reordering | |
US20040213151A1 (en) | Fabric access integrated circuit configured to bound cell reorder depth | |
US20170017424A1 (en) | System And Method For Enabling High Read Rates To Data Element Lists | |
US9753660B2 (en) | System and method for implementing hierarchical distributed-linked lists for network devices | |
KR20170082438A (en) | Hybrid module, system including the same, and method for providing memory interface of the system | |
US9063841B1 (en) | External memory management in a network device | |
US20200076742A1 (en) | Sending data using a plurality of credit pools at the receivers | |
US20170017423A1 (en) | System And Method For Enabling High Read Rates To Data Element Lists | |
US7844758B1 (en) | Dynamic resource allocation scheme for efficient use of a queue | |
US10061513B2 (en) | Packet processing system, method and device utilizing memory sharing | |
US8103788B1 (en) | Method and apparatus for dynamically reallocating buffers for use in a packet transmission | |
US20250007854A1 (en) | Data transmission scheduling for disaggregated memory systems | |
KR20220103794A (en) | Apparatus and method for managing packet transmission over memory fabric physical layer interfaces | |
EP3841484B1 (en) | Link layer data packing and packet flow control scheme | |
US20170017568A1 (en) | System And Method For Implementing Distributed-Linked Lists For Network Devices | |
US9582215B2 (en) | Packet processing system, method and device utilizing memory sharing | |
US9612950B2 (en) | Control path subsystem, method and device utilizing memory sharing | |
JP2005235046A (en) | Data transfer device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: FREESCALE SEMICONDUCTOR, INC., TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ATZMON, NIR;BAR, RON-MICHAEL;GLICKMAN, ERAN;AND OTHERS;SIGNING DATES FROM 20140213 TO 20140218;REEL/FRAME:032335/0727 |
|
AS | Assignment |
Owner name: CITIBANK, N.A., AS NOTES COLLATERAL AGENT, NEW YORK Free format text: SUPPLEMENT TO SECURITY AGREEMENT;ASSIGNOR:FREESCALE SEMICONDUCTOR, INC.;REEL/FRAME:032845/0442 Effective date: 20140502 Owner name: CITIBANK, N.A., AS NOTES COLLATERAL AGENT, NEW YORK Free format text: SUPPLEMENT TO SECURITY AGREEMENT;ASSIGNOR:FREESCALE SEMICONDUCTOR, INC.;REEL/FRAME:032845/0522 Effective date: 20140502 Owner name: CITIBANK, N.A., AS NOTES COLLATERAL AGENT, NEW YOR Free format text: SUPPLEMENT TO SECURITY AGREEMENT;ASSIGNOR:FREESCALE SEMICONDUCTOR, INC.;REEL/FRAME:032845/0442 Effective date: 20140502 Owner name: CITIBANK, N.A., AS NOTES COLLATERAL AGENT, NEW YOR Free format text: SUPPLEMENT TO SECURITY AGREEMENT;ASSIGNOR:FREESCALE SEMICONDUCTOR, INC.;REEL/FRAME:032845/0522 Effective date: 20140502 Owner name: CITIBANK, N.A., AS COLLATERAL AGENT, NEW YORK Free format text: SUPPLEMENT TO SECURITY AGREEMENT;ASSIGNOR:FREESCALE SEMICONDUCTOR, INC.;REEL/FRAME:032845/0497 Effective date: 20140502 |
|
AS | Assignment |
Owner name: FREESCALE SEMICONDUCTOR, INC., TEXAS Free format text: PATENT RELEASE;ASSIGNOR:CITIBANK, N.A., AS COLLATERAL AGENT;REEL/FRAME:037357/0763 Effective date: 20151207 |
|
AS | Assignment |
Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: ASSIGNMENT AND ASSUMPTION OF SECURITY INTEREST IN PATENTS;ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:037458/0438 Effective date: 20151207 Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: ASSIGNMENT AND ASSUMPTION OF SECURITY INTEREST IN PATENTS;ASSIGNOR:CITIBANK, N.A.;REEL/FRAME:037458/0479 Effective date: 20151207 |
|
AS | Assignment |
Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:038017/0058 Effective date: 20160218 |
|
AS | Assignment |
Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE INCORRECT APPLICATION NUMBERS 12222918, 14185362, 14147598, 14185868 & 14196276 PREVIOUSLY RECORDED AT REEL: 037458 FRAME: 0479. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT AND ASSUMPTION OF SECURITY INTEREST IN PATENTS;ASSIGNOR:CITIBANK, NA;REEL/FRAME:038665/0498 Effective date: 20151207 Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE APPLICATION NUMBERS PREVIOUSLY RECORDED AT REEL: 037458 FRAME: 0438. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT AND ASSUMPTION OF SECURITY INTEREST IN PATENTS;ASSIGNOR:CITIBANK, NA;REEL/FRAME:038665/0136 Effective date: 20151207 |
|
AS | Assignment |
Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: SUPPLEMENT TO THE SECURITY AGREEMENT;ASSIGNOR:FREESCALE SEMICONDUCTOR, INC.;REEL/FRAME:039138/0001 Effective date: 20160525 |
|
AS | Assignment |
Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12092129 PREVIOUSLY RECORDED ON REEL 038017 FRAME 0058. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:039361/0212 Effective date: 20160218 |
|
AS | Assignment |
Owner name: NXP, B.V., F/K/A FREESCALE SEMICONDUCTOR, INC., NETHERLANDS Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:040925/0001 Effective date: 20160912 Owner name: NXP, B.V., F/K/A FREESCALE SEMICONDUCTOR, INC., NE Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:040925/0001 Effective date: 20160912 |
|
AS | Assignment |
Owner name: NXP B.V., NETHERLANDS Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:040928/0001 Effective date: 20160622 |
|
AS | Assignment |
Owner name: NXP USA, INC., TEXAS Free format text: CHANGE OF NAME;ASSIGNOR:FREESCALE SEMICONDUCTOR INC.;REEL/FRAME:040579/0827 Effective date: 20161107 |
|
AS | Assignment |
Owner name: NXP USA, INC., TEXAS Free format text: CHANGE OF NAME;ASSIGNOR:FREESCALE SEMICONDUCTOR INC.;REEL/FRAME:040626/0683 Effective date: 20161107 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: NXP USA, INC., TEXAS Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNMENT DOCUMENTATION - INITIAL CONVENYANCE LISTED CHANGE OF NAME. PREVIOUSLY RECORDED ON REEL 040579 FRAME 0827. ASSIGNOR(S) HEREBY CONFIRMS THE UPDATE CONVEYANCE TO MERGER AND CHANGE OF NAME;ASSIGNOR:FREESCALE SEMICONDUCTOR INC.;REEL/FRAME:040945/0252 Effective date: 20161107 Owner name: NXP USA, INC., TEXAS Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNMENT DOCUMENTATION - INITIAL CONVENYANCE LISTED CHANGE OF NAME. PREVIOUSLY RECORDED ON REEL 040579 FRAME 0827. ASSIGNOR(S) HEREBY CONFIRMS THE UPDATE CONVEYANCE TO MERGER AND CHANGE OF NAME EFFECTIVE NOVEMBER 7, 2016;ASSIGNORS:NXP SEMICONDUCTORS USA, INC. (MERGED INTO);FREESCALE SEMICONDUCTOR, INC. (UNDER);REEL/FRAME:040945/0252 Effective date: 20161107 |
|
AS | Assignment |
Owner name: NXP USA, INC., TEXAS Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE NATURE OF CONVEYANCE PREVIOUSLY RECORDED AT REEL: 040626 FRAME: 0683. ASSIGNOR(S) HEREBY CONFIRMS THE MERGER AND CHANGE OF NAME;ASSIGNOR:FREESCALE SEMICONDUCTOR INC.;REEL/FRAME:041414/0883 Effective date: 20161107 Owner name: NXP USA, INC., TEXAS Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE NATURE OF CONVEYANCE PREVIOUSLY RECORDED AT REEL: 040626 FRAME: 0683. ASSIGNOR(S) HEREBY CONFIRMS THE MERGER AND CHANGE OF NAME EFFECTIVE NOVEMBER 7, 2016;ASSIGNORS:NXP SEMICONDUCTORS USA, INC. (MERGED INTO);FREESCALE SEMICONDUCTOR, INC. (UNDER);SIGNING DATES FROM 20161104 TO 20161107;REEL/FRAME:041414/0883 |
|
AS | Assignment |
Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12681366 PREVIOUSLY RECORDED ON REEL 039361 FRAME 0212. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:042762/0145 Effective date: 20160218 Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12681366 PREVIOUSLY RECORDED ON REEL 038017 FRAME 0058. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:042985/0001 Effective date: 20160218 |
|
AS | Assignment |
Owner name: NXP B.V., NETHERLANDS Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:050744/0097 Effective date: 20190903 Owner name: NXP B.V., NETHERLANDS Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:050745/0001 Effective date: 20190903 |
|
AS | Assignment |
Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12298143 PREVIOUSLY RECORDED ON REEL 042762 FRAME 0145. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051145/0184 Effective date: 20160218 Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12298143 PREVIOUSLY RECORDED ON REEL 039361 FRAME 0212. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051029/0387 Effective date: 20160218 Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12298143 PREVIOUSLY RECORDED ON REEL 042985 FRAME 0001. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051029/0001 Effective date: 20160218 Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION12298143 PREVIOUSLY RECORDED ON REEL 042985 FRAME 0001. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051029/0001 Effective date: 20160218 Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION12298143 PREVIOUSLY RECORDED ON REEL 039361 FRAME 0212. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051029/0387 Effective date: 20160218 Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION 12298143 PREVIOUSLY RECORDED ON REEL 038017 FRAME 0058. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051030/0001 Effective date: 20160218 Owner name: MORGAN STANLEY SENIOR FUNDING, INC., MARYLAND Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVE APPLICATION12298143 PREVIOUSLY RECORDED ON REEL 042762 FRAME 0145. ASSIGNOR(S) HEREBY CONFIRMS THE SECURITY AGREEMENT SUPPLEMENT;ASSIGNOR:NXP B.V.;REEL/FRAME:051145/0184 Effective date: 20160218 |
|
AS | Assignment |
Owner name: NXP B.V., NETHERLANDS Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVEAPPLICATION 11759915 AND REPLACE IT WITH APPLICATION11759935 PREVIOUSLY RECORDED ON REEL 040928 FRAME 0001. ASSIGNOR(S) HEREBY CONFIRMS THE RELEASE OF SECURITYINTEREST;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:052915/0001 Effective date: 20160622 |
|
AS | Assignment |
Owner name: NXP, B.V. F/K/A FREESCALE SEMICONDUCTOR, INC., NETHERLANDS Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE REMOVEAPPLICATION 11759915 AND REPLACE IT WITH APPLICATION11759935 PREVIOUSLY RECORDED ON REEL 040925 FRAME 0001. ASSIGNOR(S) HEREBY CONFIRMS THE RELEASE OF SECURITYINTEREST;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC.;REEL/FRAME:052917/0001 Effective date: 20160912 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |