Processing packets in cache memory

Electrical computers and digital processing systems: memory – Storage accessing and control – Hierarchical memories

Reexamination Certificate

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

C711S122000, C711S132000, C711S154000

Reexamination Certificate

active

06745289

ABSTRACT:

BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention is directed to processing network packets with multiple processing engines.
2. Description of the Related Art
Multi-processor computer systems include multiple processing engines performing operations at the same time. This is very useful when the computer system constantly receives new time-critical operations to perform.
For example, networking applications, such as routing, benefit from parallel processing. Routers receive multiple continuous streams of incoming data packets that need to be directed through complex network topologies. Routing determinations require a computer system to process packet data from many sources, as well as learn topological information about the network. Employing multiple processing engines speeds the routing process.
Another application benefiting from parallel processing is real-time video processing. A computer video system must perform complex compression and decompression operations under stringent time constraints. Employing multiple processors enhances system performance.
Parallel processing requires: (1) identifying operations to be performed, (2) assigning resources to execute these operations, and (3) executing the operations. Meeting these requirements under time and resource constraints places a heavy burden on a computer system. The system faces the challenges of effectively utilizing processing resources and making data available on demand for processing.
Over utilizing a system's processors results in long queues of applications waiting to be performed. Networking products employing traditional parallel processing encounter such processor utilization problems. These systems assign each incoming packet to a single processor for all applications. General processors, instead of specialized engines, perform applications requiring complex time-consuming operations. When each processor encounters a packet requiring complex processing, system execution speed drops substantially—processing resources become unavailable to receive new processing assignments or manage existing application queues.
Memory management also plays an important role in system performance. Many systems include main memory and cache memory, which is faster than main memory and more closely coupled to the system's processors. Systems strive to maintain frequently used data in cache memory to avoid time-consuming accesses to main memory.
Unfortunately, many applications, such as networking applications, require substantial use of main memory. Networking systems retrieve data packets from a communications network over a communications medium. Traditional systems initially store retrieved data packets in a local buffer, which the system empties into main memory. In order to perform applications using the data packets, the system moves the packets from main memory to cache memory—a time consuming process.
Traditional systems also incur costly memory transfer overhead when transmitting data packets. These systems transfer transmit packet data into main memory to await transmission, once processor operation on the data is complete—forcing the system to perform yet another main memory transfer to retrieve the data for transmission.
A need exists for a parallel processing system that effectively utilizes and manages processing and memory resources.
SUMMARY OF THE INVENTION
A multi-processor in accordance with the present invention efficiently manages processing resources and memory transfers. The multi-processor assigns applications to compute engines that are coupled to cache memory. Each compute engine includes a central processing unit coupled to coprocessor application engines. The application engines are specifically suited for servicing applications assigned to the compute engine. This enables a compute engine to be optimized for servicing the applications it will receive. For example, one compute engine may contain coprocessor application engines for interfacing with a network, while other coprocessors include different application engines.
The coprocessors also offload the central processing units from processing assigned applications. The coprocessors perform the applications, leaving the central processing units free to manage the allocation of applications. The coprocessors are coupled to the cache memory to facilitate their application processing. Coprocessors exchange data directly with cache memory—avoiding time consuming main memory transfers found in conventional computer systems. The multi-processor also couples cache memories from different compute engines, allowing them to exchange data directly without accessing main memory.
A multi-processor in accordance with the present invention is useful for servicing many different fields of parallel processing applications, such as video processing and networking. One example of a networking application is application based routing. A multi-processor application router in accordance with the present invention includes compute engines for performing the different applications required. For example, application engines enable different compute engines to perform different network services, including but not limited to: 1) virtual private networking; 2) secure sockets layer processing; 3) web caching; 4) hypertext mark-up language compression; and 5) virus checking.


REFERENCES:
patent: 5613136 (1997-03-01), Casavant et al.
patent: 5721855 (1998-02-01), Hinton et al.
patent: 6223260 (2001-04-01), Gujral et al.
patent: 6374329 (2002-04-01), McKinney et al.
patent: 6405289 (2002-06-01), Arimilli et al.
patent: 6430654 (2002-08-01), Mehrotra et al.
patent: 6539457 (2003-03-01), Mulla et al.
patent: 2001/0042190 (2001-11-01), Tremblay et al.
patent: 2002/0007443 (2002-01-01), Gharachorloo et al.
Tom Shanley, ‘Pentium Pro Processor System Architecture’, 1997, Mindshare, Inc, Chapters 1,2,7,9 and 11.*
Harmon, William “32-Bit Bus Master Ethernet Interface for the 68030 (Using the Macintosh SE/30),” Apr. 1993.
Troutman, Denise “DP83916EB-AT: High Performance AT Compatible Bus Master Ethernet Adapter Card,” Nov. 1992.

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Processing packets in cache memory does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Processing packets in cache memory, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Processing packets in cache memory will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-3312980

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.