Abstract

A novel packet switch architecture is disclosed. The packet switch utilizes internal queuing (i.e. recirculation loops) and output queuing (i.e. multiple paths to each destination) to provide a packet switch which offers superior performance in comparison to a packet switch which utilizes either of these queuing strategies alone. The combination of recirculation and output queues have complimentary effects. The output queuing reduces the number of recirculation loops needed and recirculation reduces the bandwidth requirements for an output buffered switch.

Description

FIELD OF THE INVENTION

The present invention relates to an architecture for a high speed and efficient packet switch. More particularly, the invention relates to a packet switch architecture which minimizes packet loss and maximizes utilization for a wide variety of traffic conditions by combining packet recirculation with the use of multiple paths to each destination port.

RELATED APPLICATION

A patent application Ser. No. 240,745 entitled "Packet Switch with Dynamic Allocation of Inputs", filed for J. N. Giacopelli and W. D. Sincoskie on even date herewith and assigned to the assignee hereof, now U.S. Pat. No. 4,866,701 Sept. 12, 1989 contains subject matter related to the subject matter of the present application.

BACKGROUND OF THE INVENTION

An important element for providing advanced telecommunications services requiring large bandwidth is a high capacity packet switch capable of interconnecting a plurality of input ports and a plurality of output ports. A packet switch that can connect any specific input port to any specific output port is known as a full access packet switch.

Typically, such a packet switch is synchronous. The packets routed therein are of fixed length and are contained in time slots. During a packet switch cycle, packets present at the input ports are routed through an interconnection network comprising part of the packet switch to specific output ports. If the interconnection network is self-routing, each packet arriving at an input port is provided with a header which contains the address of a specific output port to which the packet is to be routed. The interconnection network utilizes this address information to route the packets to the specific output ports.

A packet switch is internally non-blocking if it can deliver all packets from the input ports to the requested output ports when the packets are addressed to distinct output ports. However, there is still the possibility of external blocking, i.e. an internally non-blocking packet switch can still block if there are two simultaneous requests for the same output port. In this case, one or both packets to the same output port will be blocked. Accordingly, it is desirable for a packet switching architecture to be both internally and externally non-blocking.

One example of a minimally sized interconnection network is a banyan routing network. Even though a banyan network is sufficient for routing packets, routing decisions may cause internal collisions, even for a distinct set of addresses, reducing the throughput to an unacceptably low level. In short, the banyan network is an internally blocking network. The internal collisions in the banyan network can be eliminated by arranging packets in either ascending or descending order based on destination address before routing through the banyan network. The arrangement of packets in ascending or descending order can be accomplished through use of a Batcher sorting network connected in front of the banyan network. However, the resulting Batcher-banyan network is still externally blocking when two or more packets are simultaneously addressed to the same output.

Various packet switch architectures using Batcher and/or banyan networks have been proposed. These various architectures utilize a variety of techniques to resolve output port conflicts among packets and use a variety of techniques to buffer or queue packets that are not routed as a result of a conflict resolution process. The techniques used impact the size and complexity as well as the overall performance and efficiency of the packet switch architecture.

The performance of an architecture is generally characterized by its packet loss rate and the delay for a given link utilization. Both delay and loss are dependent on congestion due to traffic profiles, the ability of the interconnection network to route to the appropriate destination and the amount of and placement of packet buffers.

Conceptually, zero packet loss can be achieved with an ideal switch design. The ideal switch design requires full interconnectivity from each input to every output and infinitely long queues at each output. Arriving packets can be moved immediately from the inputs to the outputs where they are queued for access to outgoing trunks. In reality, full interconnectivity is expensive and the number of buffers must be finite. All packet switch architectures make design trade offs between the complexity of the interconnection network and the number and location of packet buffers provided.

Input buffered switches service packets on a first-come first-served basis by storing new arrivals in input queues to await service. This arrangement suffers from head of the queue blocking. Head of the queue blocking occurs since a packet at the top of the queue which cannot be transmitted to a particular output, blocks other packets within the queue from exiting even though they may be addressed to idle outputs. A variety of relatively complex techniques such as queue depth search have been used to solve this problem.

Output queuing generally involves the use of multiple routing paths to each output. Thus, a number of packets addressed to an output may be simultaneously routed thereto depending on the number of routing paths to the output. The packets are then queued at the output ports to obtain access to outgoing trunks. Thus, output queuing is associated with the use of relatively complex interconnection networks necessary to achieve multiple routing paths to each output.

Internal queuing may be accomplished as follows. A trap network may be located in between a Batcher sorting network and a banyan routing network at the outputs of the Batcher network. The trap network identifies packets with repeated output port addresses. The repeats can then be discarded or recirculated back to the switch input ports for later transmission. Each recirculation loop typically includes an internal queue for use by recirculating packets. The use of recirculation loops and queue for recirculating packets solves the head of the queue blocking problem for packet switches. However, prior art packet switches utilizing internal queuing are provided with a dedicated input at the Batcher sorting network for each recirculation path. Thus, for an interconnection network of given size, this substantially reduces the number of input ports which can be used for servicing newly arriving packets. Another way of stating this is that a fixed bandwidth through the network is allocated to recirculating packets.

In view of the above, it is an object of the present invention to provide a packet switch architecture which overcomes the shortcomings of switches which utilize input queuing alone, internal queuing alone, or output queuing alone. More particularly, it is an object of the invention to provide a packet switch architecture which combines a plurality of queuing approaches to form a more efficient packet switch.

SUMMARY OF THE INVENTION

In a preferred embodiment, the present invention is a packet switch architecture that is built around a Batcher sorting network and a plurality of banyan routing networks. This architecture combines both internal queuing (i.e. recirculation) and output queuing (i.e. multiple paths to each output) to form a packet switch which offers superior performance in comparison to a packet switch which utilizes either output queuing or internal queuing by itself. The combination of recirculation and output queuing have complimentary effects. The output queuing drastically reduces the number of recirculation loops, increasing the number of inputs that can serve newly arriving packets. Recirculation, on the other hand, reduces the bandwidth required for an output buffered switch.

BRIEF DESCRIPTION OF THE DRAWING

FIGS. 1A and 1B (arranged as shown in FIG. 1 and jointly referred to herein as FIG. 1) schematically illustrate a packet switch architecture utilizing internal and output queuing, in accordance with an illustrative embodiment of the present invention;

FIG. 2 shows the format of packets routed through the switch of FIG. 1;

FIG. 3 shows a comparator or trap cell forming part of the packet switch of FIG. 1;

FIG. 4 shows an output port controller forming part of the packet switch of FIG. 1;

FIG. 5 schematically illustrates the architecture of a packet switch with prioritized overflow in accordance with an alternative illustrative embodiment of the invention;

FIG. 6 schematically illustrates a packet switch architecture in which inputs are dynamically allocated between newly arriving and recirculating packets;

FIG. 7 shows the format of packets utilized in the packet switch of FIG. 6;

FIG. 8 schematically illustrates the architecture of a packet switch utilizing trunk grouping, in accordance with an illustrative embodiment of the present invention, and

FIGS. 9, 10 and 11 illustrate the format of packets utilized in the switch of FIG. 8.

In FIG. 1, the Batcher network inputs are divided into two groups. A first group of inputs a1. . . aM receives recirculating packets from the recirculation loops 30. A second group of inputs aM+1, aM+2. . . aN receives newly arriving packets via the trunks 17-M+1 . . . 17-N and input port controllers 16-M+1 . . . 16-N. It should be noted that in the network 10 of FIG. 1, the number of Batcher inputs allocated to receiving newly arriving packets is fixed and the number of Batcher inputs allocated to receiving recirculating packets is fixed. Another way of stating this is that a fixed fraction of the total bandwidth through the switch 10 is allocated to recirculating packets.

In any packet switching cycle, the packets present at the Batcher inputs are synchronously transmitted through the Batcher network and sorted, for example, in ascending order according to a destination address found in the packet header. Thus, during any packet switching cycle, the packets are sorted according to destination addresses at the Batcher outputs b1, b2. . . bN. Since the packet switch 10 of FIG. 1 is a synchronous switch, the packets may be viewed as occupying timeslots. The input signal at any Batcher input may be viewed as comprising a sequence of packets that are synchronous with the packets at any other input.

The format of each resulting packet is shown in FIG. 2. Each packet comprises a data field which contains user data and a header. The header comprises a routing field and a priority field. Both fields are ordered with their most significant bit first. The first bit of the routing field serves as an activity bit (A) where a logical "1" represents an active packet and a logical "0" represents an idle packet. Idle packets appear to have lower addresses than active packets. This causes idle packets to be grouped together at the lower numbered outputs of the Batcher sorting network 12. The remainder of the bits in the routing field comprise the destination address. For packets addressed to the same destination, the Batcher network 12 will continue to sort over the priority field. The priority field contains two sub-fields. The service priority (SP) sub-field identifies a class of service to which the packet belongs. The switch priority (PR) sub-field maintains a count of the number of slots a packet has been delayed (through the use of the recirculation loop 30) with an initial value of all logical "1 "s. Illustratively, for every slot a packet is delayed, the PR-subfield is decremented to increase the packet's priority relative to other packets within the same service class. Thus, the Batcher sorting network 12 of FIG. 1, produces at its outputs b1. . . bN a sorted list of packets ordered by destination address. Within each common destination, packets are ordered by priority of service.

In the packet switch 10 of FIG. 1, two banyan networks 14a, 14b are utilized. This means there are two routing paths directed to each output port controller 18 and two packets can simultaneously be routed to each output port controller 18. The arriving packets are queued at the output port controllers 18 to await access to the outgoing trunks 19.

If there are more than two packets present at the Batcher outputs, b1. . . bN contending for the same output port controller, a conflict resolution process is utilized to determine which packets will be permitted to exit to the output port controllers. Those packets not "winning" the conflict resolution process are recirculated via the recirculation loops 30 back to the Batcher inputs for re-transmission in the next packet switch cycle.

In the packet switch of FIG. 1, contention is resolved by identifying the two top packets within a common address block, since these will be the highest priority packets of the group. All other packets in a common address block are recirculated. The trap network 32 operates on the sorted output produced by the Batcher network 12 wherein packets with common destination addresses appear on adjacent outputs.

The trap network 32 comprises a column of trap cells, 34-3, 34-4 . . . 34-N, one for each output b3. . . bN of the Batcher network. The trap cells are address comparators. The trap network outputs are designated c1, C2. . . cN. As shown in FIG. 3, the comparator (i.e. trap cell) 34-i compares the destination address on Batcher output line bi with the destination address on Batcher output line bi-2. If the addresses are equal, then the packet on line bi losses the conflict resolution process and is recirculated. If the addresses are different, then the packet on line bi wins the contention resolution since it is either the highest or second highest priority packet to a particular destination. The packet present on Batcher output line bi emerges from the trap network comparator 34-i on the trap network output line ci. A trap bit in the packet is set if the compared destination addresses are equal and the packet is to be recirculated. The trap bit is not set if the compared destination addresses are not equal and the packet is to exit to an output port controller via a banyan network 14. Note that the top two Batcher output lines b1, b2 are not associated with comparators as these two lines contain two packets addressed to two different output port controllers or the two highest priority packets addressed to the same output port controller. In either event these packets can be routed by one or both of the banyan networks 14. However, delay elements (not shown) are utilized on the lines b1, b2 to insure that packets on lines b1, b2 remain synchronized with packets on lines b3. . . bN.

For a network with k banyan routing networks 14, k packets may be routed to a particular output port controller 16 in a packet switching cycle. In this case, the trap network compares the packet on Batcher output line bi with the packet on Batcher output line bi-k. If the addresses are equal, the packet on line bi is recirculated. Otherwise the packet on line i exits to the appropriate output port controller.

After the trapping network identifies packets to be routed to the output port controllers and packets to be recirculated, the two groups of packets should be separated so that the packets to be recirculated can be steered to the recirculation loops 30 and the packets exiting to the output port controllers can be steered to the banyan networks 14. In addition, to meet the non-blocking criterion for banyan networks, the exiting packets must be repacked into an ascending list by destination address without any gaps between them.

A concentrator network 36 performs this task by regrouping the packets into two separate lists. The packets to be recirculated are put in one list and the exiting packets are placed in the other list. A routing header constructed by the concentrator network 36 steers packets to be recirculated to one edge of the concentrator network (i.e. to the lower numbered outputs) and the exiting packets to the opposite edge (i.e. to the higher numbered outputs). Lists are built starting at opposite edges with each list growing towards the center of the concentrator outputs d1. . . dN. This procedure results in two separate sorted lists of packets.

The boundary between packets to be recirculated and exiting packets at the outputs of the concentrator 36 is arbitrary. Therefore each concentrator output desirably has access to both the recirculation loops 30 and one of the banyan networks 14a, 14b. Such access is provided by the selector network 38 which comprises a column of cells 40-1, 40-2 . . . 40-N, each having an input d1, d2. . . dN and two outputs e1a, e1b. . . eNa, eNb. The outputs e1a, e2a. . . eMa form part of the recirculation loops 30. The outputs e.sub.(M+1)a, e.sub.(M+2)a. . . eNa handle overflow packets which are lost via the lines 29. Thus any packets to be recirculated and appearing on this latter set of outputs are lost as overflow packets. Another way of looking at this is to note that the recirculation bandwidth is finite. If, in any switch cycle, the number of packets to be recirculated exceeds the recirculation bandwidth, packets are lost. Note that each of the recirculating loop 30 includes a shift register 52 for queuing recirculating packets. The shift registers together form the shared queue 54. The queue 54 is designated a shared queue because a packet from any Batcher input can be queued therein. Recirculating packets, stored in the queue 54, enter the Batcher network via the inputs a1. . . aM.

The packets present at the selector outputs e1b, e2b. . . eNb are transmitted to the banyan networks 14a, 14b for routing to particular output port controllers. When there are two packets addressed to the same output port controller present at the outputs e1b, e2b. . . eNb, the two packets are routed using separate banyan networks. To accomplish this, an inverse perfect shuffle wiring pattern 42 connects the outputs e1b, e2b. . . eNb to the banyan networks. This wiring pattern divides the sorted list of packets present at the outputs e1b, e2b. . . eNb into two separate but sorted lists having a unique set of destination addresses. In accordance with the inverse perfect shuffle wiring pattern, outputs e1b, e3b, e5b. . . e.sub.(N-1)b are connected to the inputs f1, f3, f5, . . . fN-1 of the banyan 14a and the outputs e2b, e4b. . . eNb are connected to the inputs g1, g3. . . gN-1 of the banyan network 14b. The inputs f2, f4. . . fN of the banyan 14a are set to zero and the inputs g2, g4, . . . gN of the banyan 14b are set to zero.

Illustratively, as shown in FIG. 4, each output port controller 18-i includes a shift register 20 associated with the input pi and a shift register 22 associated with the input fi. The shift registers 20, 22 serves as queues for packets arriving via the inputs pi, qi. The control device 24, enables the queues 20, 22 to access the outgoing trunk line 19-i. Note, in an alternative embodiment of the invention, there may be separate queues in the output port controllers for packets from different classes of services. Instead of shift registers, RAM devices may be used to queue packets at the output port controllers.

In FIG. 1, the network 12 is shown as a full Batcher network. However, if there are N/2 input port controllers and N/2 recirculation lines, an NxN Batcher network is not necessary. Since packets are already sorted at the trap network, a concentrator can maintain the relative positions of the trapped packets producing a sorted list of trapped packets. For this case an (N/2)xN/2 Batcher network is needed to sort the new arrivals. At this point, both the trapped packets and the new arrivals are separate but sorted lists that are then merged together. This is accomplished using a merge network which is actually the last stage of an NxN Batcher network to produce a sorted list of length N.

B. Prioritized Overflow

One problem with the packet switch architecture of FIG. 1, is that packets will be lost, when in any switching cycle, there are more packets to be circulated than there are recirculation loops. The sorting and trapping networks 12, 32 sort packets based on their destination addresses. This same order is maintained by the selector network 38. When overflow occurs, the highest addressed packets, regardless of the priority level, are lost first resulting in a network that favors low addresses.

To correct this problem, packets are preferably given access to the recirculation loops based on priority, not destination address, so that the lowest priority packets are the first to be lost at overload. To accomplish this, the packets are recorded before the recirculation. A packet switch 110 for carrying out such reordering is shown in FIG. 5. In FIG. 5, the Batcher network 66 replaces the concentrator network 36. In addition, the functions performed by the trap and selector networks 32 and 38 of FIG. 1 are modified in the networks 32' and 38' of FIG. 5. A concentrator network is limited to partitioning a list of packets into two parts. The concentrator network 36 of FIG. 1 partitions packets arriving at its inputs into lists of trapped (i.e. non-exiting packets including recirculating and overflow packets) and exiting packets. To partition trapped packets based on priority, multiple concentrator networks would be required. Such an approach would be both complex and expensive. Instead, the needed partitioning is carried out using the Batcher network 66.

A Batcher network can be used to partition a list into as many pieces as necessary by constructing a two part header whose first part groups common packets and a second part that orders packets within the groups.

The routing and priority fields of the packets (see FIG. 2) contain the necessary information to accomplish the desired partitioning. Thus, packets are compared in the trap network 32' of FIG. 5 in the same manner as they are compared in the trap network 32 of FIG. 1. However, in contrast to the network 32 of FIG. 1, trapped packets leaving the network 32' of FIG. 5 will have their routing and priority fields interchanged thus marking them for recirculation. The first bit of the priority field is now defined as a logical zero. Since an active packet's activity bit (see FIG. 2) is a logical one, trapped packets will have lower addresses than exiting packets because of the leading zero. Idle packets are constructed by the input port controllers 17-M+1 . . . 17-N to have an address between the valid ranges for trapped and exiting packets. In this case, when the packets leaving the trap network 32' are sorted using the Batcher sorting network 66, the packets at the outputs of the sorting network are arranged so that the trapped packets appear at the outputs with the lowest addresses, followed by the idle packets, followed by the exiting packets. The trapped packets are ordered from highest priority to lowest priority and the exiting packets are ordered according to destination address. Thus, the trapped packets with lowest priority are most likely to be lost as overflow packets.

The selector cells comprising the selector network 38' now make a decision based on the first bit of each arriving packet. All packets with leading zeros are directed towards the recirculation loops and all packets with leading ones are directed toward the banyan networks 14a, 14b. At the point where packets overflow (i.e. at the point where there are more packets with leading zeros than recirculating loops), the overflow packets will also be directed to the banyan networks (rather than to the distinct overflow lines 29 of FIG. 1). These packets, with leading zeros, will not effect the routing capabilities of the banyan networks. However, these overflow packets can be monitored by the output port controllers to obtain some idea of the fraction of packets being lost. In the event corresponding input port and output port controllers from duplex controllers, these overflow packets can be queued and resubmitted to the switch in a later switch cycle.

Before trapped packets enter the recirculation loops 30, the priority and routing fields are swapped again, placing the routing field in front. This header swap is carried out by the selector cells in the selector network 28' and ensures that packets reentering the switch via the inputs a1. . . an of the Batcher network 12 have the appropriate header format. In addition, the priority field of each recirculating packet will be decremented to increase its priority relative to packets which have not been recirculated or have been recirculated fewer times.

C. Dynamic Allocation of Batcher Inputs

In the packet switch architecture of FIG. 1, a fixed fraction of the packet switch bandwidth is dedicated to recirculating packets. This fraction may be represented as M/N which M is the number of Batcher inputs dedicated to recirculating packets and N is the total number of Batcher input lines. The remainder of the inputs are dedicated to newly arriving packets. Illustratively, M/N is about a third so that 33% of available bandwidth is dedicated for recirculating packets.

It should be noted, however, that the capacity of each Batcher input a1. . . aN is allocated 100% to recirculating or newly arriving packets. However, in reality the offered load on each input line dedicated to newly arriving packets is much less than 100%. For example, the offered load on an input line allocated to newly arriving packets may be 50% so that the input line is idle half the time. To take advantage of this excess capacity, Batcher inputs a1. . . aM may be dynamically allocated between newly arriving and recirculating packets. When this is done, it is not necessary to dedicate a fixed fraction of the switch bandwidth to recirculating packets.

An example of a packet switch with dynamic input allocation is shown in FIG. 6. The packet switch 80 of FIG. 6 comprises a plurality of input port controllers 16-1 . . . 16-N which receive newly arriving packets via the incoming trunks 17-1 . . . 17-N. The packet switch 80 also includes a plurality of output port controllers 18-1 . . . 18-N which interface with outgoing trunks 19-1 . . . 19-N. A conflict resolution and routing network 82 is provided. This network serves to resolve conflicts between packets addressed to the same output port and to route exiting packets to the output port controllers 18-1 . . . 18-N. Packets which cannot be routed to the output port controllers are sent via lines 85 to the recirculation network 90 for return to the input port controllers. Illustratively, the network 82 comprises a Batcher sorting network, a trap network for identifying exiting and recirculating packets, a concentrator network for separating the exiting and recirculating packets into separate lists, and a selector network for transmitting recirculating packets to the recirculation network 90 and for transmitting exiting packets to one or more banyan routing networks for routing to the output port controllers. Such a conflict-resolution and routing arrangement is illustrated in FIG. 1 and comprises elements 12, 32, 36, 38, 14a, 14b of FIG. 1. The recirculation network 90 comprises concentrator 92, multiplexer 94, and banyan network 96.

The packet switch 80 of FIG. 6 operates as follows. All newly arriving packets at each input port controller are delayed for a period in excess of one packet slot. This provides time for an input port controller to decide in time slot T if it will be idle during the time slot T+2. If an input port controller will be idle during time slot T+2, then it serves as an access point for a recirculating packet. In such a case, the input port controller submits an active steering packet to the concentrator 92 that will be paired with a recirculating packet to provide a routing header that returns the recirculating packet to the input port controller that issued the steering packet. The input port controller that receives the recirculating packet, will then resubmit the recirculating packet to the network 82 in the T+2 time slot.

The packet formats for both the data packet and the steering packets are shown in FIG. 7. The data packet 91 in packet time slot T contains a destination address field (DA) 97, a data field 98, and an additional field 99 following the data field. When a data packet is submitted by an input port controller into the switch, the additional field 99 is in the don't care state as designated by the symbol "X". As shown in FIG. 7, the steering packet 101 is the same size as the data packet 91. However in the steering packet 101, the fields 107 and 108 (corresponding to the fields 97 and 98) are in the don't care state and the field 109 (corresponding to the field 99) contains the address of the input port controller which generated it.

The network 80 dynamically allocates inputs between newly arriving packets and recirculating packets as follows. As indicated above, an input port controller that will be idle during packet slot T+2 generates an active steering packet during the packet time slot T that includes its address in the source address field 109. An input port controller that will not be idle during the packet time slot T+2 generates an idle steering packet during the packet slot T as indicated by an activity bit being set to zero. The active and idle steering packets are concentrated by the concentrator network 92. Since the input port controllers 16-1 . . . 16-N are connected in order of address to the concentrator 92, the resulting concentrated list meets the non-blocking criteria of the banyan network 96. The steering packets generated during packet slot T are timed so that they meet up synchronously with the don't care fields 99 of recirculating packets of packet slot T. The multiplexer 94 gates the source field 109 of each steering packet with the field 99 of a corresponding recirculating data packet of slot T. The source address which is now located in the field 99 of a recirculating packet of slot T serves as a header to route a recirculating data packet of slot T+1 (i.e. the immediately following recirculating packet) back to the input port controller indicated by the source address. The input port controller then resubmits the data packets to the switch during the slot T+2 when it would otherwise be idle.

The packets switch architecture 80 of FIG. 2, dynamically allocates recirculating lines by sharing access with underutilized input lines. The amount of bandwidth available for recirculating packets will fluctuate on a slot-by-slot basis. It should be noted that even if there are more recirculating packets than idle port controllers for a particular slot, these excess recirculating packets will pair up with idle steering packets and emerge from the banyan network 96 at random locations. These packets may then be queued at the input port controllers for possible later re-submission to the switch. D. Trunk Grouping

Trunk grouping allows increased bandwidth to be engineered on a per route basis. (see e.g. "Multichannel Bandwidth Allocation", U.S. patent application Ser. No. 175,239 filed for Achille Pattavina on Mar. 30, 1988 and assigned to the assignee hereof). Trunk grouping pools switch resources by assigning multiple outgoing trunks to a common group. This group is then treated as a single high bandwidth trunk. A common address (logical address) represents each trunk group but each individual trunk in a group requires a routing address (physical address) that uniquely identifies it.

A packet switch 200 utilizing recirculation, multiple routing paths to each output and trunk groups is illustrated in FIG. 8. In the packet switch 200 of FIG. 8, contention is resolved using logical addresses while routing is performed using physical addresses. Translation from logical addresses to physical addresses is carried out using control packets. The control packets are generated using the control packet generator 202 which stores the necessary translation information.

The initial packet headers are shown in FIG. 9. Each newly arriving data packet is formed by the input port controllers and contains an activity bit (A), a logical address field (LA), a logical address index field (LA INDEX), a fixed logic "1", a packet index field (PKT INDEX), a trap flag (T), a priority field (PR) and a data field. The symbol X indicates a "don't care" state. Initially for each data packet, the input port controllers set the LA INDEX, PKT INDEX and T flag to logic "0". The activity bit is set to logic 1 for active packets and zero for idle packets. The LA field contains the logical address for the packet. Packets entering the Batcher network 12 from the recirculation loops 30 have the same format as the data packet of FIG. 9.

The format of the control packets generated by the control packet generator 202 is also shown in FIG. 9. Each control packet contains an activity bit (A), logical address field (LA), logical address index field (LA INDEX), a fixed logic "0", a trap flag (T), and physical address field (PA).

During a switch cycle, the NxN Batcher network 12 produces a sorted list based on the logical addresses and, within each common logical address, packets are ordered by priority. The outputs of the NxN Batcher network 12 are connected to two running sum adder networks 204a, 204b. Each of the networks 204a, 204b is an (N/2)×(N/2)K network. The outputs of the Batcher network 12 are connected to the running sum adder networks using the inverse perfect shuffle wiring pattern discussed in connection with FIG. 1. The adder networks 204a, 204b performs two operations. First they compute a running count of the number of packets having a particular logical address and store this value within each LA INDEX field. Secondly, they compute a running count over all data packets. This value is stored in the PKT index field of each packet. The combination of fields, A, LA, LA INDEX and fixed logic "1" uniquely selects a control packet to be assigned to a particular data packet. The fixed logic "1" field serves as the least significant bit of the header of a data packet, thus making all data packets appear to have an odd address. The fixed logic "0" field of the control packet makes all control packets appear to have an even address.

The control packets are merged with the data packets using the merge networks 206a, 206b. At the outputs of the merge networks is a combined sorted list comprising data and control packets. The sorting is based on logical addresses. The trap networks 208a, 208b pair control packets to data packets if there is a control packet on line i that matches the A, LA, LA INDEX fields of the data packet on line i+1. For paired packets, the physical address (PA) field is copied from the control packet into the data packet with the priority field (PR) being shifted back one position. Data packets which do not pair up with control packets are not altered. After completing the process, the trap networks 208a, 208b rotate the A, LA, and LA INDEX fields to the back of the header. The headers at the outputs of the trap network 208a, 208b are shown in FIG. 10. More particularly, FIG. 10 shows the header of a non-paired or trapped data packet and the header of a paired packet as well as the header of a control packet. For paired packets, the trap flag is set at logic "1". For unpaired packets, the trap flag is set at logic "0".

The non-paired packets are eventually recirculated back to the inputs of the Batcher network 12 using the recirculation loops 30. Illustratively, a packet is not paired in the following circumstances. In a packet switch cycle, the number of control packets produced for each logical address is equal to the number of physical addresses corresponding to the particular logical address multiplied by the number of routing paths to each physical address. If, in a particular switch cycle, the number of data packets having a particular logical address exceeds the number of control packets for a particular logical address (and therefore exceeds the routing capacity of the switch to the logical address), the packets will not be paired and will be recirculated.

The reverse banyan networks 210a, 210b separate control packets from data packets. A reverse banyan network is a mirror image of a banyan network except it routes based on least significant bit first. It is non-blocking for a continuous set of ascending or descending addresses though idle packets may be interleaved with active packets. The reverse banyan networks 210a, 210b direct all data packets to the second Batcher network 212 using the fixed logical 1 and PKT index fields as a routing header. During the routing process, the fixed logical 1 and PKT INDEX fields are rotated to the back of the header producing the format for trapped packets (i.e. packets to be recirculated) and paired packets shown in FIG. 11. The Batcher 212 concentrates the list of data packets based on priority or physical address. The selector network 214 separates the paired and unpaired packets. The paired packets are routed through the banyan networks 14a, 14b to the output port controllers 18 based on physical address. The unpaired packets are modified by the selector network 212 so that they have the format of newly arriving packets. These packets are then routed via the recirculation loops 30 back to the Batcher 12.

CONCLUSION

A packet switch architecture which utilizes both recirculation and output queuing has been disclosed. Finally, the above described embodiments of the invention are intended to be illustrative only. Numerous alternative embodiments may be devised by those skilled in the art without departing from the spirit and scope of the following claims.

Claims (20)

What is claimed is:

1. A packet switch comprising:

a plurality of inputs and a plurality of outputs,

a sorting network connected to said inputs for sorting data packets to be routed through said switch,

a trap network connected to said sorting network for identifying said data packets as exiting packets or recirculating packets,

a plurality of recirculation loops connected to at least some of said inputs,

a plurality of distinct routing networks for providing a plurality of routing paths to each of said plurality of outputs, and

a selector network for applying said identified recirculating packets to said recirculation loops and said identified exiting packets to said routing networks.

2. The packet switch of claim 1 wherein said sorting network is a Batcher network.

3. A packet switch comprising

a plurality of inputs and a plurality of outputs,

a sorting network connected to said inputs for sorting data packets to be routed through said switch,

a trap network connected to said sorting network for identifying said data packets as exiting packets or recirculating packets,

a plurality of recirculation loops connected to at least some of said inputs,

a plurality of distinct routing networks for providing a plurality of routing paths to each of said plurality of outputs, and

a selector network for applying said identified recirculating packets to said recirculation loops and said identified exiting packets to said routing networks,

wherein said switch further includes a concentrator network located between said trap network and said selector network.

4. The packet switch of claim 1 wherein each of said routing networks is a banyan network.

5. A packet switch comprising

a plurality of inputs and a plurality of outputs,

a sorting network connected to said inputs for sorting data packets to be routed through said switch,

a trap network connected to said sorting network for identifying said data packets as exiting packets or recirculating packets,

a plurality of recirculation loops connected to at least some of said inputs,

a plurality of distinct routing networks for providing a plurality of routing paths to each of said plurality of outputs, and

a selector network for applying said identified recirculating packets to said recirculation loops and said identified exiting packets to said routing networks,

wherein each of said routing networks is a banyan network, and

wherein said banyan networks and said selector network are connected by means of an inverse perfect shuffle wiring pattern.

6. A packet switch comprising

a plurality of inputs and a plurality of outputs,

a first sorting network connected to said inputs for sorting data packets to be routed through said switch,

a trap network connected to said sorting network for identifying said data packets as exiting packets or recirculating packets,

a plurality of recirculation loops connected to at least some of said inputs,

a plurality of distinct routing networks for providing a plurality of routing paths to each of said plurality of outputs, and

a selector network for applying said identified recirculating packets to said recirculation loops and said identified exiting packets to said routing networks,

wherein a second sorting network is connected between said trap network and said selector network for ordering said recirculating packets according to priority.

7. The packet switch of claim 1 wherein said data packets include logical addresses and said switch includes means for generating control packets.

8. A packet switch comprising

a sorting network connected to said inputs for sorting data packets to routed through said switch,

a trap network for receiving said data packets sorted by said sorting network and for identifying said data packets as exiting packets or recirculating packets,

a plurality of recirculation loops connected to at least some of said inputs,

a plurality of distinct routing networks for providing a plurality of routing paths to each of said plurality of outputs, and

a selector network for applying said identified recirculating packets to said recirculation loops and said identified exiting packets to said routing networks,

wherein said data packets include logical addresses and said switch includes means for generating control packets, and

wherein said trap network identifies data packets as recirculating or exiting packets by comparing said data packets with said control packets.

9. The packet switch of claim 8 wherein the logical addresses of said exiting packets are translated into physical addresses by inserting into each exiting packet a physical address from a corresponding control packet.

10. A packet switch comprising

a plurality of inputs for receiving data packets to be routed through said switch

a plurality of outputs,

means for resolving conflicts among data packets addressed to the same output at the same time by designating the data packets as exiting or recirculating packets,

network means for routing said exiting packets to said outputs, said network means permitting at least two packets to be simultaneously routed to each of said outputs, and

means for providing a plurality of recirculation loops via which said recirculating packets are routed back to said inputs.

11. A packet switch comprising

a plurality of inputs for receiving data packets to be routed through said switch,

a plurality of outputs,

means for resolving conflict among data packets addressed to the same output at the same time so as to designate the data packets as exiting or recirculating packets,

means for providing a plurality of routing paths to each output via which said exiting packets are routed to said outputs, and

means for providing a plurality of recirculation loops via which said recirculating packets are routed back to said inputs,

wherein said packet switch includes means for ordering said recirculating packets according to priority so that if the number of recirculating packets exceeds the number of recirculating loops only packets of lower priority will be lost.

12. A packet switch comprising

a plurality of inputs for receiving data packets to be routed through said switch,

a plurality of outputs,

means for resolving conflicts among data packets addressed to the same output at the same time so as to designate the data packets as exiting or recirculating packets,

means for providing a plurality of routing paths to each output via which said exiting packets are routed to said outputs, and

means for providing a plurality of recirculation loops via which said recirculating packets are routed back to said inputs,

wherein said data packets contain logical addresses and wherein said switch includes means for generating control packets, said conflict resolving means utilizing said control packets to identify exiting and recirculating packets and to translate said logical addresses to physical addresses.

13. The packet switch of claim 10 wherein said means for providing a plurality of paths comprises a plurality of banyan networks.

14. The packet switch of claim 10 wherein each of said recirculating loops includes a shift register for queuing recirculating packets.

15. A packet switch comprising

a plurality of inputs for receiving data packets to be routed through said switch,

a plurality of outputs,

means for designating the data packets as exiting packets or recirculating packets,

a plurality of distinct routing networks via which each of said exiting packets can be routed to a particular one of said outputs,

means for queuing said exiting packets at said outputs, and

a plurality of recirculating loops including queuing means via which said recirculating packets are routed to said inputs.

16. A packet switch comprising

a plurality of outputs,

a plurality of inputs for receiving packets to be routed through said switch to said outputs,

means for designating each of said packets as a recirculating packet or an exiting packet

recirculation means for routing at least a fraction of said packets designated recirculating packets back to said inputs, and

a plurality of distinct routing networks via which each of said exiting packets can be routed to a particular one of said outputs.

17. The packet switch of claim 16 wherein said recirculation means includes means for queuing said recirculating packets.

18. The packet switch of claim 17 wherein a queue is associated with each of said outputs.

19. A packet switch comprising

a plurality of outputs,

a plurality of inputs for receiving packets to be routed through said switch to said outputs,

means for designating each of said packets as a recirculating packet or an exiting packet,

recirculation means for routing at least a fraction of said recirculating packets back to said inputs, and

network means for routing said exiting packets to said outputs, said network means enabling at least two exiting packets to be routed simultaneously to the same output.