WO1993020637A1 - Congestion control for cell networks - Google Patents

Congestion control for cell networks Download PDF

Info

Publication number
WO1993020637A1
WO1993020637A1 PCT/US1993/002097 US9302097W WO9320637A1 WO 1993020637 A1 WO1993020637 A1 WO 1993020637A1 US 9302097 W US9302097 W US 9302097W WO 9320637 A1 WO9320637 A1 WO 9320637A1
Authority
WO
WIPO (PCT)
Prior art keywords
rate
congestion
cell
cells
bursty data
Prior art date
Application number
PCT/US1993/002097
Other languages
French (fr)
Inventor
Lionel A. Bustini
Patrick D. Daley
Charles M. Corbalis
Original Assignee
Stratacom, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Stratacom, Inc. filed Critical Stratacom, Inc.
Priority to EP93907328A priority Critical patent/EP0634068A1/en
Priority to JP5517450A priority patent/JPH07508378A/en
Publication of WO1993020637A1 publication Critical patent/WO1993020637A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L12/5602Bandwidth control in ATM Networks, e.g. leaky bucket
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5614User Network Interface
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5619Network Node Interface, e.g. tandem connections, transit switching
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5629Admission control
    • H04L2012/563Signalling, e.g. protocols, reference model
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5629Admission control
    • H04L2012/5631Resource management and allocation
    • H04L2012/5632Bandwidth allocation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5629Admission control
    • H04L2012/5631Resource management and allocation
    • H04L2012/5632Bandwidth allocation
    • H04L2012/5635Backpressure, e.g. for ABR
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5638Services, e.g. multimedia, GOS, QOS
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/54Store-and-forward switching systems 
    • H04L12/56Packet switching systems
    • H04L12/5601Transfer mode dependent, e.g. ATM
    • H04L2012/5638Services, e.g. multimedia, GOS, QOS
    • H04L2012/5646Cell characteristics, e.g. loss, delay, jitter, sequence integrity
    • H04L2012/5652Cell construction, e.g. including header, packetisation, depacketisation, assembly, reassembly

Definitions

  • the present invention relates to the field of cell switching network communications. More specifically, the present invention relates to apparatus and method of traffic congestion prevention and control within the cell switching network. BACKGROUND OF THE INVENTION
  • a typical fully-integrated voice and data network using digital trunk lines includes a number of geographically distant interconnect nodes.
  • Each node acts as a cell exchanger for receiving and forwarding cell information to its destination.
  • each node is capable of interfacing to user generated voice and data streams, then segmenting and assembling the streams into a more efficient cell format for transmission over a closed network using digital lines.
  • Each node is also capable of receiving data from other network nodes and forwarding that data through to other network nodes to its ultimate destination. All terminal nodes also provide the necessary interface cards required to reassemble the data contained in the cells into a standard user data stream format.
  • a typical modern node is capable of handling six classes of cell traffic, each class having different characteristics and different service requirements.
  • the six classes of traffic include: (1) High priority ("HP") for node-to-node control messages;
  • LSS Low speed statistical
  • High speed deterministic (HSD) traffic for full period voice or data at data rates exceeding 64 Kbps;
  • Bursty data for point-to-point traffic generated by such sources as local area networks, bridges, routers or high speed packet switches within the cell network;
  • Multicast traffic which is of the same type as bursty data except that it is simultaneously broadcast from one source to many destinations (vs. point-to-point).
  • Each of these six traffic types are buffered at each network node in accordance with their particular sensitivities to network delay and cell loss.
  • Cell loss may occur due to intermittent short term overload of network bandwidth and lack of adequate buffer capacity.
  • voice traffic is relatively delay sensitive and insensitive to occasional cell loss.
  • data traffic such as file transfers, is relatively insensitive to delay but is data loss sensitive.
  • High priority data is both delay and loss sensitive.
  • each class of traffic is typically placed in a preassigned queue, each with a different service priority.
  • servicing algorithms are typically employed to discriminate between traffic classes in order to allocate bandwidth. Delay is managed by properly sizing the queue depths and prioritizing transmission within a class based upon a measure of the time that a cell has been in the network as, for example, by use of time stamps and hop counts.
  • Short-term congestion typically handled by discarding traffic at the queue, may be done haphazardly or preferably selectively by having cells marked with their "discard eligibility”.
  • Long-term congestion is controlled by admission policies that allocate resources (bandwidth and buffers) at the time a connection is established. If no resources are available, the connection is not allowed.
  • the token bucket has a fixed maximum capacity corresponding to the maximum packet burst duration. Tokens arriving to a full bucket are dropped.
  • Peak rate control is accomplished by means of a spacer that introduces a suitable delay proportional to the length of the prior transmitted packet.
  • bursty data sources are characterized by intermittent high data rate burst with significant spans of inactivity between bursts. Under these circumstances, the above described simplest system would result in underutilization of the bandwidth capacity of the system because of the prescribed safe bandwidth limit assigned to the virtual connection session.
  • Average bandwidth utilization efficiency is typically improved by introducing "colored" tokens, for example, green and red.
  • Green tokens correspond to packets received for transmission that fall within the minimum guaranteed bandwidth protocol while the red tokens correspond to packet data received for transmission in excess of the guaranteed minimum rate.
  • Intermediate nodes provide per trunk FIFO buffer service and use the colors associated with each packet for congestion control.
  • green packets are protected and passed along while red packets are discarded upon arrival whenever the chosen metric (usually queue lengths) for congestion threshold is exceeded. Even though discarding of packets implies retransmission of the lost packet data, the system is represented as improving the average utilization of bandwidth capacity.
  • the window control method indirectly controls the effective network data rate by having the user adjust the window duration controlling the number of contiguous packets that can be transmitted at a given time.
  • each network layer packet has a header that is used for "housekeeping" functions, including routing. Within that header, a bit is assigned as a congestion indication bit. The packet source clears the congestion indication bit when a packet is originally transmitted from a source node end system.
  • Each network router is equipped with congestion detection means that monitors the average queue length at the router at the time the packet arrives.
  • the average queue length is determined by the number of packets that are queued and in service, averaged over an interval. This interval corresponds to the duration of the sum of the last busy plus idle cycle duration plus the busy interval of the current cycle.
  • an acknowledgement message is generated if no error is detected and is sent together with the congestion indication bit status associated with the acknowledged packet.
  • the user transport entity updates the window size based on the number of packet congestion bits set during the last transmission window used. If at least 50% of the bits are set, the window size is reduced from its current value. Otherwise, it is increased.
  • updating of the user's transparent entity window size occurs only after a time period corresponding to the duration of the last two windows. This delay allows enough time to see the effects of the last window change.
  • the window size adjustment algorithm provides for increasing window size by linear increments and for decreasing window size by a multiplicative factor.
  • Ramakrishnan and Jain describe a system using window control at the ISO Transport layer using window duration control rather than rate control. Rate control is indirectly controlled by the limiting actions of acknowledgements and window length. Because transmission rate is a direct measure of bandwidth, better short term control of this system resource can be obtained by direct rate control.
  • the authors describe a system carrying different classes of service in which a cell rate is assigned to variable bit rate data that is less than the available peak cell rate so as to provide capacity for statistical multiplexing of data.
  • a new connection is admitted only if sufficient network resources are available to support that connection.
  • Cells offered to the network in excess of the assigned rate are immediately discarded because, in the opinion of the authors, the tagging of floating cells for possible subsequent discard appears to yield no substantial advantage.
  • Traffic enforcement is applied on the ingress direction of each virtual circuit.
  • a leaky bucket or credit manager algorithm is used to control the rate.
  • the Cooper article refers to a reactive (feedback) control method that combines cell tagging with explicit congestion notification (“ECN").
  • ECN asynchronous transfer mode
  • ATM asynchronous transfer mode
  • the serving ATM switches operate at one of two preset thresholds for rate enforcement on each variable bit rate virtual circuit using ECN. The lower threshold is used during congestion.
  • the ECN message may be transported between network nodes possibly using a separate virtual circuit, and possibly using a bit in the ATM cell header.
  • One object of the present invention is to optimize the use of available system bandwidth.
  • Another object of the present invention is the optimization of network resource allocation to bursty traffic when sharing a network digital trunk with the five other traffic classes.
  • Another object of the present invention is to provide a dynamic bandwidth (or data rate) allocation scheme that allows individual users to use unused network capacity for increasing throughput when necessary to accommodate the peak loads of individual users.
  • Another object is to provide a method used for allocation that also helps to ensure "fairness" in network resource availability to all users by using linear incremental increases and multiplicative decreases in data rates together with guaranteed minimum rate allocations.
  • Another object is to provide a cell switching network operating at the International Standards Organization ("ISO") Data Link Layer that direct rate control based on virtual circuit connection congestion.
  • ISO International Standards Organization
  • a further object of the invention is to provide for the fast recovery from congestion.
  • Another object is to provide congestion relief by tagging cells at the source node when the virtual circuit's guaranteed minimum information rate (“MIR”) is exceeded so that if a prescribed cell loss priority (“CLP”) of queue depth threshold is exceeded at any intermediate node, tagged cells using that connection may be dropped.
  • MIR virtual circuit's guaranteed minimum information rate
  • CLP cell loss priority
  • Another object is to provide an adaptive interval between sending rate updates based on the actual measured round trip of the connection. Another object is to provide each node with means for measuring virtual connection round trip transmission time for use in adaptive interval control.
  • Another object is to provide means for generating supervisory cells for transmission of congestion control information from destination node to source node in the absence of normal two-way traffic.
  • Another object is to provide a quiescent information sending rate ("QIR”) greater than the guaranteed minimum information rate (“MIR”) and less than the peak information rate (“PIR").
  • QIR quiescent information sending rate
  • MIR guaranteed minimum information rate
  • PIR peak information rate
  • Another object is to provide a no rate change feedback indicator in the feedback cell header sent from destination to source node while waiting for the effects of the previous adjustments to occur or while measuring the effects of previous rate adjustments.
  • the control system comprises a source node, optional intermediate nodes, and a destination node.
  • the source node is capable of accepting, queueing and buffering user bursty data, formatting the bursty data into cells, and transmitting the cells to a destination node by a virtual circuit connection through a cell switching network to a destination node.
  • the rate of transmission is incrementally controlled by the state of congestion on the virtual network as reported by the destination node to the source node using the existing two-way virtual connection.
  • Intermediate nodes accept, queue, buffer and forward the bursty cells toward the destination together with an associated incipient congestion indicator.
  • Incipient congestion is measured by monitoring of the virtual connection queue and buffer lengths.
  • the destination node receives the data cells together with the congestion indicator, reconstructs and formats the data for delivery to the end user, and counts the received congestion indicators over an adaptive interval.
  • a feedback rate control signal is generated from the congestion indicator count.
  • Figure 1 is a block diagram of a telecommunication network for voice and data
  • Figure 2 is a cell format diagram
  • Figure 3 is a functional block diagram of a telecommunication node
  • Figure 4 is a data flow diagram
  • - Figure 5 illustrates an example of congestion in a cell network
  • Figure 6 illustrates the effects on throughput and delay due to congestion
  • FIG. 7 illustrates a cell switching network internal congestion avoidance (“ICA) scheme
  • FIG 8 is a functional block diagram of a frame relay pad ("FRP");
  • FIG 9 is a functional block diagram of the transmit portion of a transmitter/receiver ("TXR");
  • FIG. 10 shows cell loss priority ("CLP") two level thresholding in a TXR
  • Figure 11 shows the architecture of a FRP and a frame relay interface (FRI);
  • FIG. 12 is a block diagram of the T1 transmitter/receiver
  • Figure 13 is a diagram of a bursty data queue
  • Figures 14(a) and 14(b) comprise a flow diagram of a cell traffic servicing routine;
  • Figure 15 shows the c-bit count interval timing using the regeneration cycle algorithm;
  • Figure 16 is a flow diagram of the data rate control algorithm
  • Figure 17 is a flow diagram of the ICA credit manager function.
  • Figure 18 is a flow diagram for the servicing of the muxbus cell transmit queue by the cell transmitter.
  • FIG. 1 illustrates a fully-integrated voice and data T1 telecommunication network 20 using telecommunication nodes 22, also referred to as integrated cell exchangers 22.
  • the network 20 shown is a domestic network.
  • the illustrated network 20 can be modified in known ways to accommodate international traffic by the addition of various interfaces to user equipment and T1-to-E1 interfaces to each node 22.
  • Each node 22 incorporates a T1 transmitter/receiver that includes the fair queuing and servicing circuitry.
  • the T1 transmitter/receivers support six classes of cell traffic: high priority (HP), voice, low speed statistical (LSS), high speed deterministic (HSD), bursty, and multicast.
  • HP high priority
  • LSS low speed statistical
  • HSD high speed deterministic
  • bursty and multicast.
  • each T1 transmitter/receiver supports the traffic classes via six queues and a service routine.
  • the service routine guarantees a minimum amount of bandwidth to each class of traffic under normal operation and allocates spare bandwidth according to a predefined priority scheme.
  • Network 20 of Fig. 1 includes nodes 22 in Atlanta, Boston, Chicago, Dallas, and Seattle. Each node 22 is connected to local communication equipment. For example, in Atlanta, a private branch exchange (PBX) telephone system 24, a fax machine 26, a computer 28, video equipment 30, local area networks (LANs) 32, high speed statistical equipment 34, a control terminal 36 and a printer 38 are all connected to the integrated cell exchanger 22. All communication equipment associated with a node 22 is connected through circuit T1 trunks 40. Each telecommunication node 22 takes existing voice and data streams from its associated communication equipment, assembles the streams into a more efficient cell format, and transmits the cells between nodes via cell T1 lines 42 (shown by broken lines).
  • PBX private branch exchange
  • LANs local area networks
  • each telecommunication node 22 receives cells from the cell T1 lines 42, disassembles the cells into data streams, and transmits those streams to the appropriate communication equipment via circuit T1 lines 40.
  • each integrated node 22 can function both as a source and a destination of information.
  • T1 refers to a telephone line operating at a synchronous data rate of 1.544 million bits per second. T1 lines are digital, and voice signals are digitized prior to their transmission over T1 lines. Under the DS-0 signaling standard, T1 line capacity is segmented into 24 individual channels each operating at 64 thousand bits per second to support voice transmission.
  • Each node 22 increases the apparent capacity of cell T1 lines 42 using virtual connections. In other words, rather than committing specific resources to a given source-destination pair, each node 22 connects a source-destination pair only when information, in the form of a cell, is present. When cells are not being created for a given source-destination pair, the same network resources are used to transmit cells for other source-destination pairs.
  • Cells are packets of fixed length and are typically used in systems that switch a number of traffic classes, as opposed to a single traffic class as in packet switching.
  • the short message length of cells as compared to packets permits cell switching networks to achieve very high rates of network utilization and still maintain short transmission delays.
  • the terms are used interchangeably herein.
  • a general cell format is illustrated in Fig. 2.
  • the cell includes 24 octets, or 8-bit bytes, of information.
  • the first two bytes, octet 0 and octet 1 represent the cell's virtual circuit identifier.
  • the virtual circuit identifier allows the cell to be self-routing through network 22.
  • Octet 2 the third byte, indicates the type of data contained within the cell namely, voice, video, etc.
  • Octet 2 also includes a cyclic redundancy check ("CRC") code, which is used by node 22 to perform an error check on the first four octets of the header.
  • Octet 3 the fourth byte of the cell, may contain a time stamp, congestion control bits, or hopcount depending upon the traffic class. The remaining octets of the cell are generally devoted to information or payload.
  • nodes 22 support six types of cell traffic. Identification of different traffic classes allows T1 transmitter/receivers to discriminate among classes of cell traffic for queuing and servicing purposes.
  • the classes of cell traffic supported by nodes 22 are high priority traffic, voice traffic, low speed statistical traffic.high-speed deterministic traffic, bursty traffic, and multicast traffic.
  • High priority cells are indicated by a binary 100 in octet 2.
  • High priority traffic includes inter-node cell exchanger-to-cell exchanger messages. This class of traffic is given the highest servicing priority so that network failures and overloads can be quickly remedied.
  • Voice traffic is represented by a binary 010 in octet 2.
  • Voice traffic includes pulse coded modulated (“PCM”) and adaptive differential pulse coded modulated (“ADPCM”) voice signals.
  • PCM pulse coded modulated
  • ADPCM adaptive differential pulse coded modulated
  • Low-speed statistical traffic includes cells generated by equipment operating at less than 64 thousand bits per second. Low speed statistical traffic can tolerate some delay between generation and transmission, but cannot tolerate any loss of data from a message. This class of traffic is represented by a binary 111 in octet 2.
  • Low speed statistical cells carry a timestamp in octet 3.
  • nodes 22 can determine how long low-speed statistical cells have been in network 20 and transmit the oldest cells first. Timestamps are also used to discard cells exceeding a maximum age, which is software-setable.
  • High speed statistical cells encompass high-speed, full-period voice or data signals. This class of traffic tolerates very little service delay and is equally intolerant of loss of data from a message.
  • a binary 110 in octet 2 indicates a cell of high speed statistical.
  • bursty and multicast classes of cell traffic supported by nodes 22 are quite different from the four other traffic types. Bursty and multicast traffic transport high-speed, statistical data from sources such as local area networks (LANs), bridges, routers, brouters and high-speed cell switches. Both types of traffic are characterized by burstiness. As used herein, the term "bursty data" refers to bursty point-to-point traffic between nodes 22. Multicast traffic carries bursty traffic from any one to all other nodes 22 in network 20. Both bursty traffic and multicast traffic are relatively delay insensitive. Thus, during periods of congestion, cells of these traffic classes may be buffered rather than discarded. Bursty traffic is represented in octet 2 by a binary 101 and multicast traffic is represented by a binary 011.
  • Bursty traffic and multicast traffic also differ from other traffic classes in that the octet 3 of each cell contains a hopcount. Hopcounts serve a purpose similar to timestamps, but are more practical than timestamps at high speeds.
  • a cell's hopcount indicates the number of nodes- 22 previously traversed by the cell in its journey between source and destination. The higher a cell's hopcount the greater the cell's service priority as compared to other cells within that class of traffic. In one embodiment, the maximum hopcount is 15; however this number is software selectable and may be modified to suit specific network applications.
  • the circuitry within the node 22 performs two basic functions: traffic segmentation and reassembly. While the node 22 is functioning as a source of messages, segmentation involves receiving incoming bit streams from the communication equipment and assembling it into cell formats. Trunk interface units buffer data in queues and route the cells through the network 20. Similarly, when node 22 functions as a destination for messages, the trunking interface receives cells and routes them to the appropriate circuity for reassembly. Reassembly is the process of converting cells into data streams or packet frames.
  • Node 22 includes a processor controller 50 , a data assembler/disassembler (“PAD”) group 52, a voice data assembler/disassembler (“PAD”) group 54, TXR 56, circuit/cell T1 interface 58 and FRP 59.
  • a system-wide system bus 60 which includes two buses, transports data and control information throughout node 22. Part of system bus 60, muxbus is a high speed statistical bus for switching cells. Also part of bus 60, control bus configures, manages, and monitors the components 50, 52, 54, 56, 58, and 59 of node 22.
  • -Processor controller 50 manages the various circuits within node 22, with direction from an operator using a control terminal and a printer. Processor controller 50 manages the routing of cells via network configuration databases. Processor controller 50 distributes control and configuration information via the control bus to the 50, 52, 54, 56 58, and 59 each of which is coupled to the control bus by an 80C31 controller. Additionally, network processor controllers 50 in Atlanta, Boston, Chicago, etc., cooperate to perform network-wide functions.
  • Processor controller 50 uses a 16-bit 68000 microprocessor supplied by Motorola Corporation of Schaumberg, Illinois, as its basic processing unit.
  • the 68000 microprocessor contains the software for controlling, diagnosing, and monitoring integrated cell exchanger 22. This software is stored in 2 Megabytes of flash EPROM and 8 Mbytes of DRAM.
  • the preferred embodiment of processor controller 50 also utilizes a bus controller sold by Intel Corporation of Santa Clara, California to manage the control bus.
  • a portion of the node's 22 segmentation and reassembly function is performed by voice/data processor 62.
  • Voice/data processor 62 assembles and disassembles cells of voice and data samples for cell T1 lines 42 using the DS-0 standard. This includes voice-band data generated by modems and FAX machines. Voice/data processor 62 is bidirectional.
  • Each voice connection within network 20 requires two voice/data processors 62, one at each end of the connection.
  • the source voice/data processor 62 determines when to generate cells.
  • Voice activity detection provides a 2:1 compression of PCM voice.
  • the source voice/data processor 62 extracts voice from a T1 port, packetizes it, and sends the packets to the remote voice/data processor 62.
  • the remote voice/data processor 62 reassembles the voice signal from the cell stream and passes it on to the appropriate external device, for example, PBX 24.
  • Voice compressor/decompressor 64 receives voice samples from voice/data processor 62. The voice samples are converted from PCM to compressed adaptive differential pulse code modulation (ADPCM) when the node 22 is acting as a source. Voice compressor/decompressor 64 converts ADPCM back to PCM when node 22 acts as a destination.
  • ADPCM compressed adaptive differential pulse code modulation
  • Synchronous data interface 66, synchronous data cell assembler/disassembler 68, and low-speed data cell assembler/disassembler 70 perform another portion of the segmentation and reassembly function within node 22.
  • Both cell assembler/disassemblers 68 and 70 commonly referred to as packet assemblers/dissemblers ("PADs") buffer, assemble and format data cells. The cells are transmitted to and received from system bus.
  • Synchronous data PAD 68 provides four different physical interfaces: RS232C/D (V.24), V.35, RS422/449, and X.21/V.1 .
  • Low speed data PAD 70 supports low speed data applications. Both low speed data PAD 70 and synchronous data PAD 68 can be configured for transparent, bit-for-bit transmission or for data compression using repetitive pattern suppression.
  • Frame relay PAD (FRP) 59 performs segmentation and reassembly of high-speed bursty data and multicast data.
  • Frame relay PAD 59 is typically connected to LAN devices, such as routers and bridges.
  • FRP 59 provides network cell rate control functions in cooperation with TXR 56.
  • TXR 56 contains routing tables for the cells to be transmitted on cell T1 lines 40 and buffers, or queues, for those cells. Each of the six traffic types supported by node 22 is queued separately, as will be discussed in detail below.
  • TXR 56 performs the T1 line framing and signal control, cell servicing, and alarm detection and generation.
  • Circuit/cell interface 58 provides interface support for bidirectional T1 lines 40 and 42. The conversion of bipolar signals from cell T1 lines 42 into logic-levels for use internal to node 22 is also performed by circuit/cell interface 58.
  • Data is transmitted across T1 network 20 over a virtual circuit between a pair of data PADs, 68, 70, or 59.
  • the flow of data into and out of node 22 is illustrated in Figure 4.
  • a data stream from local communication equipment is received either by synchronous data PAD 68, low speed data PAD 70, or frame relay PAD 59.
  • the source data PAD 68, 70, or 59 builds data cells at a rate proportional to the baud rate of the data.
  • the PAD 68, 70, or 59 passes the cells to system bus 60 for delivery to TXR 56.
  • Transmitter/receiver 56 places the cells in the appropriate queue according to traffic type where the cells await delivery to cell/circuit interface 58.
  • the cells are transmitted onto cell T1 lines 42 by cell/circuit interface 58. Cell traffic coming into node 22 follows the reverse path.
  • Figure 5 is an example of a four node network 20 for use in explaining the onset of congestion interval to network 20.
  • user D subsequently establishes a working bursty data connection through node B to user C2 at node 22-C involving FRP 59D1 , TXR 56D1 , TXR 56B3, TXR 56B2, TXR 56C1 , and FRP 59C2.
  • node 22-B could become congested at the common TXR 56B2 (shown shaded) if the combined bursty data rate exceeds the available capacity on trunk 42 connecting nodes 22-B and 22-C.
  • congestion could occur in the reverse direction at TXR 56B3 of node 22- B interfacing with cell trunk 42 connecting node 22-B and 22-D if the combined bursty traffic between user C2 and node 22-D plus that between, say, user B and node D exceeded the available bandwidth between nodes 22-B and 22-D.
  • Figure 6 shows the effects of congestion on effective network bandwidth and delay as a function of the offered load bandwidth.
  • bursty data tends to be high- bandwidth short-duration messages that, if uncontrolled, may either cause congestion or require that the network operate with a high percentage of average unused bandwidth.
  • the average offered load would operate at point A of Figure 6, where peak loads would not be likely to cause severe congestion.
  • the method and apparatus for congestion prevention control described below is applied to cell switching networks of the type described. These cell networks are designed to interface with user data ports operating at the CCITT asynchronous transfer mode ("ATM") adaptation layer of ISDN networks.
  • ATM asynchronous transfer mode
  • the cell network accepts frame relay data at its input nodes and produces frame relay data at its output nodes.
  • the entire operation of the cell switching network is intended to be completely transparent to the frame relay user.
  • the congestion control system is hereafter referred to as an internal congestion avoidance (“ICA”) method and apparatus, wherein "internal” means internal to the cell switching network.
  • ICA internal congestion avoidance
  • Figure 7 is an abstract representation of an internal congestion avoidance feedback control system.
  • the purpose of the ICA feedback control system is to anticipate the onset of congestion in the bursty data queues of the network nodes and correct the cause of the incipient congestion in a timely fashion. Corrective action is taken, on a per channel basis, by controlling the rate at which each bursty source node contributing to the congestion, accepts incoming user data.
  • node 22-A is the source node
  • node 22-B is an intermediate node
  • node 22-C is the destination node.
  • FRP 59 of node 22-A accepts user data from one of the input virtual data ports. It is assumed that the input data contains bursty framed data. FRP 59 disassembles the data frames into cells and regulates the rate at which they are delivered to TXR 56 for transmission over virtual connection 182 to intermediate node 22-B.
  • Other TXR 56 units (not shown) at node 22-B may each receive bursty data cells, each receiving TXR 56 directing the bursty data cells over the system bus 60 to the appropriate bursty data queue in a TXR 56 for transmission over one or more common trunks.
  • Each of the output TXR 56 units of nodes 22-A and 22-B performs traffic control and an ICA detection function by monitoring the length of its bursty data queue. If the queue length exceeds a prescribed threshold length, a congestion indicator bit (c-bit) is set in the header (octet 3, bit 5) of all bursty data cells transmitted from node B to node C while threshold is exceeded.
  • This same TXR 56 unit also performs a traffic control function including the discarding of cells when congestion conditions require.
  • a TXR 56 performs the receive functions previously described and directs the cell traffic to the appropriate queue of an output FRP 59 unit via system bus 60 where the cells are assembled into frames for delivery to the user C virtual port 40.
  • FRP 59 examines the headers of each bursty data virtual circuit and generates an estimate of the congestion status by comparing the average c-bit rate with a threshold reference value. If threshold is exceeded, indicating incipient congestion, an ICA rate control message is fed back to the source FRP 59 unit in source node 22-A, where user A's input rate is decreased so as to relieve the congestion present on its virtual circuit.
  • the FRP 59 unit of the source node 22-A also provides an additional traffic control feature by marking burst data cells that are delivered by user A at an excess rate.
  • a cell loss priority (“CLP") indicator bit is set in the cell header (octet 3, bit 4) and, like the c-bit, once set remains set throughout its journey to the destination FRP 59.
  • This CLP bit is used by the intermediate transmit TXR 56 of node 22-B for prioritizing cells for discard when required by congestion conditions. Cells marked by asserted CLP bits are discarded first. Of course, the discard of any one cell in a given data frame requires the retransmission of the entire frame of data.
  • a two threshold algorithm is used in order to discard cells that, with high probability, belong to the same frame.
  • each FRP 59 and TXR 56 has a distinct receiver function coupling node input data to system bus 60 and a distinct transmit function coupling data from system bus 60 to the node output.
  • the ICA rate control feedback path 180 is the return path of the existing virtual connection.
  • Intermediate node B is not necessary for implementing a virtual connection or for operation of the ICA system. Instead, the intermediate node B should be interpreted representing a possible multiplicity of concatenated intermediate nodes performing similar functions.
  • FIG 8 is a more detailed functional description of an FRP 59 unit.
  • Input user data arriving on user input 40 is accepted by traffic monitor 801 that compares data with the guaranteed minimum information rate ("MIR”), the committed burst duration (Be) and the quiescent information rate (“QIR”) time-out parameter, TQ. If MIR, for a given connection, is exceeded, a CLP bit is set in each frame exceeding the MIR. Frames are sorted into distinct virtual connection (“VC") queues 803, one per connection. Segmenter 805 operates on the data in VC queues 803, in an order prescribed by the serving protocol, and segments the VC frames into VC cells, which are sent to system bus 60 of node 22 by cell transmitter 807.
  • MIR guaranteed minimum information rate
  • Be committed burst duration
  • QIR quiescent information rate
  • TQ quiescent information rate
  • segmenter 805 sets a bit in the final cell of a frame if that frame was received with its DE bit set (indicating excess frame rate at the frame relay layer).
  • the rate of transmitting cells for each virtual connection by cell transmitter 807 is controlled by credit manager 809 operating cooperatively with rate controller 811 to provide incrementally variable transmission rates.
  • C a ⁇ is a configurable parameter for credit manager 809 that specifies the maximum number of cell transmission credits that may be accumulated by a given virtual connection, when the virtual connection is idle.
  • each VC cell rate initially sets each VC cell rate to a value equal to the configuration parameter QIR, representing the quiescent information rate.
  • QIR is bounded on the lower side by the guaranteed minimum information rate (MIR) and on the upper by the peak information rate parameter ("PIR").
  • MIR guaranteed minimum information rate
  • PIR peak information rate parameter
  • the rate setting is subsequently dynamically adjusted by rate commands received from the VC destination FRP 59 via cell receiver 815.
  • a two-bit coded message embedded in the cell header octet 3, bits 2 and 3
  • return traffic is used to encode a four-state rate message: increase, decrease, large decrease, and no-change.
  • controller 811 In response to the rate control message, controller 811 increases the rate by linear additive increments (typically 1/10 of MIR) and incrementally decreases the rate by a multiplicative factor of 7/8 or 1/2, the latter being used when cell loss in conjunction with congestion is detected by the destination FRP 59. Controller 811 also monitors the VC queue depth to see if sufficient data is present to warrant a data rate increase. No increase is made if sufficient data is not present. Traffic monitor 801 also provides controller 811 with a reset signal that causes the rate to be reset to QIR when no data has been presented to traffic monitor 801 for a period equal to configuration parameter, TQ.
  • Credit manager 809 performs the function of controlling the intervals at which each virtual connection is permitted by cell transmitter 807 to transmit a cell to system bus 60.
  • the receive portion of FRP 59 comprises cell receiver 815, reassembler 817, DE filter 821 , port transmit queues 819 and traffic analyzer 813.
  • Receiver 815 interfaces with the local system bus 60 from which-it receives its virtual circuit data from the connection's other terminal node.
  • the congestion indicators (c-bits) and loss data information is provided to traffic analyzer 813.
  • Cell data is passed on to reassembler 817 that reconstructs frames from cells and sets the DE bit if the last cell of the frame indicates that the DE bit was set in the frame header when received from the sending user.
  • DE filter 821 monitors the port transmit queue depths and compares them against a threshold, DEth, typically set at approximately 75% of total queue depth. If threshold is exceeded, frames with the DE bit asserted are discarded first.
  • Traffic analyzer 813 analyzes the average c-bit rate over the measured round trip delay ("RTD") provided by PCC 50 of the local node 22 of Figure 3. If the average count of c-bits exceeds a prescribed threshold, TAth, (typically 1/2 of total possible count), a moderate rate decrease message is generated and supplied to cell transmitter 807 where it is embedded in a cell header and transmitted to the other terminal FRP 59. If cell loss is also present, a large rate decrease message is sent. If threshold is not exceeded a rate increase message is generated. A no-change rate message is sent with all outgoing cells for an interval of two times RTD, so that the effects of the previous rate change message may be observed before introducing additional changes. This delay in adjustments is necessary for stable congestion control.
  • RTD round trip delay
  • Traffic analyzer 813 also monitors port transmit queues 819. If the output port transmit queue used by a particular VC shows incipient congestion, any rate increase message generated by traffic analyzer 813, based on the VC average c-bit count not exceeding the TAth threshold, is changed to a rate decrease message even though the c-bit count threshold, TAth, is not exceeded.
  • output port incipient congestion is determined by comparing the average queue depth of port transmit queues 819 against a configurable threshold, PQth.
  • the PQth threshold is normally set at four kilobytes, a fraction of the 64 kilobyte queue capacity. Average queue depth is calculated by traffic analyzer 813 using a regeneration cycle averaging period.
  • a regeneration cycle is defined as a busy plus idle interval of burst data. This average may be used for the entire duration of the next cycle. Alternatively, when the regeneration cycle is very long, the averaging may include the previous regeneration cycle count as well as the current count, averaged over the previous regeneration cycle time plus the current portion of the subsequent regeneration period.
  • test generator 861 , echo unit 863 and RTD measure unit 865 comprise the means for measuring the RTD of a virtual connection.
  • PCC 50 of local node 22 initiates an RTD measurement test by instructing test generator 861 , via system bus 60, to generate a high priority RTD cell for transmission by cell transmitter 807.
  • the RTD cell is received at the other terminal node by its FRP 59 cell receiver 815, it is directed to RTD echo unit 863 and retransmitted as a high priority RTD cell via its associated cell transmitter 807.
  • the initiating node 22 cell receiver directs the cell to RTD measure unit 865 that measures the elapsed time between the time of original RTD cell transmission and the receipt of its echo.
  • RTD measure unit 865 provides this data to PCC 50 via system bus 60 which in turn supplies this information as an RTD configuration parameter to traffic analyzer 813.
  • FIG. 9 shows a more detailed functional block diagram of the transmit portion of TXR 56.
  • This part of TXR 56 provides the traffic control and congestion detection in the preferred embodiment.
  • TXR 56 interfaces with system bus 60 of its local node 22 from which it receives cells destined to be transmitted over a common trunk 42.
  • Bursty cells arriving at the input of TXR 56 are sorted into queues 135 according to traffic class. Bursty data cells are queued in BD queue 855 where the average depth is monitored for indications of incipient congestion by congestion detector 857 by comparing the average queue depth with configurable threshold parameter, ECNth. If threshold is exceeded, the next bursty cell processed by cell server 853 will have a c- bit set in its header. Threshold ECNth is selected so as to introduce a tolerable amount of delay due to queue length while minimizing the amount of cells discarded. A value of 30 cells has been found acceptable.
  • the average BD queue 855 queue depth is also based on a regeneration cycle averaging period. Also, the BD queue 855 depth is used to control the discard of cells by CLP filter 851 which discards arriving bursty cells with the CLP bit set if threshold CLPthl is exceeded. Discarding of CLP labelled cells continues as long as a second threshold, CLPth2, is exceeded (CLPth2 > CLPthl ). In this manner, hysterisis is provided in the control mechanism so that it tends to discard the following cells belonging to the ⁇ ame frame, as shown in Figure 10. This strategy recognizes that once a cell belonging to a given frame is discarded, the integrity of the entire frame is violated, requiring retransmission of the entire frame.
  • Thresholds CLPthl and CLPth2 are configurable parameters supplied by PCC 50 of Figure 3 and typically set a approximately 75% and 25% of queue capacity, respectively.
  • FIG 11 shows the architecture of a FRP 59 and of a frame relay interface (FRI) unit 300 used to interface FRP 59 to four V.35 data ports, in one preferred embodiment.
  • DMA controller 301 of FRI 300 is a dedicated Motorola Corporation 56001 digital signal processor ("DSP") that interfaces between global frame RAM 230 and two-way serial communication controller (“SCC”) 303.
  • DSP digital signal processor
  • SCC two-way serial communication controller
  • V.35 ports are provided, each capable of operating full duplex at bit rates up to 2.048 Mbits/sec.
  • DMA 301 receives frames from the ports via SCC 303 and stores them in frame buffers 237 of global frame RAM 230. Also, DMA 301 transmits frames from frame buffers 237 to the data ports through SCC 303.
  • DMA 301 handles start and end of frame interrupts, extracts the cyclical redundancy check ("CRC") indicator from SCC 303 and appends it to the frame, and interfaces with frame transmit controller 202 and frame receive controller 201 in ACP 200 through queues in global frame RAM 230.
  • CRC cyclical redundancy check
  • DMA 301 takes pointers to frame buffers 237 from port free buffer queue 234 for use in storing incoming frame data. Because all ports may operate simultaneously, DMA 301 maintains a separate control block for each port.
  • DMA 301 notifies frame receive controller 201 through port frame receive queue 231.
  • DMA 301 polls frame transmit queues 232 (one per port).
  • DMA 301 maintains the following information for each transmitted or received frame: pointer to the start of frame in global frame RAM 230, pointer to the current work location in RAM 230, and current length of the frame.
  • FRP 59 comprises four major units: administration/communication processor (ACP) 200, multiported global frame RAM 230, cell receiver 210, and cell transmitter 220.
  • ACP administration/communication processor
  • processors are included on the FRP 59 board: an Intel Corporation 8751 microcomputer acting as a control bus kernel, interfacing FRP 59 to the control bus of system bus 60; a Motorola Corporation 68020 microprocessor that serves an administrative and main communication control (ACP 200) functions; two Motorola Corporation 56001 digital signal processors, one dedicated to cell transmitter 220 and the other to cell receiver 210.
  • Figure 5 shows two paths between the 68020 microprocessor of ACP 200 and the 56001 microprocessors in receiver 210, transmitter 220, and DMA 301.
  • the solid lines represent the typical flow of data and control information between the microprocessors and global frame RAM 230. Normal frame relay communications is done through these data paths.
  • the dashed lines represent a secondary control path between ACP 200 and the host port of each 56001 microprocessor for handling administrative functions such as diagnostics and statistics collection.
  • Control bus command server 203 of ACP 200 responds to commands received from processor controller 50 ( Figure 3) over the control bus 95 portion of system bus 60, sending responses to event server 204 for forwarding to processor controller 50 via a FIFO queue to the control bus kernel microcomputer, or for forwarding commands to the communication units, i.e.; frame receive controller 201 and frame transmit controller 202.
  • Controllers 201 and 202 perform real-time control functions.
  • Cell transmitter 220 maps frames received from FRI 300 data ports to cell network connections, applies the ICA control scheme, fragments frames into cells, and transmits cells on muxbus 96 toward the destination node.
  • Cell receiver 210 coordinates receipt of frame fragments (cells) from muxbus 96, reassembles cells into frames.
  • Frame transmit controller 202 transmits those frames to the proper user port, linking cell receiver 210 to DMA 301 through RAM 230, where reassembly takes place. Transmit controller 202 maintains muxbus free buffer queue 233 for the use of receiver 210 to determine the next buffer to use to store reassembled frame data.
  • LMI 205 is an optional local management interface for verifying the integrity of the V.35 physical link and for providing status from the cell network to the user device.
  • These ACP functions (receive controller 201 , transmit controller 202, command server 203, event server 204 and LMI 205) are tasks executed under the operating system running on the 68020 microprocessor.
  • Cell receiver 210 and transmitter 220 operate as dedicated controllers. Communications between them and communication controller tasks 201 , 202 and DMA 301 is primarily through queues in global frame RAM 230. Using these queues simplifies the timing requirements on the communication control tasks, allowing them to operate less frequently on larger groups of requests.
  • Queue 236 points to the location of the cell data in frame buffers 237 of RAM 230 together with the number of octets to send, the virtual circuit identifier for the cell, the initial timestamp value to be placed in the cell header, and an indication of whether the cell is a start-, middle-, or end-of-frame fragment.
  • the cell is sent to muxbus 96 via the muxbus transmit FIFO of cell transmitter 220.
  • Cell transmitter 220 controls the rate of transmission of cells onto muxbus 96 by means of a credit manager (similar to that described for servicing cell traffic by TXR 56). After the start of frame indication, queueing of cells for transmission on muxbus 96 does not occur until one of the following occurs:
  • Cell transmit queue 236 is polled successively to see if one of these two conditions (ensuring sufficient data to transmit a cell) prevail. If so, the control is subjected to the ICA credit manager control scheme to be described later.
  • FIG. 12 illustrates T1 transmitter/receiver 56 that implements the circuitry and method for queuing and servicing cell traffic.
  • T1 TXR 56 is divided into two sections, a transmitter 72 and a receiver 70.
  • Receive state control 74 coordinates the operation of T1 receiver 70.
  • Incoming data from another node 22 is received as a T1 bit stream 80 and an extracted clock 82 sent by circuit/cell interface 58.
  • T1 deframer 84 strips the T1 framing pattern from the incoming bit stream and sends the received octets of the cell through descrambler 86 and CRC checker 88. If the CRC for the cell is in error the cell is destroyed. If the CRC is correct, the cell is placed into receive packet FIFO 90.
  • muxbus output control 92 When muxbus output control 92 is notified by muxbus address and control signals 94 that it may dispatch a cell onto muxbus 96, the first cell stored in receive packet FIFO 90 is driven onto muxbus data bus 98 through output register 100. If there are no cells in receive packet FIFO 90, a cell consisting of "all ones" is written to muxbus data bus 98.
  • Queue manager 76 includes a high-speed microprocessor 110, RAM 112 and program ROM 114.
  • the fair queueing and service routines are implemented in software, which runs on microprocessor 110. The fair queuing and servicing routines will be described below, following a general description of the structure and operation of transmitter 72.
  • the arrival of a cell via muxbus 96 to transmitter 72 is accompanied by a combination of asserted address and control signals 116 to input state control 118.
  • the first two octets of the cell are received into the pipeline registers 120 from muxbus data bus 96.
  • input state control 118 performs a series of functions. First, a new packet start address is fetched from packet start address FIFO 124 which is passed together with the output of counter 126, through mux 128. This forms the address for queue memory 135. As each successive octet of the cell exits pipeline registers 120 onto queue memory bus 140, each byte is written into a location in queue memory 135 pointed to by the packet start address and indexed by counter 126, which increments with each octet.
  • a packet information block is created for the cell.
  • Octet 2 and 3 of the packet and the current time which is used as a queue input time, are written into packet information FIFO 152. This is done through mux 154, and forms the packet information block.
  • queue manager 76 is notified that a cell has just been queued, where that cell is stored, the traffic type, its timestamp value or hopcount, if applicable, and when the cell arrived.
  • Current time counter 158 is incremented by the 125 microsecond clock 160 from muxbus 96. Current time counter 158 may be interrogated by queue manager 76.
  • queue manager 76 When queue manager 76 decides to dispatch a cell, it writes the packet start address and updated timestamp or updated hopcount, if applicable, to output state control 144.
  • Output state control 144 fetches each byte of the cell to be dispatched from queue memory 130 by applying the appropriate addresses to queue memory 135 through mux 128. Each octet of the cell is moved via queue memory bus 140 through MUX 146 to CRC generator 142. Queue memory 135 is time-shared between input state control 118 and output state control 144 .
  • an updated timestamp previously written to output state control 144 by queue manager 76, will replace the fourth octet of the cell. This is achieved by activating mux 146.
  • CRC generator 142 calculates and then writes a CRC-5 code into the lower bit field of the traffic type/CRC octet. The CRC covers the first four octets of the cell.
  • the cell hopcount will be updated prior to dispatch by queue manager 76. Again, the cell CRC will be recalculated by CRC generator 142.
  • queue manager 76 performs two major functions: queueing cells and servicing cells. During queueing and servicing, queue manager 76 treats the various classes of cell traffic differently.
  • queue manager 76 fetches packet information blocks from packet information FIFO 152.
  • Packet information blocks contain a representation of the cells that have been stored in queue memory 130.
  • the traffic type field of the packet information block is interrogated to determine into which queue the packet information block should be appended.
  • the queues are linked lists of packet information blocks.
  • Figure 13 illustrates the queue structure of the bursty and multicast traffic queues, which utilize hopcounts.
  • bursty traffic and multicast traffic queues are nearly identical. Thus, while the description that follows refers to the bursty traffic queue, it will be understood that the following description applies equally to the multicast traffic queue. Queuing of bursty traffic differs from the other queuing schemes. The difference arises, in part, because these traffic queues include a number of subqueues as shown, for example, in Figure 7 and numbered 1 through 15. Each of the subqueues is a FIFO queue including a set of linked-lists. As each cell of bursty traffic is received it is placed into one of the subqueues according to the hopcount stored in octet 3 of the cell.
  • a cell of bursty traffic with a hopcount of 13 will be placed at the end of subqueue 13 of the bursty data queue.
  • Cells of bursty traffic having hopcounts greater than 15 are simply discarded, as they have used more network resource (delay) than permitted.
  • the maximum hopcount is 3 for the bursty queue for one embodiment due to limited coding bits in octet 3 of the cell header. This maximum could, however, be increased to allow cells of bursty data traffic to use more network resource by increasing the hop-count coding capacity. Because of this coding bit limitation, cells are not discarded when the count exceeds 3 but continue to carry a maximum hop-count of 3.
  • Cells within the bursty traffic queue are serviced according to subqueue priority order, with subqueue 15 having the highest priority and subqueue 1 having the lowest priority.
  • a subqueue pointer indicates the highest ranked non-empty subqueue. The indicated subqueue will be serviced first whenever bursty data traffic is serviced. The subqueue pointer is updated whenever cells are placed into the bursty data queue or the queue is serviced.
  • servicing the bursty traffic queue includes taking the cell designated by the subqueue pointer and updating both the linked lists and queue depths. Additionally, during the servicing of the bursty data queue the hopcount of the serviced cell is incremented by 1. If the serviced cell was the last cell in the subqueue, the subqueue pointer is updated to indicate the highest priority non-empty subqueue.
  • the result of the priority scheme used in the bursty traffic queue is that priority is given to cells which have used more network resource (delay). Possible undesirable effects of this priority scheme are freezing of lower priority subqueues during periods of congestion and lengthy queuing delays. These effects are avoided by an aging mechanism.
  • the aging mechanism allows overaged cells to be discarded, while cells that have spent too long in one subqueue are moved to the next subqueue with a higher priority. A cell's queuing priority is effectively increased by moving to another subqueue, but the hopcount is unaffected.
  • Table 1 is an example that details the mapping of service order, j, and spare bandwidth priorities, k, for each class of traffic, i, in the preferred embodiment. Note that the service priority is according to assigned minimum bandwidth.
  • the servicing routine uses a credit accrual scheme to ensure that each class of traffic receives a selectable minimum bandwidth.
  • N denote the total available bandwidth on a cell T1 trunk and let T denote the queue server tick interval.
  • the unit of N is not relevant; it can be specified as a number of cells per second, or any other throughput unit.
  • N 8000 cells per second.
  • T can be given in any convenient unit of time.
  • the tick interval T equals 125 microseconds.
  • the product N * T represents the capacity of the cell trunk per tick interval, or the quantum of bandwidth.
  • Each class of traffic is assigned a minimum amount of the quantum of bandwidth, with the exception of high priority traffic. This is because all high priority traffic will be serviced regardless of the required bandwidth.
  • the sum of the minimum class bandwidths must be less than N to allow some bandwidth for high priority traffic. In other words, if i represents the class number, and Nj represents the minimum bandwidth assigned to the i th traffic class, then N* ⁇ + N ⁇ + N3 + N 4 + N5 ⁇ N.
  • Each minimum bandwidth Nj can be transformed into a timer value, Df, representative of the number of tick intervals T that must elapse for traffic class i to acquire its quantum of bandwidth.
  • a credit accrual routine runs simultaneously with the service routine.
  • Each class of traffic i is assigned a timer Tj, which is initialized to the associated timer value, Dj.
  • the timer Tj is decremented every T units of time.
  • a transmission credit Cj accrues for traffic class i.
  • Nj and Dj the greater the allocated minimum bandwidth for a class of traffic, the faster the rate at which it acquires transmission credit.
  • timer Tj is updated by adding Dj to the previous value of Tj. Using this method of accrual, each class of traffic i accrues Nj credit in a tick interval of T.
  • the maximum number of credits, Cmaxj, that may be accrued for each class of traffic i is selectable.
  • the maximum credit that may accrue to any traffic class is 1.
  • Figure 14(a) is a flow diagram of the service routine for a single tick interval implemented by queue manager 76.
  • queue manager 76 Using a credit based strategy for servicing cell traffic, queue manager 76 guarantees each class of traffic a minimum bandwidth.
  • index B which indicates the number of spare bandwidth credits available is incremented in step 416.
  • Step 418 checks to see if the priority order index, j, has been exhausted, and if not, returns to step 406 where index j is incremented. If all values of j have been exhausted, step 420 checks to see if B > 0, indicating that spare bandwidth is available for distribution in accordance with protocol 800 referenced in step 422. Otherwise, the process terminates.
  • the traffic class index, i is set equal to the value of k.
  • the described method of allocating spare bandwidth between various traffic classes by TXR 56 is an open-loop control system because the data rate is controlled by the sending node without any feedback from the cell switching network. This procedure leads to a conservative allocation of network resources because each terminal network node acts independently and without specific knowledge of the traffic state of the network. In order to achieve higher bandwidth utilization by bursty traffic, without undo congestion on a given virtual connection, it is necessary to provided the ICA feedback information about the level of bursty traffic being handled by all FRPs involved with a given virtual connection.
  • ICA is configurable on a per connection basis.
  • the configurable MIR and PIR guarantee that each connection gets at least its minimum allocated bandwidth, independent of other traffic.
  • System software resident in each process controller 50 implements the user interface and ICA node functions, and further comprises the following functions:
  • MIR minimum information rate expressed in Kbps which is translated into cells per second for internal node use
  • PIR peak information rate expressed in Kbps (internal cells per second);
  • TQ quiescent information rate (QIR) time-out parameter
  • VCqd virtual connection maximum queue depth in bytes
  • ECNth virtual connection (VC) queue threshold in bytes for explicit congestion notification (ECN);
  • CLPthl ,2 cell loss priority thresholds in bytes (CLPthl > CLPth2) to control loss of newly arriving cells with CLP bit set at 1 (indicating input rates in excess of MIR) are configurable for each TXR;
  • TAth traffic analyzer average c-bit count threshold
  • PQth output port transmit queue threshold
  • VCqd, PIR or, (CIR, Be, PIC), or (CIR, VCqd, Be) may be used interchangeably as optional parameter sets.)
  • the two identical TXR queues previously labelled as bursty and multicast are redefined to accommodate the ICA feature. All ICA bursty traffic uses the bursty data queue, while non-ICA bursty data and multicast traffic share the same multicast queue. In this manner, both fCA non-ICA bursty data can be accommodated in the same cell switching network.
  • TXR 56 further comprises the following firmware functions:
  • FRP 59 firmware provides augmented functions in support of ICA operation comprising:
  • Additional FRP functions operating in conjunction with the extemal frame relay network comprise:
  • each frame transmit port queue 232 with a DE threshold (DEth) so that when reached, incoming frames from muxbus 60 with DE-bit set are discarded.
  • DEth DE threshold
  • ICA control is based on the detection of incipient congestion at each node traversed by the virtual connection. Incipient congestion obtains whenever the average bursty data queue length in TXR 56 is greater than congestion threshold parameter ECNlti. The average queue length is based on queue length measurements taken over a regeneration cycle consisting of a busy plus idle interval of the bursty data. If ECNth is exceeded, a c-bit is set in the cell header of the next cell to be transmitted by queue manager 76.
  • Destination ACP 200 of FRP 59 counts (averages) the c-bits over a RTD interval as shown in Figure 15. At the end of the averaging period, a rate adjustment message may be sent by the destination FRP 59 through cell transmitter 220 onto muxbus 96 and thence through TXR 56 and the virtual connection to the source FRP 59.
  • the source FRP 59 may adjust the rate once per two RTD delay units of time. Any change in congestion condition on the virtual connection due to a rate adjustment is detected in the average value measurement starting one RTD unit later.
  • FIG 16 is a flow diagram that describes the per virtual connection rate (bandwidth) change process 500 by which data rate changes imposed on cell transmitter 220 are adjusted by frame receive controller 201.
  • the it ⁇ * bursty channel bandwidth, Ni is initialized by setting the transmission rate to the quiescent rate, QIRj.
  • Step 504 checks to see if the queue has been inactive for a period of time greater than TQ, a prescribed configuration parameter. But because the channel has just been activated, the process passes on to step 507. Otherwise, it would go to step 506 where the quiescent rate, QIRjfis assigned. If no rate change has been received, test step 507 moves the process back to step 504, forming a wait loop. If a rate change moves the process to step 508 where it is determined if it is a rate increase or decrease.
  • step 516 determines if it is a fast decrease due to cell loss at the other terminal FRP 59. If so, the process moves to step 518 where the rate is reduced by a factor of 1/2, otherwise to step 520 where it is reduced by a moderate factor of 7/8. After either reduction, the process moves to step 522 where it is determined if the reduction resulted in a rate less than the guaranteed minimum, MIRj. If so, the rate is set at MIRj, or otherwise left unchanged and the process moves back to step 504.
  • step 507 If the rate change in step 507 is an increase, the process moves to step 509 where it is determined if data is available, as indicated the boolean flag Qj. If not, no rate increase is made and the process goes back to step 504. If data is available, step 510 increases Nj by the linear increment of 1/10 of MIRj and then checks (step 512) if the rate exceeds the maximum allowable, PIRj. If so, the rate is set at PIRj, in step 514. In either case, the process returns to step 504.
  • the credit manager function resident in cell transmitter 220 determines cell receiver 210's per channel output rate by assigning credits in accordance with the state of congestion and data availability. In the four V.35 port embodiment, each channel is serviced round-robin. The relative priority given to each (up to 252) virtual connections is determined by the bandwidth assignment algorithm in conjunction with the credit manager servicing algorithm as shown in the flow diagram of Figure 17.
  • Test step 606 checks to see if the resulting value of Ti is equal to or less than zero, indicating that a credit should not be added to the ilti credit index, Ci, and if so, passes on to step 616, where the round-robin procedure is initiated by incrementing index i. Otherwise, step 608 is invoked by crediting (incrementing) Ci.
  • Step 610 checks if the incremented value of Ci exceeds the upper limit, Cmax, and if so, moves to step 612 where Ci is set equal to Cmax. The process moves on to step 614. Step 614 restores a positive non-zero value to Ti by adding Di to the value obtained from step 604.
  • Step 616 leads to test step 618 to see if all of the VC connections have been served, and if so, a wait of one tick period, T, is imposed by step 620 • before repeating the process by returning to step 602. Otherwise, the process returns to step 604.
  • Step 704 decrements the credit count, Ci, indicating that the queue has been serviced.
  • Step 706 movements the VC index and passes to test step 708 that checks if all VC connections have been attended and if so, passes back to the initial 700 steps. Otherwise, the remaining VC connections are attended by returning to step 702.

Abstract

A feedback control system for congestion prevention in a cell (packet) switching communication network is described. Congestion control is accomplished by controlling the transmission rate of bursty traffic in the presence of high priority, voice, low speed statistical, high speed deterministic and multicast data. Because bursty traffic is relatively insensitive to delay, adequate buffercapacity can be provided at the network nodes in order to minimize bursty data cell loss. By monitoring the buffer queue lengths at the nodes, a control signal can be generated at each intermediate node indicating the state of congestion. Excess queue length indicates incipient congestion while short queue lengths indicate excess capacity. Queue status is forwarded to the destination node where it is interpreted and sent back to the source node as a feedback rate control signal using a 2-bit code. The source node regulates the rate of bursty data transmission over the cell network in accordance with the feedback control signal thus minimizing congestion and concomitant data loss while efficiently utilizing available network bandwidth.

Description

CONGESTION CONTROL FOR CELL NETWORKS FIELD OF INVENTION
The present invention relates to the field of cell switching network communications. More specifically, the present invention relates to apparatus and method of traffic congestion prevention and control within the cell switching network. BACKGROUND OF THE INVENTION
The desire to integrate data, voice, image, and video over high speed digital trunks has led to the development of a packet switching technique called cell relay or asynchronous transfer mode.
A typical fully-integrated voice and data network using digital trunk lines (e.g., T1 , FT1 , E1 , T3, etc.) includes a number of geographically distant interconnect nodes. Each node acts as a cell exchanger for receiving and forwarding cell information to its destination. By the use of a variety of interface cards, each node is capable of interfacing to user generated voice and data streams, then segmenting and assembling the streams into a more efficient cell format for transmission over a closed network using digital lines. Each node is also capable of receiving data from other network nodes and forwarding that data through to other network nodes to its ultimate destination. All terminal nodes also provide the necessary interface cards required to reassemble the data contained in the cells into a standard user data stream format.
A typical modern node is capable of handling six classes of cell traffic, each class having different characteristics and different service requirements. The six classes of traffic include: (1) High priority ("HP") for node-to-node control messages;
(2) Voice for pulse code or adaptive pulse code voice signals;
(3) Low speed statistical ("LSS") traffic sent at rates of less than 64 Kbps which includes statistical and deterministic full-period traffic;
(4) High speed deterministic ("HSD") traffic for full period voice or data at data rates exceeding 64 Kbps;
(5) Bursty data for point-to-point traffic generated by such sources as local area networks, bridges, routers or high speed packet switches within the cell network; and
(6) Multicast traffic which is of the same type as bursty data except that it is simultaneously broadcast from one source to many destinations (vs. point-to-point).
Each of these six traffic types are buffered at each network node in accordance with their particular sensitivities to network delay and cell loss. Cell loss may occur due to intermittent short term overload of network bandwidth and lack of adequate buffer capacity. For example, voice traffic is relatively delay sensitive and insensitive to occasional cell loss. In contrast, data traffic, such as file transfers, is relatively insensitive to delay but is data loss sensitive. High priority data is both delay and loss sensitive. To accommodate these differences, each class of traffic is typically placed in a preassigned queue, each with a different service priority. During periods of network traffic congestion, when network traffic demand exceeds the network's bandwidth capacity, servicing algorithms are typically employed to discriminate between traffic classes in order to allocate bandwidth. Delay is managed by properly sizing the queue depths and prioritizing transmission within a class based upon a measure of the time that a cell has been in the network as, for example, by use of time stamps and hop counts.
Even with these sophisticated queueing and service algorithms, congestion (due to excess arriving traffic) can occur. This congestion is typically divided into three categories: short-term, medium-term, and long-term. Short-term congestion, typically handled by discarding traffic at the queue, may be done haphazardly or preferably selectively by having cells marked with their "discard eligibility". Long-term congestion is controlled by admission policies that allocate resources (bandwidth and buffers) at the time a connection is established. If no resources are available, the connection is not allowed.
Medium-term congestion control has been an active topic of research during the last couple of years. Two types of techniques that have been researched are (1) open-loop control techniques involving no explicit feedback from the network in which congestion is controlled by smoothing the flow of input traffic and (2) closed-loop techniques that sense the level of congestion on the virtual circuit and control the in-flow of traffic based on feedback of congesting status information to the source terminal.
An example of a general rate regulation scheme for a bursty data source on a per virtual connection basis is described in a paper by K. Bala, et al., entitled, Congestion Control for High Speed Packet Switched Networks, published in the proceedings of the IEEE INFOCOM, June 5-7, 1990, pages 520-526. At the initial establishment of a virtual connection, a minimum amount of guaranteed bandwidth is allocated. The simplest system described uses the concept of a "leaky bucket" input rate controller that uses "tokens" and "spacers" to control the average data rate introduced into the packet switched network. Tokens arrive at the controller at a fixed rate. Each token corresponds to fixed number of bytes. The controller buffers the packet until enough tokens are collected for transmitting the entire packet. The token bucket has a fixed maximum capacity corresponding to the maximum packet burst duration. Tokens arriving to a full bucket are dropped. Thus, the system can handle different length packets which are transmitted without fragmentation. Peak rate control is accomplished by means of a spacer that introduces a suitable delay proportional to the length of the prior transmitted packet.
-A given session on a virtual connection may last for long periods of time (up to hours). Bursty data sources are characterized by intermittent high data rate burst with significant spans of inactivity between bursts. Under these circumstances, the above described simplest system would result in underutilization of the bandwidth capacity of the system because of the prescribed safe bandwidth limit assigned to the virtual connection session.
Average bandwidth utilization efficiency is typically improved by introducing "colored" tokens, for example, green and red. Green tokens correspond to packets received for transmission that fall within the minimum guaranteed bandwidth protocol while the red tokens correspond to packet data received for transmission in excess of the guaranteed minimum rate. Intermediate nodes provide per trunk FIFO buffer service and use the colors associated with each packet for congestion control. In general, green packets are protected and passed along while red packets are discarded upon arrival whenever the chosen metric (usually queue lengths) for congestion threshold is exceeded. Even though discarding of packets implies retransmission of the lost packet data, the system is represented as improving the average utilization of bandwidth capacity.
An example of prior art feedback control is a window control method, described by K.K. Ramakrishnan and Raj Jain in an article entitled A Binary Feedback Scheme for Congestion Avoidance in Computer Networks. ACM Transactions on Computer Systems, Vol. 8, pages 158-181 , May 1990. The window control method indirectly controls the effective network data rate by having the user adjust the window duration controlling the number of contiguous packets that can be transmitted at a given time.
For ISO Transport network architectures, each network layer packet has a header that is used for "housekeeping" functions, including routing. Within that header, a bit is assigned as a congestion indication bit. The packet source clears the congestion indication bit when a packet is originally transmitted from a source node end system. Each network router is equipped with congestion detection means that monitors the average queue length at the router at the time the packet arrives.
The average queue length is determined by the number of packets that are queued and in service, averaged over an interval. This interval corresponds to the duration of the sum of the last busy plus idle cycle duration plus the busy interval of the current cycle. When certain average queue length threshold conditions are met at any router, the congestion indication bit is set and forwarded with its associated packet. Once set, the congestion bit remains set even though it subsequently passes through routers with queue lengths not exceeding threshold.
At the destination end system (user transport entity) an acknowledgement message is generated if no error is detected and is sent together with the congestion indication bit status associated with the acknowledged packet. The user transport entity updates the window size based on the number of packet congestion bits set during the last transmission window used. If at least 50% of the bits are set, the window size is reduced from its current value. Otherwise, it is increased. In order to avoid instability in this feedback control system, updating of the user's transparent entity window size occurs only after a time period corresponding to the duration of the last two windows. This delay allows enough time to see the effects of the last window change. In order to promote "fairness7 of access by all network users and to achieve stable operation, the window size adjustment algorithm provides for increasing window size by linear increments and for decreasing window size by a multiplicative factor.
In summary, Ramakrishnan and Jain describe a system using window control at the ISO Transport layer using window duration control rather than rate control. Rate control is indirectly controlled by the limiting actions of acknowledgements and window length. Because transmission rate is a direct measure of bandwidth, better short term control of this system resource can be obtained by direct rate control.
CA. Cooper and K.I. Park in an article entitled Toward a Broadband Congestion Control Strategy. IEEE Network Magazine, May 1990, pp. 18-23, discuss a congestion control system for variable bit rate traffic using an ATM based fixed length cell switching network which may have elements of both open-loop and closed-loop control. The authors describe a system carrying different classes of service in which a cell rate is assigned to variable bit rate data that is less than the available peak cell rate so as to provide capacity for statistical multiplexing of data. A new connection is admitted only if sufficient network resources are available to support that connection. Cells offered to the network in excess of the assigned rate are immediately discarded because, in the opinion of the authors, the tagging of floating cells for possible subsequent discard appears to yield no substantial advantage. Traffic enforcement is applied on the ingress direction of each virtual circuit. A leaky bucket or credit manager algorithm is used to control the rate.
The Cooper article refers to a reactive (feedback) control method that combines cell tagging with explicit congestion notification ("ECN"). When an asynchronous transfer mode ("ATM") switch detects congestion on a given route, it sends an ECN message over the variable rate virtual circuits on that route to the serving ATM switches which may, in turn, relay these ECN messages to the involved customer premises node. The serving ATM switches operate at one of two preset thresholds for rate enforcement on each variable bit rate virtual circuit using ECN. The lower threshold is used during congestion. Additionally, the ECN message may be transported between network nodes possibly using a separate virtual circuit, and possibly using a bit in the ATM cell header. SUMMARY AND OBJECTS OF THE INVENTION
One object of the present invention is to optimize the use of available system bandwidth.
Another object of the present invention is the optimization of network resource allocation to bursty traffic when sharing a network digital trunk with the five other traffic classes.
Another object of the present invention is to provide a dynamic bandwidth (or data rate) allocation scheme that allows individual users to use unused network capacity for increasing throughput when necessary to accommodate the peak loads of individual users.
Another object is to provide a method used for allocation that also helps to ensure "fairness" in network resource availability to all users by using linear incremental increases and multiplicative decreases in data rates together with guaranteed minimum rate allocations.
Another object is to provide a cell switching network operating at the International Standards Organization ("ISO") Data Link Layer that direct rate control based on virtual circuit connection congestion.
A further object of the invention is to provide for the fast recovery from congestion.
Another object is to provide congestion relief by tagging cells at the source node when the virtual circuit's guaranteed minimum information rate ("MIR") is exceeded so that if a prescribed cell loss priority ("CLP") of queue depth threshold is exceeded at any intermediate node, tagged cells using that connection may be dropped.
Another object is to provide an adaptive interval between sending rate updates based on the actual measured round trip of the connection. Another object is to provide each node with means for measuring virtual connection round trip transmission time for use in adaptive interval control.
Another object is to provide means for generating supervisory cells for transmission of congestion control information from destination node to source node in the absence of normal two-way traffic.
Another object is to provide a quiescent information sending rate ("QIR") greater than the guaranteed minimum information rate ("MIR") and less than the peak information rate ("PIR").
Another object is to provide a no rate change feedback indicator in the feedback cell header sent from destination to source node while waiting for the effects of the previous adjustments to occur or while measuring the effects of previous rate adjustments.
A method and apparatus for congestion prevention control, on a per virtual circuit connection basis, in a cell switching communication network handling bursty traffic is described. The control system comprises a source node, optional intermediate nodes, and a destination node. The source node is capable of accepting, queueing and buffering user bursty data, formatting the bursty data into cells, and transmitting the cells to a destination node by a virtual circuit connection through a cell switching network to a destination node. The rate of transmission is incrementally controlled by the state of congestion on the virtual network as reported by the destination node to the source node using the existing two-way virtual connection. Intermediate nodes accept, queue, buffer and forward the bursty cells toward the destination together with an associated incipient congestion indicator. Incipient congestion is measured by monitoring of the virtual connection queue and buffer lengths. The destination node receives the data cells together with the congestion indicator, reconstructs and formats the data for delivery to the end user, and counts the received congestion indicators over an adaptive interval. A feedback rate control signal is generated from the congestion indicator count.
Other objects, features, and advantages of the present invention will be apparent from the accompanying drawings and detailed description that follows.
BRIEF DESCRIPTION OF THE DRAWINGS
The present invention is illustrated by way of example, and not limitation, in the accompanying figures in which like references indicate similar elements and in which:
Figure 1 is a block diagram of a telecommunication network for voice and data;
Figure 2 is a cell format diagram;
Figure 3 is a functional block diagram of a telecommunication node;
Figure 4 is a data flow diagram;
-Figure 5 illustrates an example of congestion in a cell network;
Figure 6 illustrates the effects on throughput and delay due to congestion;
Figure 7 illustrates a cell switching network internal congestion avoidance ("ICA") scheme;
Figure 8 is a functional block diagram of a frame relay pad ("FRP");
Figure 9 is a functional block diagram of the transmit portion of a transmitter/receiver ("TXR");
Figure 10 shows cell loss priority ("CLP") two level thresholding in a TXR;
Figure 11 shows the architecture of a FRP and a frame relay interface (FRI);
Figure 12 is a block diagram of the T1 transmitter/receiver;
Figure 13 is a diagram of a bursty data queue;
Figures 14(a) and 14(b) comprise a flow diagram of a cell traffic servicing routine; Figure 15 shows the c-bit count interval timing using the regeneration cycle algorithm;
Figure 16 is a flow diagram of the data rate control algorithm;
Figure 17 is a flow diagram of the ICA credit manager function; and
Figure 18 is a flow diagram for the servicing of the muxbus cell transmit queue by the cell transmitter.
DETAILED DESCRIPTION
Figure 1 illustrates a fully-integrated voice and data T1 telecommunication network 20 using telecommunication nodes 22, also referred to as integrated cell exchangers 22. The network 20 shown is a domestic network. Those skilled in the art will appreciate that the illustrated network 20 can be modified in known ways to accommodate international traffic by the addition of various interfaces to user equipment and T1-to-E1 interfaces to each node 22.
Each node 22 incorporates a T1 transmitter/receiver that includes the fair queuing and servicing circuitry. The T1 transmitter/receivers support six classes of cell traffic: high priority (HP), voice, low speed statistical (LSS), high speed deterministic (HSD), bursty, and multicast. As will be discussed in detail below, each T1 transmitter/receiver supports the traffic classes via six queues and a service routine. The service routine guarantees a minimum amount of bandwidth to each class of traffic under normal operation and allocates spare bandwidth according to a predefined priority scheme.
Network 20 of Fig. 1 includes nodes 22 in Atlanta, Boston, Chicago, Dallas, and Seattle. Each node 22 is connected to local communication equipment. For example, in Atlanta, a private branch exchange (PBX) telephone system 24, a fax machine 26, a computer 28, video equipment 30, local area networks (LANs) 32, high speed statistical equipment 34, a control terminal 36 and a printer 38 are all connected to the integrated cell exchanger 22. All communication equipment associated with a node 22 is connected through circuit T1 trunks 40. Each telecommunication node 22 takes existing voice and data streams from its associated communication equipment, assembles the streams into a more efficient cell format, and transmits the cells between nodes via cell T1 lines 42 (shown by broken lines). Similarly, each telecommunication node 22 receives cells from the cell T1 lines 42, disassembles the cells into data streams, and transmits those streams to the appropriate communication equipment via circuit T1 lines 40. Thus, each integrated node 22 can function both as a source and a destination of information.
The term "T1" refers to a telephone line operating at a synchronous data rate of 1.544 million bits per second. T1 lines are digital, and voice signals are digitized prior to their transmission over T1 lines. Under the DS-0 signaling standard, T1 line capacity is segmented into 24 individual channels each operating at 64 thousand bits per second to support voice transmission.
Each node 22 increases the apparent capacity of cell T1 lines 42 using virtual connections. In other words, rather than committing specific resources to a given source-destination pair, each node 22 connects a source-destination pair only when information, in the form of a cell, is present. When cells are not being created for a given source-destination pair, the same network resources are used to transmit cells for other source-destination pairs.
Information is transmitted between nodes over the T1 lines 42 in the form of cells. Cells are packets of fixed length and are typically used in systems that switch a number of traffic classes, as opposed to a single traffic class as in packet switching. The short message length of cells as compared to packets permits cell switching networks to achieve very high rates of network utilization and still maintain short transmission delays. Despite the distinction between cells and packets, the terms are used interchangeably herein.
A general cell format is illustrated in Fig. 2. The cell includes 24 octets, or 8-bit bytes, of information. The first two bytes, octet 0 and octet 1 , represent the cell's virtual circuit identifier. The virtual circuit identifier allows the cell to be self-routing through network 22. Octet 2, the third byte, indicates the type of data contained within the cell namely, voice, video, etc. Octet 2 also includes a cyclic redundancy check ("CRC") code, which is used by node 22 to perform an error check on the first four octets of the header. Octet 3, the fourth byte of the cell, may contain a time stamp, congestion control bits, or hopcount depending upon the traffic class. The remaining octets of the cell are generally devoted to information or payload.
Using the traffic type bits in octet 2, nodes 22 support six types of cell traffic. Identification of different traffic classes allows T1 transmitter/receivers to discriminate among classes of cell traffic for queuing and servicing purposes. The classes of cell traffic supported by nodes 22 are high priority traffic, voice traffic, low speed statistical traffic.high-speed deterministic traffic, bursty traffic, and multicast traffic.
High priority cells are indicated by a binary 100 in octet 2. High priority traffic includes inter-node cell exchanger-to-cell exchanger messages. This class of traffic is given the highest servicing priority so that network failures and overloads can be quickly remedied. Voice traffic is represented by a binary 010 in octet 2. Voice traffic includes pulse coded modulated ("PCM") and adaptive differential pulse coded modulated ("ADPCM") voice signals. Such traffic is relatively intolerant of servicing delays but, given its redundant statistical nature, can withstand some intermittent loss of cells without a noticeable adverse effect.
Low-speed statistical traffic includes cells generated by equipment operating at less than 64 thousand bits per second. Low speed statistical traffic can tolerate some delay between generation and transmission, but cannot tolerate any loss of data from a message. This class of traffic is represented by a binary 111 in octet 2.
Low speed statistical cells carry a timestamp in octet 3. Using timestamps, nodes 22 can determine how long low-speed statistical cells have been in network 20 and transmit the oldest cells first. Timestamps are also used to discard cells exceeding a maximum age, which is software-setable.
High speed statistical cells encompass high-speed, full-period voice or data signals. This class of traffic tolerates very little service delay and is equally intolerant of loss of data from a message. A binary 110 in octet 2 indicates a cell of high speed statistical.
The bursty and multicast classes of cell traffic supported by nodes 22 are quite different from the four other traffic types. Bursty and multicast traffic transport high-speed, statistical data from sources such as local area networks (LANs), bridges, routers, brouters and high-speed cell switches. Both types of traffic are characterized by burstiness. As used herein, the term "bursty data" refers to bursty point-to-point traffic between nodes 22. Multicast traffic carries bursty traffic from any one to all other nodes 22 in network 20. Both bursty traffic and multicast traffic are relatively delay insensitive. Thus, during periods of congestion, cells of these traffic classes may be buffered rather than discarded. Bursty traffic is represented in octet 2 by a binary 101 and multicast traffic is represented by a binary 011.
Bursty traffic and multicast traffic also differ from other traffic classes in that the octet 3 of each cell contains a hopcount. Hopcounts serve a purpose similar to timestamps, but are more practical than timestamps at high speeds. A cell's hopcount indicates the number of nodes- 22 previously traversed by the cell in its journey between source and destination. The higher a cell's hopcount the greater the cell's service priority as compared to other cells within that class of traffic. In one embodiment, the maximum hopcount is 15; however this number is software selectable and may be modified to suit specific network applications.
A block diagram of a node 22, which includes the T1 transmitter/receiver ("TXR") and frame relay pad ("FRP") of the present invention, is shown in Fig. 3. The circuitry within the node 22 performs two basic functions: traffic segmentation and reassembly. While the node 22 is functioning as a source of messages, segmentation involves receiving incoming bit streams from the communication equipment and assembling it into cell formats. Trunk interface units buffer data in queues and route the cells through the network 20. Similarly, when node 22 functions as a destination for messages, the trunking interface receives cells and routes them to the appropriate circuity for reassembly. Reassembly is the process of converting cells into data streams or packet frames.
Node 22 includes a processor controller 50 , a data assembler/disassembler ("PAD") group 52, a voice data assembler/disassembler ("PAD") group 54, TXR 56, circuit/cell T1 interface 58 and FRP 59. A system-wide system bus 60 which includes two buses, transports data and control information throughout node 22. Part of system bus 60, muxbus is a high speed statistical bus for switching cells. Also part of bus 60, control bus configures, manages, and monitors the components 50, 52, 54, 56, 58, and 59 of node 22.
-Processor controller 50 manages the various circuits within node 22, with direction from an operator using a control terminal and a printer. Processor controller 50 manages the routing of cells via network configuration databases. Processor controller 50 distributes control and configuration information via the control bus to the 50, 52, 54, 56 58, and 59 each of which is coupled to the control bus by an 80C31 controller. Additionally, network processor controllers 50 in Atlanta, Boston, Chicago, etc., cooperate to perform network-wide functions.
Processor controller 50 uses a 16-bit 68000 microprocessor supplied by Motorola Corporation of Schaumberg, Illinois, as its basic processing unit. The 68000 microprocessor contains the software for controlling, diagnosing, and monitoring integrated cell exchanger 22. This software is stored in 2 Megabytes of flash EPROM and 8 Mbytes of DRAM. The preferred embodiment of processor controller 50 also utilizes a bus controller sold by Intel Corporation of Santa Clara, California to manage the control bus. A portion of the node's 22 segmentation and reassembly function is performed by voice/data processor 62. Voice/data processor 62 assembles and disassembles cells of voice and data samples for cell T1 lines 42 using the DS-0 standard. This includes voice-band data generated by modems and FAX machines. Voice/data processor 62 is bidirectional.
Each voice connection within network 20 requires two voice/data processors 62, one at each end of the connection. Utilizing voice activity detection the source voice/data processor 62 determines when to generate cells. Voice activity detection provides a 2:1 compression of PCM voice. The source voice/data processor 62 extracts voice from a T1 port, packetizes it, and sends the packets to the remote voice/data processor 62. The remote voice/data processor 62 reassembles the voice signal from the cell stream and passes it on to the appropriate external device, for example, PBX 24.
Voice compressor/decompressor 64 receives voice samples from voice/data processor 62. The voice samples are converted from PCM to compressed adaptive differential pulse code modulation (ADPCM) when the node 22 is acting as a source. Voice compressor/decompressor 64 converts ADPCM back to PCM when node 22 acts as a destination.
Synchronous data interface 66, synchronous data cell assembler/disassembler 68, and low-speed data cell assembler/disassembler 70 perform another portion of the segmentation and reassembly function within node 22. Both cell assembler/disassemblers 68 and 70, commonly referred to as packet assemblers/dissemblers ("PADs") buffer, assemble and format data cells. The cells are transmitted to and received from system bus. Synchronous data PAD 68 provides four different physical interfaces: RS232C/D (V.24), V.35, RS422/449, and X.21/V.1 . Low speed data PAD 70 supports low speed data applications. Both low speed data PAD 70 and synchronous data PAD 68 can be configured for transparent, bit-for-bit transmission or for data compression using repetitive pattern suppression.
Frame relay PAD (FRP) 59 performs segmentation and reassembly of high-speed bursty data and multicast data. Frame relay PAD 59 is typically connected to LAN devices, such as routers and bridges. In the present invention, FRP 59 provides network cell rate control functions in cooperation with TXR 56.
Queuing and transmission functions for node 22 are provided by T1 TXR 56 and circuit/cell interface 58 of Figure 3. TXR 56 contains routing tables for the cells to be transmitted on cell T1 lines 40 and buffers, or queues, for those cells. Each of the six traffic types supported by node 22 is queued separately, as will be discussed in detail below. TXR 56 performs the T1 line framing and signal control, cell servicing, and alarm detection and generation. Circuit/cell interface 58 provides interface support for bidirectional T1 lines 40 and 42. The conversion of bipolar signals from cell T1 lines 42 into logic-levels for use internal to node 22 is also performed by circuit/cell interface 58.
Data is transmitted across T1 network 20 over a virtual circuit between a pair of data PADs, 68, 70, or 59. The flow of data into and out of node 22 is illustrated in Figure 4. A data stream from local communication equipment is received either by synchronous data PAD 68, low speed data PAD 70, or frame relay PAD 59. The source data PAD 68, 70, or 59 builds data cells at a rate proportional to the baud rate of the data. The PAD 68, 70, or 59 passes the cells to system bus 60 for delivery to TXR 56. Transmitter/receiver 56 places the cells in the appropriate queue according to traffic type where the cells await delivery to cell/circuit interface 58. The cells are transmitted onto cell T1 lines 42 by cell/circuit interface 58. Cell traffic coming into node 22 follows the reverse path.
Figure 5 is an example of a four node network 20 for use in explaining the onset of congestion interval to network 20. Assume that a bursty data working connection between user A at node 22-A and user C1 at 22-C via node 22-B involving FRP 59A1 , TXR 56A1 , TXR 56B1 , TXR 56B2, TXR 56C1 , and FRP 59C1. Further, assume that user D subsequently establishes a working bursty data connection through node B to user C2 at node 22-C involving FRP 59D1 , TXR 56D1 , TXR 56B3, TXR 56B2, TXR 56C1 , and FRP 59C2. If either user A or D should increase their transmission rate because of an excess available input load, node 22-B could become congested at the common TXR 56B2 (shown shaded) if the combined bursty data rate exceeds the available capacity on trunk 42 connecting nodes 22-B and 22-C. Similarly, congestion could occur in the reverse direction at TXR 56B3 of node 22- B interfacing with cell trunk 42 connecting node 22-B and 22-D if the combined bursty traffic between user C2 and node 22-D plus that between, say, user B and node D exceeded the available bandwidth between nodes 22-B and 22-D. Figure 6 shows the effects of congestion on effective network bandwidth and delay as a function of the offered load bandwidth. At low average load rates (Region I), throughput increases linearly with load, while delay shows a moderate rate of increase with load. When throughput approaches the network's information rate capacity, mild congestion results (Region II) as average queue lengths increase. Throughput saturates, increasing only slightly with increased offered load, while delay exhibits sharp increases. Finally, if the offered load increases some more, threshold is exceeded when Region III is reached, causing a breakdown in throughput (severe congestion) and hence unlimited delay because of data losses requiring constant retransmission.
In order to efficiently use the bandwidth resource of a cell switching network, it is desirable that peak offered loads be accommodated by adequate buffering, particularly for bursty data which is more tolerant of network delay. Bursty data tends to be high- bandwidth short-duration messages that, if uncontrolled, may either cause congestion or require that the network operate with a high percentage of average unused bandwidth. Ideally, the average offered load would operate at point A of Figure 6, where peak loads would not be likely to cause severe congestion.
The method and apparatus for congestion prevention control described below is applied to cell switching networks of the type described. These cell networks are designed to interface with user data ports operating at the CCITT asynchronous transfer mode ("ATM") adaptation layer of ISDN networks. The cell network accepts frame relay data at its input nodes and produces frame relay data at its output nodes. The entire operation of the cell switching network is intended to be completely transparent to the frame relay user. For this reason, the congestion control system is hereafter referred to as an internal congestion avoidance ("ICA") method and apparatus, wherein "internal" means internal to the cell switching network.
Figure 7 is an abstract representation of an internal congestion avoidance feedback control system. The purpose of the ICA feedback control system is to anticipate the onset of congestion in the bursty data queues of the network nodes and correct the cause of the incipient congestion in a timely fashion. Corrective action is taken, on a per channel basis, by controlling the rate at which each bursty source node contributing to the congestion, accepts incoming user data.
In Figure 7, assume that node 22-A is the source node, node 22-B is an intermediate node and node 22-C is the destination node. FRP 59 of node 22-A accepts user data from one of the input virtual data ports. It is assumed that the input data contains bursty framed data. FRP 59 disassembles the data frames into cells and regulates the rate at which they are delivered to TXR 56 for transmission over virtual connection 182 to intermediate node 22-B. Other TXR 56 units (not shown) at node 22-B may each receive bursty data cells, each receiving TXR 56 directing the bursty data cells over the system bus 60 to the appropriate bursty data queue in a TXR 56 for transmission over one or more common trunks.
Each of the output TXR 56 units of nodes 22-A and 22-B performs traffic control and an ICA detection function by monitoring the length of its bursty data queue. If the queue length exceeds a prescribed threshold length, a congestion indicator bit (c-bit) is set in the header (octet 3, bit 5) of all bursty data cells transmitted from node B to node C while threshold is exceeded. This same TXR 56 unit also performs a traffic control function including the discarding of cells when congestion conditions require.
At the input to node 22-C a TXR 56 performs the receive functions previously described and directs the cell traffic to the appropriate queue of an output FRP 59 unit via system bus 60 where the cells are assembled into frames for delivery to the user C virtual port 40. In addition, FRP 59 examines the headers of each bursty data virtual circuit and generates an estimate of the congestion status by comparing the average c-bit rate with a threshold reference value. If threshold is exceeded, indicating incipient congestion, an ICA rate control message is fed back to the source FRP 59 unit in source node 22-A, where user A's input rate is decreased so as to relieve the congestion present on its virtual circuit.
The FRP 59 unit of the source node 22-A also provides an additional traffic control feature by marking burst data cells that are delivered by user A at an excess rate. A cell loss priority ("CLP") indicator bit is set in the cell header (octet 3, bit 4) and, like the c-bit, once set remains set throughout its journey to the destination FRP 59. This CLP bit is used by the intermediate transmit TXR 56 of node 22-B for prioritizing cells for discard when required by congestion conditions. Cells marked by asserted CLP bits are discarded first. Of course, the discard of any one cell in a given data frame requires the retransmission of the entire frame of data. A two threshold algorithm is used in order to discard cells that, with high probability, belong to the same frame.
Although this overview of the ICA control system was explained in the context of a data transmission virtual link between source and destination, it should be noted that the virtual connections are usually two-way connections sharing common resources. Hence, for one embodiment both source and destination nodes have symmetrically interchangeable roles. As will be seen in the following discussion, each FRP 59 and TXR 56 has a distinct receiver function coupling node input data to system bus 60 and a distinct transmit function coupling data from system bus 60 to the node output.
The ICA rate control feedback path 180 is the return path of the existing virtual connection. Intermediate node B is not necessary for implementing a virtual connection or for operation of the ICA system. Instead, the intermediate node B should be interpreted representing a possible multiplicity of concatenated intermediate nodes performing similar functions.
Figure 8 is a more detailed functional description of an FRP 59 unit. Input user data arriving on user input 40 is accepted by traffic monitor 801 that compares data with the guaranteed minimum information rate ("MIR"), the committed burst duration (Be) and the quiescent information rate ("QIR") time-out parameter, TQ. If MIR, for a given connection, is exceeded, a CLP bit is set in each frame exceeding the MIR. Frames are sorted into distinct virtual connection ("VC") queues 803, one per connection. Segmenter 805 operates on the data in VC queues 803, in an order prescribed by the serving protocol, and segments the VC frames into VC cells, which are sent to system bus 60 of node 22 by cell transmitter 807. Additionally, segmenter 805 sets a bit in the final cell of a frame if that frame was received with its DE bit set (indicating excess frame rate at the frame relay layer). The rate of transmitting cells for each virtual connection by cell transmitter 807 is controlled by credit manager 809 operating cooperatively with rate controller 811 to provide incrementally variable transmission rates. C aχ is a configurable parameter for credit manager 809 that specifies the maximum number of cell transmission credits that may be accumulated by a given virtual connection, when the virtual connection is idle.
- Rate controller 811 initially sets each VC cell rate to a value equal to the configuration parameter QIR, representing the quiescent information rate. QIR is bounded on the lower side by the guaranteed minimum information rate (MIR) and on the upper by the peak information rate parameter ("PIR"). The rate setting is subsequently dynamically adjusted by rate commands received from the VC destination FRP 59 via cell receiver 815. For one preferred embodiment, a two-bit coded message embedded in the cell header (octet 3, bits 2 and 3) of return traffic is used to encode a four-state rate message: increase, decrease, large decrease, and no-change. In response to the rate control message, controller 811 increases the rate by linear additive increments (typically 1/10 of MIR) and incrementally decreases the rate by a multiplicative factor of 7/8 or 1/2, the latter being used when cell loss in conjunction with congestion is detected by the destination FRP 59. Controller 811 also monitors the VC queue depth to see if sufficient data is present to warrant a data rate increase. No increase is made if sufficient data is not present. Traffic monitor 801 also provides controller 811 with a reset signal that causes the rate to be reset to QIR when no data has been presented to traffic monitor 801 for a period equal to configuration parameter, TQ.
Credit manager 809 performs the function of controlling the intervals at which each virtual connection is permitted by cell transmitter 807 to transmit a cell to system bus 60.
The receive portion of FRP 59 comprises cell receiver 815, reassembler 817, DE filter 821 , port transmit queues 819 and traffic analyzer 813. Receiver 815 interfaces with the local system bus 60 from which-it receives its virtual circuit data from the connection's other terminal node. The congestion indicators (c-bits) and loss data information is provided to traffic analyzer 813. Cell data is passed on to reassembler 817 that reconstructs frames from cells and sets the DE bit if the last cell of the frame indicates that the DE bit was set in the frame header when received from the sending user. DE filter 821 monitors the port transmit queue depths and compares them against a threshold, DEth, typically set at approximately 75% of total queue depth. If threshold is exceeded, frames with the DE bit asserted are discarded first.
Traffic analyzer 813 analyzes the average c-bit rate over the measured round trip delay ("RTD") provided by PCC 50 of the local node 22 of Figure 3. If the average count of c-bits exceeds a prescribed threshold, TAth, (typically 1/2 of total possible count), a moderate rate decrease message is generated and supplied to cell transmitter 807 where it is embedded in a cell header and transmitted to the other terminal FRP 59. If cell loss is also present, a large rate decrease message is sent. If threshold is not exceeded a rate increase message is generated. A no-change rate message is sent with all outgoing cells for an interval of two times RTD, so that the effects of the previous rate change message may be observed before introducing additional changes. This delay in adjustments is necessary for stable congestion control.
Traffic analyzer 813 also monitors port transmit queues 819. If the output port transmit queue used by a particular VC shows incipient congestion, any rate increase message generated by traffic analyzer 813, based on the VC average c-bit count not exceeding the TAth threshold, is changed to a rate decrease message even though the c-bit count threshold, TAth, is not exceeded. In one embodiment, output port incipient congestion is determined by comparing the average queue depth of port transmit queues 819 against a configurable threshold, PQth. The PQth threshold is normally set at four kilobytes, a fraction of the 64 kilobyte queue capacity. Average queue depth is calculated by traffic analyzer 813 using a regeneration cycle averaging period.
A regeneration cycle is defined as a busy plus idle interval of burst data. This average may be used for the entire duration of the next cycle. Alternatively, when the regeneration cycle is very long, the averaging may include the previous regeneration cycle count as well as the current count, averaged over the previous regeneration cycle time plus the current portion of the subsequent regeneration period.
In addition, test generator 861 , echo unit 863 and RTD measure unit 865 comprise the means for measuring the RTD of a virtual connection. Periodically, PCC 50 of local node 22 initiates an RTD measurement test by instructing test generator 861 , via system bus 60, to generate a high priority RTD cell for transmission by cell transmitter 807. When the RTD cell is received at the other terminal node by its FRP 59 cell receiver 815, it is directed to RTD echo unit 863 and retransmitted as a high priority RTD cell via its associated cell transmitter 807. Upon receipt of this echoed RTD cell, the initiating node 22 cell receiver directs the cell to RTD measure unit 865 that measures the elapsed time between the time of original RTD cell transmission and the receipt of its echo. RTD measure unit 865 provides this data to PCC 50 via system bus 60 which in turn supplies this information as an RTD configuration parameter to traffic analyzer 813.
Figure 9 shows a more detailed functional block diagram of the transmit portion of TXR 56. This part of TXR 56 provides the traffic control and congestion detection in the preferred embodiment. On one side, TXR 56 interfaces with system bus 60 of its local node 22 from which it receives cells destined to be transmitted over a common trunk 42.
Bursty cells arriving at the input of TXR 56 are sorted into queues 135 according to traffic class. Bursty data cells are queued in BD queue 855 where the average depth is monitored for indications of incipient congestion by congestion detector 857 by comparing the average queue depth with configurable threshold parameter, ECNth. If threshold is exceeded, the next bursty cell processed by cell server 853 will have a c- bit set in its header. Threshold ECNth is selected so as to introduce a tolerable amount of delay due to queue length while minimizing the amount of cells discarded. A value of 30 cells has been found acceptable.
The average BD queue 855 queue depth is also based on a regeneration cycle averaging period. Also, the BD queue 855 depth is used to control the discard of cells by CLP filter 851 which discards arriving bursty cells with the CLP bit set if threshold CLPthl is exceeded. Discarding of CLP labelled cells continues as long as a second threshold, CLPth2, is exceeded (CLPth2 > CLPthl ). In this manner, hysterisis is provided in the control mechanism so that it tends to discard the following cells belonging to the^ame frame, as shown in Figure 10. This strategy recognizes that once a cell belonging to a given frame is discarded, the integrity of the entire frame is violated, requiring retransmission of the entire frame. Also, this control method has the added advantage of more probably relieving the congestion, rather than oscillating about a single threshold value. Thresholds CLPthl and CLPth2 are configurable parameters supplied by PCC 50 of Figure 3 and typically set a approximately 75% and 25% of queue capacity, respectively.
In order to facilitate the explanation and understanding of the various embodiments, the following is a more detailed description of the FRP 59 and TXR 56 architecture and hardware wherein the invention is implemented.
Figure 11 shows the architecture of a FRP 59 and of a frame relay interface (FRI) unit 300 used to interface FRP 59 to four V.35 data ports, in one preferred embodiment. DMA controller 301 of FRI 300 is a dedicated Motorola Corporation 56001 digital signal processor ("DSP") that interfaces between global frame RAM 230 and two-way serial communication controller ("SCC") 303. Four V.35 ports are provided, each capable of operating full duplex at bit rates up to 2.048 Mbits/sec. DMA 301 receives frames from the ports via SCC 303 and stores them in frame buffers 237 of global frame RAM 230. Also, DMA 301 transmits frames from frame buffers 237 to the data ports through SCC 303. In addition to data transfer, DMA 301 handles start and end of frame interrupts, extracts the cyclical redundancy check ("CRC") indicator from SCC 303 and appends it to the frame, and interfaces with frame transmit controller 202 and frame receive controller 201 in ACP 200 through queues in global frame RAM 230. Once the data ports are configured and enabled, DMA 301 takes pointers to frame buffers 237 from port free buffer queue 234 for use in storing incoming frame data. Because all ports may operate simultaneously, DMA 301 maintains a separate control block for each port. When a start or end of frame interrupt occurs during frame reception, DMA 301 notifies frame receive controller 201 through port frame receive queue 231. For frame transmission, DMA 301 polls frame transmit queues 232 (one per port). Separate control blocks are maintained for each port. An entire chained frame may be transmitted or received without ACP 200 intervention for each frame buffer. DMA 301 maintains the following information for each transmitted or received frame: pointer to the start of frame in global frame RAM 230, pointer to the current work location in RAM 230, and current length of the frame. FRP 59 comprises four major units: administration/communication processor (ACP) 200, multiported global frame RAM 230, cell receiver 210, and cell transmitter 220. Four processors are included on the FRP 59 board: an Intel Corporation 8751 microcomputer acting as a control bus kernel, interfacing FRP 59 to the control bus of system bus 60; a Motorola Corporation 68020 microprocessor that serves an administrative and main communication control (ACP 200) functions; two Motorola Corporation 56001 digital signal processors, one dedicated to cell transmitter 220 and the other to cell receiver 210. Figure 5 shows two paths between the 68020 microprocessor of ACP 200 and the 56001 microprocessors in receiver 210, transmitter 220, and DMA 301. The solid lines represent the typical flow of data and control information between the microprocessors and global frame RAM 230. Normal frame relay communications is done through these data paths. The dashed lines represent a secondary control path between ACP 200 and the host port of each 56001 microprocessor for handling administrative functions such as diagnostics and statistics collection.
Control bus command server 203 of ACP 200 responds to commands received from processor controller 50 (Figure 3) over the control bus 95 portion of system bus 60, sending responses to event server 204 for forwarding to processor controller 50 via a FIFO queue to the control bus kernel microcomputer, or for forwarding commands to the communication units, i.e.; frame receive controller 201 and frame transmit controller 202. Controllers 201 and 202 perform real-time control functions. Cell transmitter 220 maps frames received from FRI 300 data ports to cell network connections, applies the ICA control scheme, fragments frames into cells, and transmits cells on muxbus 96 toward the destination node. Cell receiver 210 coordinates receipt of frame fragments (cells) from muxbus 96, reassembles cells into frames. Frame transmit controller 202 transmits those frames to the proper user port, linking cell receiver 210 to DMA 301 through RAM 230, where reassembly takes place. Transmit controller 202 maintains muxbus free buffer queue 233 for the use of receiver 210 to determine the next buffer to use to store reassembled frame data. LMI 205 is an optional local management interface for verifying the integrity of the V.35 physical link and for providing status from the cell network to the user device. These ACP functions (receive controller 201 , transmit controller 202, command server 203, event server 204 and LMI 205) are tasks executed under the operating system running on the 68020 microprocessor.
Cell receiver 210 and transmitter 220 operate as dedicated controllers. Communications between them and communication controller tasks 201 , 202 and DMA 301 is primarily through queues in global frame RAM 230. Using these queues simplifies the timing requirements on the communication control tasks, allowing them to operate less frequently on larger groups of requests.
Queue 236 points to the location of the cell data in frame buffers 237 of RAM 230 together with the number of octets to send, the virtual circuit identifier for the cell, the initial timestamp value to be placed in the cell header, and an indication of whether the cell is a start-, middle-, or end-of-frame fragment. The cell is sent to muxbus 96 via the muxbus transmit FIFO of cell transmitter 220. Cell transmitter 220 controls the rate of transmission of cells onto muxbus 96 by means of a credit manager (similar to that described for servicing cell traffic by TXR 56). After the start of frame indication, queueing of cells for transmission on muxbus 96 does not occur until one of the following occurs:
(1 ) the end-of-frame indication is received in frame receive queue 231 , so that the exact length of frame is known, or
(2) the data received since the start-of-frame indication is sufficient for a full cell's payload data to have been received.
Cell transmit queue 236 is polled successively to see if one of these two conditions (ensuring sufficient data to transmit a cell) prevail. If so, the control is subjected to the ICA credit manager control scheme to be described later.
Figure 12 illustrates T1 transmitter/receiver 56 that implements the circuitry and method for queuing and servicing cell traffic. T1 TXR 56 is divided into two sections, a transmitter 72 and a receiver 70.
Receive state control 74 coordinates the operation of T1 receiver 70. Incoming data from another node 22 is received as a T1 bit stream 80 and an extracted clock 82 sent by circuit/cell interface 58. T1 deframer 84 strips the T1 framing pattern from the incoming bit stream and sends the received octets of the cell through descrambler 86 and CRC checker 88. If the CRC for the cell is in error the cell is destroyed. If the CRC is correct, the cell is placed into receive packet FIFO 90.
When muxbus output control 92 is notified by muxbus address and control signals 94 that it may dispatch a cell onto muxbus 96, the first cell stored in receive packet FIFO 90 is driven onto muxbus data bus 98 through output register 100. If there are no cells in receive packet FIFO 90, a cell consisting of "all ones" is written to muxbus data bus 98.
The operation of the T1 transmitter 72 is coordinated by queue manager 76. Queue manager 76 includes a high-speed microprocessor 110, RAM 112 and program ROM 114. The fair queueing and service routines are implemented in software, which runs on microprocessor 110. The fair queuing and servicing routines will be described below, following a general description of the structure and operation of transmitter 72.
The arrival of a cell via muxbus 96 to transmitter 72 is accompanied by a combination of asserted address and control signals 116 to input state control 118. Coincident with these control signals 116, the first two octets of the cell, comprising a virtual circuit identifier, are received into the pipeline registers 120 from muxbus data bus 96.
If the cell is to be queued, then input state control 118 performs a series of functions. First, a new packet start address is fetched from packet start address FIFO 124 which is passed together with the output of counter 126, through mux 128. This forms the address for queue memory 135. As each successive octet of the cell exits pipeline registers 120 onto queue memory bus 140, each byte is written into a location in queue memory 135 pointed to by the packet start address and indexed by counter 126, which increments with each octet.
Simultaneously, a packet information block is created for the cell. Octet 2 and 3 of the packet and the current time, which is used as a queue input time, are written into packet information FIFO 152. This is done through mux 154, and forms the packet information block. Thus, queue manager 76 is notified that a cell has just been queued, where that cell is stored, the traffic type, its timestamp value or hopcount, if applicable, and when the cell arrived.
Current time counter 158 is incremented by the 125 microsecond clock 160 from muxbus 96. Current time counter 158 may be interrogated by queue manager 76.
When queue manager 76 decides to dispatch a cell, it writes the packet start address and updated timestamp or updated hopcount, if applicable, to output state control 144. Output state control 144 fetches each byte of the cell to be dispatched from queue memory 130 by applying the appropriate addresses to queue memory 135 through mux 128. Each octet of the cell is moved via queue memory bus 140 through MUX 146 to CRC generator 142. Queue memory 135 is time-shared between input state control 118 and output state control 144 .
In the case of timestamped data cells, an updated timestamp, previously written to output state control 144 by queue manager 76, will replace the fourth octet of the cell. This is achieved by activating mux 146. CRC generator 142 calculates and then writes a CRC-5 code into the lower bit field of the traffic type/CRC octet. The CRC covers the first four octets of the cell.
Similarly, in the case of cells bearing hopcounts, the cell hopcount will be updated prior to dispatch by queue manager 76. Again, the cell CRC will be recalculated by CRC generator 142.
The octets are then sent through scrambler 148 to T1 framer 150, which inserts cells into T1 frames. Given this general description of the operation of transmitter 72, it will be appreciated that the difference in treatment between the various classes of cell traffic is determined by queue manager 76. Queue manager 76 performs two major functions: queueing cells and servicing cells. During queueing and servicing, queue manager 76 treats the various classes of cell traffic differently.
During queueing, queue manager 76 fetches packet information blocks from packet information FIFO 152. Packet information blocks contain a representation of the cells that have been stored in queue memory 130. The traffic type field of the packet information block is interrogated to determine into which queue the packet information block should be appended. There are six traffic queues: high priority, voice, low/speed statistical, high speed deterministic, bursty, and multicast.
The queues are linked lists of packet information blocks. Figure 13 illustrates the queue structure of the bursty and multicast traffic queues, which utilize hopcounts.
The bursty traffic and multicast traffic queues are nearly identical. Thus, while the description that follows refers to the bursty traffic queue, it will be understood that the following description applies equally to the multicast traffic queue. Queuing of bursty traffic differs from the other queuing schemes. The difference arises, in part, because these traffic queues include a number of subqueues as shown, for example, in Figure 7 and numbered 1 through 15. Each of the subqueues is a FIFO queue including a set of linked-lists. As each cell of bursty traffic is received it is placed into one of the subqueues according to the hopcount stored in octet 3 of the cell. Thus, a cell of bursty traffic with a hopcount of 13 will be placed at the end of subqueue 13 of the bursty data queue. Cells of bursty traffic having hopcounts greater than 15 are simply discarded, as they have used more network resource (delay) than permitted. The maximum hopcount is 3 for the bursty queue for one embodiment due to limited coding bits in octet 3 of the cell header. This maximum could, however, be increased to allow cells of bursty data traffic to use more network resource by increasing the hop-count coding capacity. Because of this coding bit limitation, cells are not discarded when the count exceeds 3 but continue to carry a maximum hop-count of 3.
Cells within the bursty traffic queue are serviced according to subqueue priority order, with subqueue 15 having the highest priority and subqueue 1 having the lowest priority. For the bursty traffic queue a subqueue pointer indicates the highest ranked non-empty subqueue. The indicated subqueue will be serviced first whenever bursty data traffic is serviced. The subqueue pointer is updated whenever cells are placed into the bursty data queue or the queue is serviced.
Servicing the bursty traffic queue includes taking the cell designated by the subqueue pointer and updating both the linked lists and queue depths. Additionally, during the servicing of the bursty data queue the hopcount of the serviced cell is incremented by 1. If the serviced cell was the last cell in the subqueue, the subqueue pointer is updated to indicate the highest priority non-empty subqueue.
The result of the priority scheme used in the bursty traffic queue is that priority is given to cells which have used more network resource (delay). Possible undesirable effects of this priority scheme are freezing of lower priority subqueues during periods of congestion and lengthy queuing delays. These effects are avoided by an aging mechanism. The aging mechanism allows overaged cells to be discarded, while cells that have spent too long in one subqueue are moved to the next subqueue with a higher priority. A cell's queuing priority is effectively increased by moving to another subqueue, but the hopcount is unaffected.
Table 1 is an example that details the mapping of service order, j, and spare bandwidth priorities, k, for each class of traffic, i, in the preferred embodiment. Note that the service priority is according to assigned minimum bandwidth.
Table 1
Figure imgf000041_0001
*Service order determined by minimum configured bandwidth. **Spare bandwidth priority for classes i = 3, 4, and 5 are equal.
The servicing routine uses a credit accrual scheme to ensure that each class of traffic receives a selectable minimum bandwidth. In selecting minimum bandwidths for each class of traffic, let N denote the total available bandwidth on a cell T1 trunk and let T denote the queue server tick interval. The unit of N is not relevant; it can be specified as a number of cells per second, or any other throughput unit. For a non- fractional T1 trunk N= 8000 cells per second. Similarly T can be given in any convenient unit of time. For one embodiment of the node, the tick interval T equals 125 microseconds. Thus, the product N*T represents the capacity of the cell trunk per tick interval, or the quantum of bandwidth.
Each class of traffic is assigned a minimum amount of the quantum of bandwidth, with the exception of high priority traffic. This is because all high priority traffic will be serviced regardless of the required bandwidth. The sum of the minimum class bandwidths must be less than N to allow some bandwidth for high priority traffic. In other words, if i represents the class number, and Nj represents the minimum bandwidth assigned to the ith traffic class, then N*ι + N≥ + N3 + N4 + N5 < N.
Each minimum bandwidth Nj can be transformed into a timer value, Df, representative of the number of tick intervals T that must elapse for traffic class i to acquire its quantum of bandwidth. The timer value Dj = (1/Nj)/T. Note that Dj may not be an integer value because it represents a ratio of bandwidths, i.e., Dj = N/Ni because N*T=1.
Given selected timer values Dj for i=1 , 2, 3, 4, 5, a credit accrual routine runs simultaneously with the service routine. Each class of traffic i is assigned a timer Tj, which is initialized to the associated timer value, Dj. The timer Tj is decremented every T units of time. When the value of timer Tj is less than or equal to zero a transmission credit Cj accrues for traffic class i. Because of the inverse relationship between Nj and Dj, the greater the allocated minimum bandwidth for a class of traffic, the faster the rate at which it acquires transmission credit. The presence of a transmission credit permits a cell from traffic class i to be serviced. After servicing of class i, timer Tj is updated by adding Dj to the previous value of Tj. Using this method of accrual, each class of traffic i accrues Nj credit in a tick interval of T.
The maximum number of credits, Cmaxj, that may be accrued for each class of traffic i is selectable. For one embodiment, the maximum credit that may accrue to any traffic class is 1.
Figure 14(a) is a flow diagram of the service routine for a single tick interval implemented by queue manager 76. Using a credit based strategy for servicing cell traffic, queue manager 76 guarantees each class of traffic a minimum bandwidth.
At the beginning of the tick interval queue manager 76 initializes the servicing routine at step 400, and at step 402 examines flag Qo for high priority traffic. If Qo =1 , then a cell of high priority traffic is queued at step 404, and queue manager 76 will service a cell of high priority traffic. Thus, it will be understood that no credit is necessary for high priority traffic to be serviced. The only requirement for servicing of high priority . traffic is the presence of a cell in the high priority queue.
If, on the other hand, Qo = 0 because no cells are present in the high priority queue, then queue manager 76 begins examining the availability of cells and credit for each class. This examination occurs in order of traffic class service priority indicated by index j. Thus, it will be understood that the next step (406) is setting the service order, j = 1 , by incrementing index j. In state 408, the queue manager 76 finds class i associated with service order j. For the ith class, found in step 408 queue manager 76 examines the associated boolean flag Qj in step 410 and the associated credit C; in step 412. A cell of traffic for class i will be serviced if credit has accrued to that class of traffic and a cell of that class is present in the associated queue. In other words, a cell of traffic from class i will be serviced in step 414 only if Qr = 1 and Ci = 1.
If Qj or Cj =0, no credit is available for class i or no cell is queued for class i, then spare bandwidth results. Consequently, index B, which indicates the number of spare bandwidth credits available is incremented in step 416.
Step 418 checks to see if the priority order index, j, has been exhausted, and if not, returns to step 406 where index j is incremented. If all values of j have been exhausted, step 420 checks to see if B > 0, indicating that spare bandwidth is available for distribution in accordance with protocol 800 referenced in step 422. Otherwise, the process terminates.
Figure 14(b) is a flow diagram for the spare bandwidth allocation process 800 which is initiated by setting the spare bandwidth priority index so that k=1. In step 802, the traffic class index, i, is set equal to the value of k. Step 804 checks boolean flag Qj to see if data is present, and if so, proceeds to step 806 where the credit, Cj, is checked. If Cj=1 , then the ith class is serviced in step 808 and the excess bandwidth index B is decremented in step 810. Step 812 checks if any excess bandwidth remains and, if not, the process ends. If excess bandwidth is not exhausted, or if Qj=0 or Cj=0, the process moves to step 814 where index k is incremented. Step 816 checks the value of index k: if k is less than 3, the process returns to step 802; if 3 < k < 6, then the process proceeds to step 818; and if k=6, the process moves to step 822.
If 3 < k < 6, then one of three possible and equal priority queues may be serviced. In order to ensure fair and equal distribution of excess bandwidth to voice (i=3), bursty (i=4), or multicast (i=5) data, steps 818 and 820 service these three round-robin by incrementing index n (mod 3) in step 818 and setting i=3+n in step 820. The process then proceeds back to step 804. When the process ends because all excess bandwidth has been allocated, index n remains set at its last value. The next time that excess bandwidth obtains after class i=1 and i=2 have been serviced, index n picks up the next round-robin value in step 818.
If the test in step 816 indicates that k=6, all five classes have been serviced. Step 822 tests to see if excess bandwidth still exists and if so, repeats the sequence by initializing the priority index so that k=1 and then proceeds to step 802. Otherwise, the process ends.
The order of allocating spare bandwidth described causes the impact of heavy high priority traffic to be born primarily by bursty data, multicast data and voice data. Correspondingly, low-speed statistical data and high speed statistical are less affected by periods of heavy high priority data.
The described method of allocating spare bandwidth between various traffic classes by TXR 56 is an open-loop control system because the data rate is controlled by the sending node without any feedback from the cell switching network. This procedure leads to a conservative allocation of network resources because each terminal network node acts independently and without specific knowledge of the traffic state of the network. In order to achieve higher bandwidth utilization by bursty traffic, without undo congestion on a given virtual connection, it is necessary to provided the ICA feedback information about the level of bursty traffic being handled by all FRPs involved with a given virtual connection.
ICA is configurable on a per connection basis. The configurable MIR and PIR guarantee that each connection gets at least its minimum allocated bandwidth, independent of other traffic.
System software resident in each process controller 50, implements the user interface and ICA node functions, and further comprises the following functions:
(1) enables/disables the ICA feature on a per node basis;
(2) configures connections using the following parameters:
MIR: minimum information rate expressed in Kbps which is translated into cells per second for internal node use;
PIR: peak information rate expressed in Kbps (internal cells per second);
QIR: quiescent information rate corresponding to the initial rate after a period of ICA inactivity, expressed in Kbps (MIR
< QIR < PIR);
TQ: quiescent information rate (QIR) time-out parameter;
VCqd: virtual connection maximum queue depth in bytes;
Be: committed burst in bytes (ICA only);
Be: excess burst in bytes (ICA only); Cmax: maximum credit count, expressed in cell units and used to ensure fairness of access by guarding against an unreasonable credit count, typically set at Cmax = 10 corresponding to two typical frames of five cells each; ECNth: virtual connection (VC) queue threshold in bytes for explicit congestion notification (ECN); CLPthl ,2: cell loss priority thresholds in bytes (CLPthl > CLPth2) to control loss of newly arriving cells with CLP bit set at 1 (indicating input rates in excess of MIR) are configurable for each TXR;
TAth: traffic analyzer average c-bit count threshold; and PQth: output port transmit queue threshold. (All of the above are not independent - (CIR, Be, Be) or, (CIR,
VCqd, PIR) or, (CIR, Be, PIC), or (CIR, VCqd, Be) may be used interchangeably as optional parameter sets.)
(3) periodically requests a route's RTD measurement from the destination FRP and makes the necessary reconfigurations to perform the measurement;
(4) controls the routing of ICA connections only through ICA- capable nodes; and
(5) collects operating statistics from TXRs and FRPs.
For one embodiment, the two identical TXR queues previously labelled as bursty and multicast are redefined to accommodate the ICA feature. All ICA bursty traffic uses the bursty data queue, while non-ICA bursty data and multicast traffic share the same multicast queue. In this manner, both fCA non-ICA bursty data can be accommodated in the same cell switching network.
Also, for one ICA embodiment, the description of TXR 56 further comprises the following firmware functions:
(1 ) calculating the ICA average queue depth using the "regeneration cycle" averaging period;
(2) setting the congestion bit (c-bit) in the cell header when congestion is incipient as defined by configuration parameter ECNth;
(3) configuring and setting thresholds for ICA queues using -ECNth for setting the c-bit and CLPth 1 and 2 for discarding new excess rate cells with CLP bit set; and
(4) maintaining statistics on the number of cells with c-bits set and the number of cells with CLP-bits set.
FRP 59 firmware provides augmented functions in support of ICA operation comprising:
(1) selectively setting the initial network cell rate at QIR, and varying the cell rate between MIR and PIR in accordance with ICA information received from the destination FRP which, for one embodiment, provides a configurable additive rate increase (+10% of MIR), multiplicative rate decrease (7/8 of current rate), or multiplicative fast decrease (1/2 of current rate), and also provides for ignoring "increase rate" message if the rate is at PIR, or if there is no offered load in the VC queue (cell transmit queue 236), ignoring "decrease" rate or "fast decrease" rate message if the rate is at MIR, and resetting rate to QIR if traffic is absent over a prescribed period of time;
(2) measuring the adaptive rate adjustment period equal to twice the round trip delay for each connection;
(3) counting the c-bit generated by the last rate change instruction to the source FRP, i.e., after a delay of one RTD and counted over an additional RTD interval or until a cell is ready to be sent back, whichever is longer;
(4) measuring configuration parameter RTD approximately every five seconds for one of each connection being served by -generating and sending a HP cell when an RTD measurement is requested by system software, and echoing back to source any RTD measurement HP cell received;
(5) generating a feedback control signal from the filtered or averaged c-bit count, comparing the average count with threshold TAth and sending it back to the source node 22 over the same two way connection in a queued cell or supervisory cell header and coded as follows:
(a) if less than TAth of the possible c-bits are set, send an "increase rate" adjustment message back to the source node;
(b) if greater than TAth, send a "decrease rate" adjustment message;
(c) if a cell loss is detected and some c-bits are set, send a "fast decrease" adjustment message; and (d) send a "no-rate-change" message during regeneration cycles while the destination FRP is waiting for the effects of the previous rate change; and
(e) send a decrease rate message to the VC source node if the threshold, PQth, of the output port transmit queue 819 associated with that VC is exceeded, even if threshold TAth is not exceeded.
Additional FRP functions operating in conjunction with the extemal frame relay network comprise:
(1) monitoring incoming traffic rate and setting the CLP bit for all cells generated from frames received at rates exceeding the committed burst, Bc;
(2) copying the DE-bit of incoming frames to the CLP-bit of the cell header; .
(3) setting the DE-bit of the extemal frame header using the same criteria as for the CLP-bit of the internal cell header;
(4) configuring FRP 59 so that each frame VC queue length can be thresholded so that if the frame DE-bit is set, incoming frames are discarded when threshold is reached; and
(5) configuring each frame transmit port queue 232 with a DE threshold (DEth) so that when reached, incoming frames from muxbus 60 with DE-bit set are discarded.
In order to further explain the operation of the ICA system, a more detailed description of the ICA algorithms follows for counting c-bits and controlling source node transmission rate. ICA control is based on the detection of incipient congestion at each node traversed by the virtual connection. Incipient congestion obtains whenever the average bursty data queue length in TXR 56 is greater than congestion threshold parameter ECNlti. The average queue length is based on queue length measurements taken over a regeneration cycle consisting of a busy plus idle interval of the bursty data. If ECNth is exceeded, a c-bit is set in the cell header of the next cell to be transmitted by queue manager 76.
Destination ACP 200 of FRP 59 counts (averages) the c-bits over a RTD interval as shown in Figure 15. At the end of the averaging period, a rate adjustment message may be sent by the destination FRP 59 through cell transmitter 220 onto muxbus 96 and thence through TXR 56 and the virtual connection to the source FRP 59. The source FRP 59 may adjust the rate once per two RTD delay units of time. Any change in congestion condition on the virtual connection due to a rate adjustment is detected in the average value measurement starting one RTD unit later.
Figure 16 is a flow diagram that describes the per virtual connection rate (bandwidth) change process 500 by which data rate changes imposed on cell transmitter 220 are adjusted by frame receive controller 201. At step 500, the itø* bursty channel bandwidth, Ni, is initialized by setting the transmission rate to the quiescent rate, QIRj. Step 504 checks to see if the queue has been inactive for a period of time greater than TQ, a prescribed configuration parameter. But because the channel has just been activated, the process passes on to step 507. Otherwise, it would go to step 506 where the quiescent rate, QIRjfis assigned. If no rate change has been received, test step 507 moves the process back to step 504, forming a wait loop. If a rate change moves the process to step 508 where it is determined if it is a rate increase or decrease.
If it is a rate decrease, step 516 determines if it is a fast decrease due to cell loss at the other terminal FRP 59. If so, the process moves to step 518 where the rate is reduced by a factor of 1/2, otherwise to step 520 where it is reduced by a moderate factor of 7/8. After either reduction, the process moves to step 522 where it is determined if the reduction resulted in a rate less than the guaranteed minimum, MIRj. If so, the rate is set at MIRj, or otherwise left unchanged and the process moves back to step 504.
If the rate change in step 507 is an increase, the process moves to step 509 where it is determined if data is available, as indicated the boolean flag Qj. If not, no rate increase is made and the process goes back to step 504. If data is available, step 510 increases Nj by the linear increment of 1/10 of MIRj and then checks (step 512) if the rate exceeds the maximum allowable, PIRj. If so, the rate is set at PIRj, in step 514. In either case, the process returns to step 504.
The credit manager function, resident in cell transmitter 220 determines cell receiver 210's per channel output rate by assigning credits in accordance with the state of congestion and data availability. In the four V.35 port embodiment, each channel is serviced round-robin. The relative priority given to each (up to 252) virtual connections is determined by the bandwidth assignment algorithm in conjunction with the credit manager servicing algorithm as shown in the flow diagram of Figure 17. Step 600 initializes the credit process by setting the ith virtual circuit (VC) connection's credit, Ti, equal to Di, where Di = N/Ni or the number of tick intervals, T, that must elapse for VC connection i to acquire access to the cell network. Thus, Di is derived from the current value of Ni generated by bandwidth assignment process 500. Step 602 sets the VC index 1 = 1. The corresponding ith interval, Ti, is decremented by one tick interval, T, in step 604. Test step 606 checks to see if the resulting value of Ti is equal to or less than zero, indicating that a credit should not be added to the ilti credit index, Ci, and if so, passes on to step 616, where the round-robin procedure is initiated by incrementing index i. Otherwise, step 608 is invoked by crediting (incrementing) Ci. Step 610 checks if the incremented value of Ci exceeds the upper limit, Cmax, and if so, moves to step 612 where Ci is set equal to Cmax. The process moves on to step 614. Step 614 restores a positive non-zero value to Ti by adding Di to the value obtained from step 604. (In this manner, non- integer values of Ti are accommodated without the loss in precision that would result if Ti were to be rounded to the nearest integer value.) Step 616 leads to test step 618 to see if all of the VC connections have been served, and if so, a wait of one tick period, T, is imposed by step 620 • before repeating the process by returning to step 602. Otherwise, the process returns to step 604.
For another embodiment, the wait period, T, may be extended by some multiplicative factor greater than 1 so that the credit manager does not need to accrue credits as often. For example, if a wait of ten tick periods, 10T, is invoked at step 620, complete round-robin servicing by the credit manager would typically occur every millisecond for T = 100 us. This would require that step 604 decrement Ti by 10 (Ti = Ti - 10), and step 608 increment by 10 (Ci = Ci + 10).
Figure 18 shows the flow diagram for servicing of muxbus cell transmit queue 236 by cell transmitter 220. Step 700 initializes the VC queue index by setting i = 1. Step 702 services the ith channel cell transmit queue 236 if Ci > 1 and Qi = 1 , indicating that data is present in frame buffers 237 at the location indicated by VC queue 236. If not, the process moves to step 706. Step 704 decrements the credit count, Ci, indicating that the queue has been serviced. Step 706 movements the VC index and passes to test step 708 that checks if all VC connections have been attended and if so, passes back to the initial 700 steps. Otherwise, the remaining VC connections are attended by returning to step 702.
In the foregoing specification, the invention has been described with reference to specific exemplary embodiments thereof. It will, however, be clear that various modifications and changes may be made thereto without departing from the broader spirit and scope of the invention as set forth in the appended claims. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.

Claims

CLAIMS What is claimed is:
1. A feedback control system for congestion prevention control on a virtual connection of a cell switching communications network handling bursty traffic, comprising: a) a source node for accepting, queueing, and buffering user framed bursty data, for formatting the bursty data into bursty data cells, and for transmitting the bursty data cells over the virtual circuit connection through the cell switching communications network at an incrementally variable controllable transmission rate, wherein the source node and any intermediate node comprising the virtual connection include means for accepting, queueing, and buffering the bursty data cells, means for detecting incipient congestion in the virtual connection queues and buffers, and means for forwarding the bursty data cells over the virtual connection together with an associated incipient congestion indicator; and b) a destination node, which is a terminus of the virtual connection, for accepting the bursty data cells and the associated incipient congestion indicator, for reconstructing, formatting, and outputting the framed bursty data, for monitoring the congestion indicator, for creating a rate control signal based on the congestion indicator, for transmitting the rate control signal back to the source node, wherein the incrementally variable controllable transmission rate of the source node is adjusted in response to the rate control signal sent from the destination node in order to minimize congestion.
2. The feedback control system of claim 1 , wherein the source node further comprises means for accepting data and control cells from the destination node using a same virtual connection between the source node and the destination node.
3. The feedback control system of claim 1 wherein the rate control signal is transmitted in a cell header by the destination node over a same virtual connection between the source node and the destination node.
-4. The feedback control system of claim 3, wherein the rate control signal is transmitted in a supervisory cell header if no return cell traffic is available.
5. The feedback control system of claim 1 , further comprising means for measuring virtual connection round-trip delay times, wherein the measuring means comprise: a) destination node means for generating a high priority (HP) time-stamped cell and for transmitting the HP timestamped cell to the source node; b) source node means for echoing HP timestamped cell back to the destination node; c) means for comparing a time-stamp of the HP timestamped cell with a destination node arrival time to determine round-trip delay time.
6. The feedback control system of claim 1 , wherein the source node further comprises: a) cell transmitter rate controller that permits a queued bursty data cell to be transmitted only if an accumulated transmission credit is greater than zero and, if transmission is permitted, decrements the accumulated transmission credit. b) credit manager for incrementing the accumulated transmission credit each time a prescribed interval of time has elapsed, wherein the prescribed time interval corresponds to a clock period of the assigned cell rate, wherein the accumulated transmission credit is sent to the cell transmitter rate controller.
7. The feedback control system of claim 6, wherein the credit manager further comprises: a) means for accepting a maximum credit count configuration parameter; and b) means for limiting the accumulated transmission credit to the maximum credit count configuration parameter.
8. The feedback control system of claim 1 , wherein the nodes are two-way nodes using a common virtual connection between source and destination nodes.
9. A feedback control system for congestion prevention control on a virtual connection of a cell switching communications network handling bursty traffic, comprising: a) a source node for accepting, queueing, and buffering user framed bursty data, for formatting the bursty data into bursty data cells, and for transmitting the cells over the virtual circuit connection through the cell switching communications network at an incrementally variable controllable transmission rate wherein the source node and any intermediate node comprising the virtual connection include means for accepting, queueing, and buffering the bursty data cells, means for detecting incipient congestion in the virtual connection queues and buffers, and means for forwarding the bursty data cells over the virtual connection together with an associated incipient congestion indicator; and - b) a destination node, being the terminus of the communication network virtual connection for accepting the bursty data cells and the associated incipient congestion indicator, for reconstructing, formatting, and outputting the framed bursty data, for monitoring the congestion indicators, for creating a four state, two binary bit, rate control signal representing rate increase, rate decrease, fast rate decrease, and no-rate-change messages based on the congestion indicators, transmitting the rate control signal back to the source node, wherein the incrementally controllable transmission rate of the source node is adjusted in response to the rate control signal sent from the destination node in order to minimize congestion.
10. The feedback control system of claim 9 wherein the means for incrementally controlling the transmission rate is responsive to the four state control signal causing an additive fractional increase in response to a rate increase signal, a large fractional multiplicative factor decrease in response to a rate decrease signal, and a small fractional multiplicative factor rate decrease in response to a fast rate decrease signal.
11. A feedback control system for congestion prevention control on a virtual connection of a cell switching communications network handling bursty traffic, comprising: a) a source node for accepting, queueing, and buffering user framed bursty data, for formatting the bursty data into bursty data cells, and for transmitting the cells over the virtual circuit connection through the cell switching communications network at an incrementally variable controllable transmission rate that is variable between a prescribed lower guaranteed minimum rate and an upper maximum rate, the source node and any intermediate node comprising the virtual connection include means for accepting, queueing and buffering the bursty data cells, means for detecting incipient congestion in the virtual connection queues and buffers, and means for forwarding the cells over the network virtual connection together with an associated incipient congestion indicator; and b) a destination node being the terminus of the communication network virtual connection having means for accepting the bursty data cells and the associated incipient congestion indicator, for reconstructing, formatting and outputting the framed bursty data, for monitoring the congestion indicators, for creating a rate control signal based on the congestion indicators, for transmitting the rate control signal back to the source node, wherein the incrementally controllable transmission rate of the source node is adjusted in response to the rate control signal sent from the destination node in order to minimize congestion.
12. The feedback control system of claim 11 , wherein the rate control further comprises means for assigning an initial quiescent rate between the lower and upper rate.
13. The feedback control system of claim 12, wherein the rate control further comprises means for sensing lack of user input bursty data activity and resetting the rate to the quiescent rate.
14. The feedback control system of claim 11 , wherein the rate control means ignores the rate increase signal if the user input rate does not create an excess virtual circuit queue load.
15. A feedback control system for congestion prevention control on a virtual connection of a ceil switching communications network handling bursty traffic, comprising: a) a source node for accepting, queueing and buffering user framed bursty data, for formatting the bursty data into bursty data cells, and for transmitting the cells over the virtual circuit connection through the cell switching communications network at an incrementally variable controllable transmission rate, the source node and any intermediate node comprising the virtual connection include means for accepting, queueing and buffering the bursty data cells, means for detecting incipient congestion in the virtual connection queues and buffers by monitoring the average queue length over a regeneration cycle time interval, means for comparing the average queue length to a preset threshold, means for generating an associated incipient congestion indicator when threshold is exceeded, and means for forwarding the ceils over the network virtual connection together with the associated incipient congestion indicator; and b) a destination node being the terminus of the communication network virtual connection having means for accepting the bursty data cells and the associated incipient congestion indicator, for reconstructing, formatting, and outputting the framed bursty data, for monitoring the congestion indicators, for creating a rate control signal based on the congestion indicators, for transmitting the rate control signal back to the source node, wherein the incrementally controllable transmission rate of the source node is adjusted in response to the rate control signal sent from the destination node in order to minimize congestion.
16. The feedback control system of claim 15, wherein the incipient congestion threshold is set to provide a tolerable delay in combination with a tolerable concomitant cell loss rate.
17. The feedback system of claim 15, further comprising means for discarding cells that arrive while queue length is exceeded.
18. The feedback control system of claim 15, wherein the incipient congestion control indicator is encoded in the associated cell header.
19. A feedback control system for congestion prevention control on a virtual connection of a cell switching communications network handling bursty traffic, comprising: a) a source node for accepting, queueing and buffering user framed bursty data, for formatting the bursty data into bursty data cells, and for transmitting the cells over the virtual circuit connection through the cell switching communications network at an incrementally variable controllable transmission rate, the source node and any intermediate node comprising the virtual connection having means for accepting, queueing and buffering the bursty data cells, means for detecting incipient congestion in the virtual connection queues and buffers, and means for forwarding the cells over the network virtual connection together with an associated incipient congestion indicator; and b) a destination node being the terminus of the communication network virtual connection having means for accepting the bursty data cells and the associated incipient congestion indicator, for reconstructing, formatting and outputting the framed bursty data, for counting cells with the associated incipient congestion indicator asserted, for creating a rate control signal based on the cell count, for transmitting the rate control signal back to the source node, wherein the incrementally controllable transmission rate of the source node is adjusted in response to the rate control signal sent from the destination node in order to minimize congestion.
20. The feedback control system of claim 19, wherein the cell count means begins counting after a round-trip delay interval after the last rate control signal transmission from the destination node to the source node.
21. The feedback control system of claim 19, wherein the congestion monitoring means further comprises means for programmable threshold level detection and means for generating a rate control signal indicating a rate increase if the threshold is not exceeded by the average incipient congestion count, and for generating a decrease rate control signal if the threshold is exceeded, and means for transmitting the rate control signal to the source node.
22. The feedback control system of claim 19, wherein the control signal generation means further comprises means for detecting cell loss together with the detection of the average incipient congestion count exceeding threshold, means for generating a fast rate reduction signal when both conditions prevail, and means for transmitting the control signal to the source node.
23. The feedback control system of claim 19, wherein the control signal generation means generates a no-rate-change signal for each cell while waiting for the effects on congestion due to the previous rate change control signal.
24. A feedback control system for congestion prevention control on a virtual connection of a cell switching communications network handling bursty traffic, comprising: a) a source node for accepting, queueing and buffering user framed bursty data, for formatting the bursty data into bursty data cells, for indicating those cells derived from user input data received at rates exceeding a prescribed guaranteed minimum rate, and for transmitting the cells together with the indicators over the virtual circuit connection through the cell switching communications network at an incrementally variable controllable transmission rate, the source node and any intermediate node comprising the virtual connection having means for accepting, queueing and buffering the bursty data cells, means for detecting incipient congestion in the virtual connection queues and buffers, and means for forwarding the cells over the network virtual connection together with an associated incipient congestion indicator; and b) a destination node being the terminus of the communication network virtual connection having means for accepting the bursty data cells and the associated incipient congestion indicator, for reconstructing, formatting and outputting the framed bursty data, means for monitoring the congestion indicators, for creating a rate control signal based on the congestion indicators, means for transmitting the rate control signal back to the source node, wherein the incrementally controllable transmission rate of the source node is adjusted in response to the rate control signal sent from the destination node in order to minimize congestion.
25. The feedback control system of claim 24, wherein the means for indicating those cells derived from user input frames rates exceeding a prescribed guaranteed minimum rate further comprises means for encoding the indications in the cell header.
26. The feedback control system of claim 24, wherein the source node and any intermediate node further comprises means for discarding the cells which have been indicated as being derived from input data exceeding the minimum guaranteed rate whenever the virtual connection queues exceed a prescribed cell loss priority threshold length.
27. The feedback control system of claim 26, wherein the cell discarding means discards all subsequent cells, belonging substantially to the same data frame as that which was being queued when the priority threshold was exceeded.
28. A feedback control system for congestion prevention control on a virtual connection of a cell switching communications network handling bursty traffic, comprising: a) a source node for accepting, queueing and buffering user framed bursty data including an associated expendable data indicator that indicates an excess input frame data rate, for formatting the bursty data into bursty data cells, for encoding the expendable data indicator in the cell header, for discarding cells bearing the indicator for relief of congestion, and for transmitting the non-discarded cells over the virtual circuit connection through a cell switching communications network at an incrementally variable controllable transmission rate, the source node and any intermediate nodes comprising the virtual connection having means for accepting, queueing and buffering the bursty data cells, means for detecting incipient congestion in the virtual connection queues and buffers, and means for forwarding the cells over the network virtual connection together with an associated incipient congestion indicator; and - b) a destination node being the terminus of the communication network virtual connection having means for accepting the bursty data cells and the associated incipient congestion indicator, for reconstructing, formatting and outputting the framed bursty data, for monitoring the congestion indicators, for creating a rate control signal based on the congestion indicators, means for transmitting the rate control signal back to the source node, wherein the incrementally controllable transmission rate of the source node is adjusted in response to the rate control signal sent from the destination node in order to minimize congestion.
29. The feedback control system of claim 28, wherein the source node and any intermediate node further comprises means for discarding cells associated with the expendable data indicator.
30. The feedback control system of claim 28, wherein the destination node further comprises: a) at least one output queue for buffering reconstructed data frames; b) means for monitoring the output queue length; c) means for comparing the output queue length with a prescribed threshold; and d) means for discarding the reconstructed bursty data frames having an associated asserted expendable data indicator whenever the prescribed output queue threshold is exceeded.
31. The feedback control system of claim 30, further comprising: a) means for calculating an average queue length of the port transmit queue associated with the virtual connection; b) means for comparing the average output port transmit queue length with a prescribed average output port transmit queue length threshold; and c) means for generating a rate decrease signal if the average output port transmit queue length threshold is exceeded.
32. The feedback control system of claim 31 , wherein the average output queue length is computed over a regeneration cycle time interval.
33. A feedback control method for congestion prevention on a virtual circuit connection of a cell switching network handling bursty data traffic, comprising the steps of: a) queuing and buffering user bursty data frames at the input source node; b) formatting the bursty data frames into bursty data cells; c) queuing and buffering the bursty data cells within the source node and any intermediate nodes, for transmission to the next node comprising the virtual connection; d) detecting incipient congestion of the bursty data cells that are buffered and queued for transmission to the next node by monitoring queue lengths; e) generating an incipient congestion indicator when the queue lengths are excessive; f) associating the congestion indicator with corresponding cells that are contributing to the incipient congestion; g) transmitting the cells and associated congestion indicators to the destination node over a virtual circuit connection at a controllable rate; h) incrementally controlling the transmission rate of bursty cells in response to the congestion indicators so as to control congestion; i) reconstructing frames at the destination node from received bursty cells; j) monitoring the congestion indicators at the destination node for determining the congestion status of the virtual connection; k) creating a rate control signal based on the congestion status; and
I) transmitting the rate control signal to the source node for incremental cell rate transmission control in order to minimize congestion.
34. The feedback control method of claim 33, wherein the step of creating a rate control signal based on the congestion status comprises generating a four state, two binary bit, signal representing rate increase, rate decrease, fast rate decrease, and no-rate-change.
35. The feedback control method of claim 34, wherein the step of incrementally controlling the transmission of cells and associated indicators is responsive to the four state control signal causing an additive fractional increase in response to rate increase signal, a moderate fractional multiplicative factor decrease in response to a rate decrease signal, a large fractional multiplicative factor decrease in response to a fast decrease signal, and no rate change in response to a no-rate-change signal.
36. The feedback control method of claim 35, wherein the step of incrementally controlling the transmission of cells and associated indicators varies the transmission rate between a lower guaranteed minimum rate and an upper maximum rate.
37. The feedback control method of claim 36 wherein the step of incrementally controlling the transmission of cells and associated indicators assigns an initial rate between the lower and upper rate.
38. The feedback control method of claim 36, wherein the step of controlling the transmission of cells and associated indicators provides for sensing lack of user input bursty data activity, causing the rate to be reset to a prescribed quiescent rate.
39. The feedback control method of claim 36, wherein the step of incrementally controlling the transmission of cells and associated indicators ignores the rate increase signal if the user input rate does not create an excess virtual circuit queue load.
40. The feedback control method of claim 36, wherein the step of incrementally controlling the transmission of cells and associated indicators further comprises the following steps: a) determining the assigned inter-cell interval corresponding to the assigned virtual circuit transmission rate expressed in queue server tick intervals; b) decrementing the assigned interval each base rate period; c) incrementing a transmission credit counter whenever the decremented assigned interval is equal to or less than zero, and renewing the decremented assigned interval by adding to it the prior assigned interval thereby creating a new assigned interval for use in step (b); d) transmitting a cell at queue server tick intervals whenever the transmission credit count is greater than zero and a cell is queued for transmission; and e) decrementing the credit counter whenever a cell is transmitted.
41. The feedback control method of claim 40, further comprising the step of limiting the transmission credit counter to a prescribed maximum count.
-42. The feedback control method of claim 33, further comprising the following additional steps:
(a) monitoring the average length of any output port transmit queue associated with the bursty cell virtual connection;
(b) comparing the average queue length to a configurable queue length threshold value; and
(c) generating a rate decrease signal if the threshold level is exceeded.
PCT/US1993/002097 1992-04-01 1993-03-08 Congestion control for cell networks WO1993020637A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP93907328A EP0634068A1 (en) 1992-04-01 1993-03-08 Congestion control for cell networks
JP5517450A JPH07508378A (en) 1992-04-01 1993-03-08 Congestion control for cell networks

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US07/861,761 US5313454A (en) 1992-04-01 1992-04-01 Congestion control for cell networks
US07/861,761 1992-04-01

Publications (1)

Publication Number Publication Date
WO1993020637A1 true WO1993020637A1 (en) 1993-10-14

Family

ID=25336688

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1993/002097 WO1993020637A1 (en) 1992-04-01 1993-03-08 Congestion control for cell networks

Country Status (5)

Country Link
US (1) US5313454A (en)
EP (1) EP0634068A1 (en)
JP (1) JPH07508378A (en)
AU (1) AU3797093A (en)
WO (1) WO1993020637A1 (en)

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2281005A (en) * 1993-06-21 1995-02-15 Plessey Telecomm Network management to avoid congestion
EP0658999A2 (en) * 1993-12-15 1995-06-21 Nec Corporation Traffic control for ATM networks
EP0706297A1 (en) * 1994-10-07 1996-04-10 International Business Machines Corporation Method for operating traffic congestion control in a data communication network and system for implementing said method
WO1997009814A1 (en) * 1995-09-07 1997-03-13 Ericsson Australia Pty. Ltd. Controlling traffic congestion in intelligent electronic networks
US5633861A (en) * 1994-12-19 1997-05-27 Alcatel Data Networks Inc. Traffic management and congestion control for packet-based networks
US5726977A (en) * 1994-12-22 1998-03-10 Korea Telecommunication Authority Apparatus and method for determining a network node congestion state in order to control the congestion in an ATM network
WO2004002085A1 (en) * 2002-06-20 2003-12-31 Telefonaktiebolaget Lm Ericsson (Publ) Apparatus and method for resource allocation
WO2008005696A1 (en) * 2006-06-30 2008-01-10 Intel Corporation End-point based tamper resistant congestion management
US7570584B1 (en) 2002-03-29 2009-08-04 Cisco Technology, Inc. Network-wide congestion control of SPVC signaling messages
WO2011022095A1 (en) * 2009-08-19 2011-02-24 Opanga Networks, Inc Enhanced data delivery based on real time analysis of network communications quality and traffic
US7969979B2 (en) 2003-09-29 2011-06-28 Runcom Technologies Ltd. Distribution of multicast data to users
US8161182B1 (en) 2000-01-26 2012-04-17 Cisco Technology, Inc. Managing network congestion using dynamically advertised congestion status
US8886790B2 (en) 2009-08-19 2014-11-11 Opanga Networks, Inc. Systems and methods for optimizing channel resources by coordinating data transfers based on data type and traffic
US8909807B2 (en) 2005-04-07 2014-12-09 Opanga Networks, Inc. System and method for progressive download using surplus network capacity
US9065595B2 (en) 2005-04-07 2015-06-23 Opanga Networks, Inc. System and method for peak flow detection in a communication network
US9143341B2 (en) 2008-11-07 2015-09-22 Opanga Networks, Inc. Systems and methods for portable data storage devices that automatically initiate data transfers utilizing host devices
WO2018004412A1 (en) * 2016-07-01 2018-01-04 Telefonaktiebolaget Lm Ericsson (Publ) Round trip time skew control methods and arrangements
US11258531B2 (en) 2005-04-07 2022-02-22 Opanga Networks, Inc. System and method for peak flow detection in a communication network

Families Citing this family (563)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5167035A (en) * 1988-09-08 1992-11-24 Digital Equipment Corporation Transferring messages between nodes in a network
US6847611B1 (en) 1990-12-10 2005-01-25 At&T Corp. Traffic management for frame relay switched data service
US5809012A (en) * 1991-01-31 1998-09-15 Fujitsu Limited Connectionless communication system
JPH0548638A (en) * 1991-08-15 1993-02-26 Fujitsu Ltd End to end congestion coping control processing system in atm network
US5734926A (en) * 1992-07-15 1998-03-31 Advanced Hardware Architectures Direct memory access controller in an integrated circuit
JPH06169320A (en) * 1992-10-02 1994-06-14 Toshiba Corp Atm cell making device
FR2699359B1 (en) * 1992-12-16 1995-02-24 Thierry Houdoin Method for simulating transmission on a transmission network by asynchronous transfer and transmission simulator on such a network.
FR2702113B1 (en) * 1993-02-26 1995-04-28 Alcatel Business Systems Asynchronous switching network for communication installation core.
US6771617B1 (en) 1993-06-17 2004-08-03 Gilat Satellite Networks, Ltd. Frame relay protocol-based multiplex switching scheme for satellite mesh network
US5434850A (en) 1993-06-17 1995-07-18 Skydata Corporation Frame relay protocol-based multiplex switching scheme for satellite
US5448567A (en) * 1993-07-27 1995-09-05 Nec Research Institute, Inc. Control architecture for ATM networks
US5598581A (en) * 1993-08-06 1997-01-28 Cisco Sytems, Inc. Variable latency cut through bridge for forwarding packets in response to user's manual adjustment of variable latency threshold point while the bridge is operating
US5457687A (en) * 1993-09-02 1995-10-10 Network Equipment Technologies, Inc. Method and apparatus for backward explicit congestion notification (BECN) in an ATM network
JP3187230B2 (en) * 1993-09-06 2001-07-11 株式会社東芝 Congestion control method and congestion control device
ZA946674B (en) * 1993-09-08 1995-05-02 Qualcomm Inc Method and apparatus for determining the transmission data rate in a multi-user communication system
WO1995008805A1 (en) * 1993-09-20 1995-03-30 Codex, Inc., A Subsidiary Company Of Motorola Inc. Distribution of network management communication via frame relay
US5557724A (en) * 1993-10-12 1996-09-17 Intel Corporation User interface, method, and apparatus selecting and playing channels having video, audio, and/or text streams
US6279029B1 (en) * 1993-10-12 2001-08-21 Intel Corporation Server/client architecture and method for multicasting on a computer network
US5446726A (en) * 1993-10-20 1995-08-29 Lsi Logic Corporation Error detection and correction apparatus for an asynchronous transfer mode (ATM) network device
FR2711466B1 (en) * 1993-10-20 1996-01-26 Tremel Jean Yves ATM type telecommunication network test and ATM connection performance measurement.
DE69324274T2 (en) * 1993-10-23 1999-10-14 Ibm SELECTIVE OVERLOAD CONTROL FOR INFORMATION NETWORKS
US5901140A (en) * 1993-10-23 1999-05-04 International Business Machines Corporation Selective congestion control mechanism for information networks
JPH07162435A (en) * 1993-12-06 1995-06-23 Fujitsu Ltd Method and device for confirming semi-fixed path
JP3307085B2 (en) * 1993-12-10 2002-07-24 ソニー株式会社 Communication method and electronic equipment
US5574861A (en) * 1993-12-21 1996-11-12 Lorvig; Don Dynamic allocation of B-channels in ISDN
US5497375A (en) * 1994-01-05 1996-03-05 Motorola, Inc. Device and method for ATM end system cell flow regulation
US5495426A (en) * 1994-01-26 1996-02-27 Waclawsky; John G. Inband directed routing for load balancing and load distribution in a data communication network
US5617409A (en) * 1994-01-28 1997-04-01 Digital Equipment Corporation Flow control with smooth limit setting for multiple virtual circuits
US5579301A (en) * 1994-02-28 1996-11-26 Micom Communications Corp. System for, and method of, managing voice congestion in a network environment
US5784358A (en) * 1994-03-09 1998-07-21 Oxford Brookes University Broadband switching network with automatic bandwidth allocation in response to data cell detection
US5450394A (en) * 1994-03-10 1995-09-12 Northern Telecom Limited Delay monitoring of telecommunication networks
US5434860A (en) * 1994-04-20 1995-07-18 Apple Computer, Inc. Flow control for real-time data streams
US5583861A (en) * 1994-04-28 1996-12-10 Integrated Telecom Technology ATM switching element and method having independently accessible cell memories
US6151301A (en) * 1995-05-11 2000-11-21 Pmc-Sierra, Inc. ATM architecture and switching element
FI98774C (en) * 1994-05-24 1997-08-11 Nokia Telecommunications Oy Method and apparatus for prioritizing traffic in an ATM network
US5475682A (en) * 1994-06-10 1995-12-12 At&T Corp. Method of regulating backpressure traffic in a packet switched network
JP3309020B2 (en) * 1994-06-20 2002-07-29 富士通株式会社 Transmission throughput control apparatus and method
US5475813A (en) * 1994-07-18 1995-12-12 International Business Machines Corporation Routing transactions in the presence of failing servers
US5799190A (en) * 1994-08-09 1998-08-25 Apple Computer, Inc. Communications coprocessor for realizing bidirectional isochronous communications link with host computer
US5793978A (en) * 1994-12-29 1998-08-11 Cisco Technology, Inc. System for routing packets by separating packets in to broadcast packets and non-broadcast packets and allocating a selected communication bandwidth to the broadcast packets
US5867666A (en) * 1994-12-29 1999-02-02 Cisco Systems, Inc. Virtual interfaces with dynamic binding
CA2160393C (en) * 1994-12-30 1999-05-04 Anwar Elwalid Method for integrated congestion control in networks
CN1105441C (en) * 1995-02-23 2003-04-09 Ntt移动通信网株式会社 Method and device for multi-cell transmission
US5854898A (en) 1995-02-24 1998-12-29 Apple Computer, Inc. System for automatically adding additional data stream to existing media connection between two end points upon exchange of notifying and confirmation messages therebetween
JPH08307420A (en) * 1995-03-03 1996-11-22 Fujitsu Ltd Congestion control system for cell exchange
AU720173B2 (en) 1995-03-08 2000-05-25 British Telecommunications Public Limited Company Broadband switching network
NZ303240A (en) * 1995-03-08 1999-06-29 British Telecomm Bandwidth allocation depending on load for broadband switching system
CA2214838C (en) 1995-03-08 2008-02-26 Avril Joy Smith Broadband switching system
US5953350A (en) 1995-03-13 1999-09-14 Selsius Systems, Inc. Multimedia client for multimedia/hybrid network
US5570360A (en) * 1995-03-20 1996-10-29 Stratacom, Inc. Method and apparatus for implementing communication service contract using cell arrival information
JP3606941B2 (en) * 1995-03-23 2005-01-05 株式会社東芝 Flow control apparatus and flow control method
US5586121A (en) * 1995-04-21 1996-12-17 Hybrid Networks, Inc. Asymmetric hybrid access system and method
US5740241A (en) * 1995-05-12 1998-04-14 Carrier Access Corporation T1 channel bank control process and apparatus
US5574724A (en) * 1995-05-26 1996-11-12 Lucent Technologies Inc. Adjustment of call bandwidth during a communication call
US5701295A (en) * 1995-05-26 1997-12-23 Lucent Technologies Inc. Variable communication bandwidth for providing automatic call back and call hold
JP3062041B2 (en) * 1995-05-29 2000-07-10 日本電気株式会社 Congestion control method in asynchronous transfer networks
GB9511319D0 (en) * 1995-06-05 1995-08-02 Gen Datacomm Adv Res Controlling the flow of ATM cells in an ATM network
EP1686742B1 (en) * 1995-06-05 2008-10-01 NEC Electronics Corporation Communiction control device and method for an ATM system applicable to an ABR mode
US5675576A (en) * 1995-06-05 1997-10-07 Lucent Technologies Inc. Concestion control system and method for packet switched networks providing max-min fairness
US5636371A (en) * 1995-06-07 1997-06-03 Bull Hn Information Systems Inc. Virtual network mechanism to access well known port application programs running on a single host system
US5734865A (en) * 1995-06-07 1998-03-31 Bull Hn Information Systems Inc. Virtual local area network well-known port routing mechanism for mult--emulators in an open system environment
US5838915A (en) * 1995-06-21 1998-11-17 Cisco Technology, Inc. System for buffering data in the network having a linked list for each of said plurality of queues
JPH098814A (en) * 1995-06-21 1997-01-10 Fuji Xerox Co Ltd Data transmission device and method thereof and data communication system
GB9514206D0 (en) 1995-07-12 1995-09-13 British Telecomm Rate controller
JPH0927811A (en) * 1995-07-13 1997-01-28 Fujitsu Ltd Congestion supervisory controller
JPH11510004A (en) 1995-07-19 1999-08-31 フジツウ ネットワーク コミュニケーションズ,インコーポレイテッド Point-to-multipoint transmission using subqueues
US6097718A (en) 1996-01-02 2000-08-01 Cisco Technology, Inc. Snapshot routing with route aging
US6147996A (en) 1995-08-04 2000-11-14 Cisco Technology, Inc. Pipelined multiple issue packet switch
US5793759A (en) * 1995-08-25 1998-08-11 Terayon Corporation Apparatus and method for digital data transmission over video cable using orthogonal cyclic codes
US6356555B1 (en) 1995-08-25 2002-03-12 Terayon Communications Systems, Inc. Apparatus and method for digital data transmission using orthogonal codes
US6307868B1 (en) 1995-08-25 2001-10-23 Terayon Communication Systems, Inc. Apparatus and method for SCDMA digital data transmission using orthogonal codes and a head end modem with no tracking loops
US5805583A (en) * 1995-08-25 1998-09-08 Terayon Communication Systems Process for communicating multiple channels of digital data in distributed systems using synchronous code division multiple access
US5768269A (en) * 1995-08-25 1998-06-16 Terayon Corporation Apparatus and method for establishing frame synchronization in distributed digital data communication systems
US5991308A (en) * 1995-08-25 1999-11-23 Terayon Communication Systems, Inc. Lower overhead method for data transmission using ATM and SCDMA over hybrid fiber coax cable plant
US5745837A (en) * 1995-08-25 1998-04-28 Terayon Corporation Apparatus and method for digital data transmission over a CATV system using an ATM transport protocol and SCDMA
US6665308B1 (en) 1995-08-25 2003-12-16 Terayon Communication Systems, Inc. Apparatus and method for equalization in distributed digital data transmission systems
WO1997010656A1 (en) 1995-09-14 1997-03-20 Fujitsu Network Communications, Inc. Transmitter controlled flow control for buffer allocation in wide area atm networks
DE19534940C2 (en) * 1995-09-20 1998-07-02 Siemens Ag Method for recognizing overload situations in subscriber line modules of a communication system
US6917966B1 (en) 1995-09-29 2005-07-12 Cisco Technology, Inc. Enhanced network services using a subnetwork of communicating processors
US6182224B1 (en) 1995-09-29 2001-01-30 Cisco Systems, Inc. Enhanced network services using a subnetwork of communicating processors
US7246148B1 (en) 1995-09-29 2007-07-17 Cisco Technology, Inc. Enhanced network services using a subnetwork of communicating processors
US6122279A (en) * 1995-10-02 2000-09-19 Virata Limited Asynchronous transfer mode switch
EP1457896A3 (en) 1995-10-26 2005-04-20 Matsushita Electric Industrial Co., Ltd. File system
US5546377A (en) * 1995-10-31 1996-08-13 Digital Equipment Corporation Efficient distributed method for computing max-min fair rates of a limited resource in ATM networks
US5687324A (en) * 1995-11-08 1997-11-11 Advanced Micro Devices, Inc. Method of and system for pre-fetching input cells in ATM switch
US6075768A (en) 1995-11-09 2000-06-13 At&T Corporation Fair bandwidth sharing for video traffic sources using distributed feedback control
US5917822A (en) * 1995-11-15 1999-06-29 Xerox Corporation Method for providing integrated packet services over a shared-media network
US5699346A (en) * 1995-11-17 1997-12-16 Telecommunications Techniques Corporation Measuring burst rate and burst size in ATM network virtual connections
DE19544976A1 (en) * 1995-12-01 1997-06-05 Siemens Ag Device and method for communication between transmitting and / or receiving stations and a switching station
US5751969A (en) * 1995-12-04 1998-05-12 Motorola, Inc. Apparatus and methods for predicting and managing congestion in a network
US5701292A (en) * 1995-12-04 1997-12-23 Lucent Technologies Inc. Method and apparatus for controlling data transfer rates of data sources in asynchronous transfer mode-based networks
US6108530A (en) 1995-12-14 2000-08-22 Lucent Technologies Inc. System and method for transmitting a displayable message between short message entities in more than one data package
US6091725A (en) * 1995-12-29 2000-07-18 Cisco Systems, Inc. Method for traffic management, traffic prioritization, access control, and packet forwarding in a datagram computer network
US5777987A (en) * 1995-12-29 1998-07-07 Symbios, Inc. Method and apparatus for using multiple FIFOs to improve flow control and routing in a communications receiver
US6035105A (en) 1996-01-02 2000-03-07 Cisco Technology, Inc. Multiple VLAN architecture system
AU709365B2 (en) 1996-01-09 1999-08-26 British Telecommunications Public Limited Company A service multiplexer
US5765032A (en) * 1996-01-11 1998-06-09 Cisco Technology, Inc. Per channel frame queuing and servicing in the egress direction of a communications network
AU1697697A (en) 1996-01-16 1997-08-11 Fujitsu Limited A reliable and flexible multicast mechanism for atm networks
JP2929991B2 (en) * 1996-01-29 1999-08-03 日本電気株式会社 Optimization credit control method
US5724358A (en) * 1996-02-23 1998-03-03 Zeitnet, Inc. High speed packet-switched digital switch and method
US5920561A (en) * 1996-03-07 1999-07-06 Lsi Logic Corporation ATM communication system interconnect/termination unit
US5982749A (en) * 1996-03-07 1999-11-09 Lsi Logic Corporation ATM communication system interconnect/termination unit
US6535512B1 (en) 1996-03-07 2003-03-18 Lsi Logic Corporation ATM communication system interconnect/termination unit
JP3686493B2 (en) * 1996-03-07 2005-08-24 富士通株式会社 Feedback control method and apparatus in ATM switch
US5841772A (en) * 1996-03-07 1998-11-24 Lsi Logic Corporation ATM communication system interconnect/termination unit
US6373846B1 (en) 1996-03-07 2002-04-16 Lsi Logic Corporation Single chip networking device with enhanced memory access co-processor
US5848068A (en) * 1996-03-07 1998-12-08 Lsi Logic Corporation ATM communication system interconnect/termination unit
JP3643637B2 (en) 1996-03-08 2005-04-27 株式会社日立コミュニケーションテクノロジー Cell output control circuit and control method
US5777984A (en) * 1996-04-01 1998-07-07 Motorola Inc. Method and apparatus for controlling cell transmission rate in a cell based network in the presence of congestion
JP3335069B2 (en) * 1996-04-11 2002-10-15 富士通株式会社 Fixed-length cell multiplex transmission device, fixed-length cell multiplex transmission method, fixed-length cell transmission device, and fixed-length cell multiplex device
US6973034B1 (en) * 1999-06-29 2005-12-06 Cisco Technology, Inc. Technique for collecting operating information from network elements, and for controlling network element behavior in a feedback-based, adaptive data network
US5864539A (en) * 1996-05-06 1999-01-26 Bay Networks, Inc. Method and apparatus for a rate-based congestion control in a shared memory switch
EP1021765A4 (en) * 1996-05-10 2005-06-08 Fujitsu Network Communications Method and apparatus for enabling flow control over multiple networks having disparate flow control capability
US6034945A (en) 1996-05-15 2000-03-07 Cisco Technology, Inc. Method and apparatus for per traffic flow buffer management
US6308148B1 (en) * 1996-05-28 2001-10-23 Cisco Technology, Inc. Network flow data export
US6243667B1 (en) 1996-05-28 2001-06-05 Cisco Systems, Inc. Network flow switching and flow data export
EP0810808B1 (en) * 1996-05-29 2009-08-12 Nippon Telegraph And Telephone Corporation ATM cell transport equipment
US5826031A (en) * 1996-06-10 1998-10-20 Sun Microsystems, Inc. Method and system for prioritized downloading of embedded web objects
US6201790B1 (en) * 1996-06-26 2001-03-13 U.S. Philips Corporation Network access control method, and device and system for implementing this method
US6212182B1 (en) 1996-06-27 2001-04-03 Cisco Technology, Inc. Combined unicast and multicast scheduling
US6434120B1 (en) 1998-08-25 2002-08-13 Cisco Technology, Inc. Autosensing LMI protocols in frame relay networks
US6961341B1 (en) * 1996-07-02 2005-11-01 Microsoft Corporation Adaptive bandwidth throttling for network services
US5799002A (en) * 1996-07-02 1998-08-25 Microsoft Corporation Adaptive bandwidth throttling for network services
US6222856B1 (en) 1996-07-02 2001-04-24 Murali R. Krishnan Adaptive bandwidth throttling for individual virtual services supported on a network server
JP2930020B2 (en) * 1996-07-18 1999-08-03 日本電気株式会社 Communication control device
US5748905A (en) * 1996-08-30 1998-05-05 Fujitsu Network Communications, Inc. Frame classification using classification keys
US5918182A (en) * 1996-08-30 1999-06-29 Motorola, Inc. Method and apparatus for mitigating data congestion in an integrated voice/data radio communications system
US6046999A (en) 1996-09-03 2000-04-04 Hitachi, Ltd. Router apparatus using ATM switch
US6144639A (en) * 1996-09-03 2000-11-07 Sbc Technology Resources, Inc. Apparatus and method for congestion control in high speed networks
US5970229A (en) * 1996-09-12 1999-10-19 Cabletron Systems, Inc. Apparatus and method for performing look-ahead scheduling of DMA transfers of data from a host memory to a transmit buffer memory
US5941952A (en) * 1996-09-12 1999-08-24 Cabletron Systems, Inc. Apparatus and method for transferring data from a transmit buffer memory at a particular rate
US5999980A (en) * 1996-09-12 1999-12-07 Cabletron Systems, Inc. Apparatus and method for setting a congestion indicate bit in an backwards RM cell on an ATM network
US5966546A (en) 1996-09-12 1999-10-12 Cabletron Systems, Inc. Method and apparatus for performing TX raw cell status report frequency and interrupt frequency mitigation in a network node
US5867480A (en) * 1996-09-12 1999-02-02 Cabletron Systems, Inc. Method and apparatus for controlling congestion in a network node
US5922046A (en) * 1996-09-12 1999-07-13 Cabletron Systems, Inc. Method and apparatus for avoiding control reads in a network node
US5995995A (en) * 1996-09-12 1999-11-30 Cabletron Systems, Inc. Apparatus and method for scheduling virtual circuit data for DMA from a host memory to a transmit buffer memory
US5982748A (en) 1996-10-03 1999-11-09 Nortel Networks Corporation Method and apparatus for controlling admission of connection requests
US6240084B1 (en) 1996-10-10 2001-05-29 Cisco Systems, Inc. Telephony-enabled network processing device with separate TDM bus and host system backplane bus
JP2001502503A (en) 1996-10-11 2001-02-20 サーノフ コーポレイション Apparatus and method for bitstream analysis
JP3672683B2 (en) 1996-10-28 2005-07-20 富士通株式会社 Traffic control method, network system, and frame relay switch
DE19745021B4 (en) * 1996-10-29 2008-08-28 Keymile Ag Network for packet-oriented traffic
US6904037B2 (en) 1996-11-05 2005-06-07 Cisco Technology, Inc. Asymmetric implementation of DSVD for voice/data internet access
US5867483A (en) * 1996-11-12 1999-02-02 Visual Networks, Inc. Method and apparatus for measurement of peak throughput in packetized data networks
JP2001500338A (en) 1996-12-06 2001-01-09 フジツウ ネットワーク コミュニケーションズ,インコーポレイテッド Method for controlling flow of ATM traffic
US6304546B1 (en) 1996-12-19 2001-10-16 Cisco Technology, Inc. End-to-end bidirectional keep-alive using virtual circuits
US6028842A (en) * 1996-12-23 2000-02-22 Nortel Networks Corporation Dynamic traffic conditioning
US5951644A (en) * 1996-12-24 1999-09-14 Apple Computer, Inc. System for predicting and managing network performance by managing and monitoring resourse utilization and connection of network
US6201789B1 (en) * 1996-12-30 2001-03-13 Compaq Computer Corporation Network switch with dynamic backpressure per port
US6014706A (en) * 1997-01-30 2000-01-11 Microsoft Corporation Methods and apparatus for implementing control functions in a streamed video display system
JP3653569B2 (en) 1997-01-30 2005-05-25 マイクロソフト コーポレーション A VCR-like feature that renders video on demand
US5978865A (en) * 1997-02-04 1999-11-02 Advanced Micro Devices, Inc. System for performing DMA transfers where an interrupt request signal is generated based on the value of the last of a plurality of data bits transmitted
US6111858A (en) * 1997-02-18 2000-08-29 Virata Limited Proxy-controlled ATM subnetwork
US5742587A (en) * 1997-02-28 1998-04-21 Lanart Corporation Load balancing port switching hub
US5922056A (en) * 1997-03-03 1999-07-13 International Business Machines Corporation Computer system with peripheral device characteristic sensing and automatic communications speed setting
US6292834B1 (en) 1997-03-14 2001-09-18 Microsoft Corporation Dynamic bandwidth selection for efficient transmission of multimedia streams in a computer network
US6118817A (en) * 1997-03-14 2000-09-12 Microsoft Corporation Digital video signal encoder and encoding method having adjustable quantization
US6115420A (en) * 1997-03-14 2000-09-05 Microsoft Corporation Digital video signal encoder and encoding method
US5918002A (en) * 1997-03-14 1999-06-29 Microsoft Corporation Selective retransmission for efficient and reliable streaming of multimedia packets in a computer network
US6081843A (en) * 1997-03-20 2000-06-27 Nokia Telecommunications System using simulation cell and simulation buffer for regulating cell transfer rate according to occupancy level of the simulation buffer
US6167030A (en) * 1997-03-20 2000-12-26 Nokia Telecommunications, Oy Buffer-based traffic measurement system and method for nominal bit rate (NBR) service
US6449653B2 (en) 1997-03-25 2002-09-10 Microsoft Corporation Interleaved multiple multimedia stream for synchronized transmission over a computer network
US6151325A (en) * 1997-03-31 2000-11-21 Cisco Technology, Inc. Method and apparatus for high-capacity circuit switching with an ATM second stage switch
US6490251B2 (en) * 1997-04-14 2002-12-03 Nortel Networks Limited Method and apparatus for communicating congestion information among different protocol layers between networks
US6002933A (en) * 1997-04-29 1999-12-14 Qualcomm Incorporated Inter-system soft handoff
US6094708A (en) 1997-05-06 2000-07-25 Cisco Technology, Inc. Secondary cache write-through blocking mechanism
US6356530B1 (en) 1997-05-23 2002-03-12 Cisco Technology, Inc. Next hop selection in ATM networks
US6122272A (en) * 1997-05-23 2000-09-19 Cisco Technology, Inc. Call size feedback on PNNI operation
US6260072B1 (en) * 1997-06-12 2001-07-10 Lucent Technologies Inc Method and apparatus for adaptive routing in packet networks
US6178448B1 (en) * 1997-06-18 2001-01-23 International Business Machines Corporation Optimal link scheduling for multiple links by obtaining and utilizing link quality information
US6385168B1 (en) * 1997-06-19 2002-05-07 Alcatel Canada Inc. Fair share bandwidth allocation algorithm and device
US5951707A (en) * 1997-06-27 1999-09-14 International Business Machines Corporation Method of partitioning CRC calculation for a low-cost ATM adapter
US6081512A (en) * 1997-06-30 2000-06-27 Sun Microsystems, Inc. Spanning tree support in a high performance network device
US6487202B1 (en) 1997-06-30 2002-11-26 Cisco Technology, Inc. Method and apparatus for maximizing memory throughput
US6049528A (en) * 1997-06-30 2000-04-11 Sun Microsystems, Inc. Trunking ethernet-compatible networks
US6052738A (en) * 1997-06-30 2000-04-18 Sun Microsystems, Inc. Method and apparatus in a packet routing switch for controlling access at different data rates to a shared memory
US6044418A (en) * 1997-06-30 2000-03-28 Sun Microsystems, Inc. Method and apparatus for dynamically resizing queues utilizing programmable partition pointers
US6081522A (en) * 1997-06-30 2000-06-27 Sun Microsystems, Inc. System and method for a multi-layer network element
US6201813B1 (en) 1997-06-30 2001-03-13 Cisco Technology, Inc. Method and apparatus for using ATM queues for segmentation and reassembly of data frames
US6094435A (en) * 1997-06-30 2000-07-25 Sun Microsystems, Inc. System and method for a quality of service in a multi-layer network element
US6088356A (en) * 1997-06-30 2000-07-11 Sun Microsystems, Inc. System and method for a multi-layer network element
US6430191B1 (en) 1997-06-30 2002-08-06 Cisco Technology, Inc. Multi-stage queuing discipline
US6044087A (en) * 1997-06-30 2000-03-28 Sun Microsystems, Inc. Interface for a highly integrated ethernet network element
US6021132A (en) * 1997-06-30 2000-02-01 Sun Microsystems, Inc. Shared memory management in a switched network element
US6052726A (en) 1997-06-30 2000-04-18 Mci Communications Corp. Delay calculation for a frame relay network
US5920566A (en) * 1997-06-30 1999-07-06 Sun Microsystems, Inc. Routing in a multi-layer distributed network element
US6119196A (en) * 1997-06-30 2000-09-12 Sun Microsystems, Inc. System having multiple arbitrating levels for arbitrating access to a shared memory by network ports operating at different data rates
US6128666A (en) * 1997-06-30 2000-10-03 Sun Microsystems, Inc. Distributed VLAN mechanism for packet field replacement in a multi-layered switched network element using a control field/signal for indicating modification of a packet with a database search engine
US6246680B1 (en) 1997-06-30 2001-06-12 Sun Microsystems, Inc. Highly integrated multi-layer switch element architecture
US6016310A (en) * 1997-06-30 2000-01-18 Sun Microsystems, Inc. Trunking support in a high performance network device
US6081524A (en) 1997-07-03 2000-06-27 At&T Corp. Frame relay switched data service
US6188671B1 (en) * 1997-07-03 2001-02-13 At&T Corp Traffic management for frame relay switched data service
US6078590A (en) 1997-07-14 2000-06-20 Cisco Technology, Inc. Hierarchical routing knowledge for multicast packet routing
AU8479298A (en) * 1997-07-15 1999-02-10 Comsat Corporation Method and apparatus for adaptive loss-less compression of cell/packet headers
US6813259B1 (en) 1997-07-15 2004-11-02 Viasat, Inc. Method and apparatus for minimizing delay variance via time divided transmissions
US6003062A (en) * 1997-07-16 1999-12-14 Fore Systems, Inc. Iterative algorithm for performing max min fair allocation
US6160818A (en) * 1997-07-17 2000-12-12 At &T Corp Traffic management in packet communication networks having service priorities and employing effective bandwidths
US6047000A (en) * 1997-07-24 2000-04-04 The Hong Kong University Of Science & Technology Packet scheduling system
US6330599B1 (en) 1997-08-05 2001-12-11 Cisco Technology, Inc. Virtual interfaces with dynamic binding
US6212183B1 (en) 1997-08-22 2001-04-03 Cisco Technology, Inc. Multiple parallel packet routing lookup
US6157641A (en) * 1997-08-22 2000-12-05 Cisco Technology, Inc. Multiprotocol packet recognition and switching
US6512766B2 (en) 1997-08-22 2003-01-28 Cisco Systems, Inc. Enhanced internet packet routing lookup
US6343085B1 (en) 1997-08-28 2002-01-29 Microsoft Corporation Adaptive bandwidth throttling for individual virtual services supported on a network server
US6167027A (en) * 1997-09-09 2000-12-26 Cisco Technology, Inc. Flow control technique for X.25 traffic in a high speed packet switching network
US6205119B1 (en) * 1997-09-16 2001-03-20 Silicon Graphics, Inc. Adaptive bandwidth sharing
JP3413788B2 (en) * 1997-09-25 2003-06-09 日本電信電話株式会社 Communication method having communication protocol for performing flow control between layers and data communication terminal
US6343072B1 (en) 1997-10-01 2002-01-29 Cisco Technology, Inc. Single-chip architecture for shared-memory router
US6198724B1 (en) 1997-10-02 2001-03-06 Vertex Networks, Inc. ATM cell scheduling method and apparatus
US6424624B1 (en) 1997-10-16 2002-07-23 Cisco Technology, Inc. Method and system for implementing congestion detection and flow control in high speed digital network
KR100233091B1 (en) 1997-10-23 1999-12-01 윤종용 Atm traffic control apparatus and method
US6075791A (en) * 1997-10-28 2000-06-13 Lucent Technologies Inc. System for guaranteeing data transfer rates and delays in packet networks
US6252878B1 (en) 1997-10-30 2001-06-26 Cisco Technology, Inc. Switched architecture access server
US6359862B1 (en) * 1997-11-25 2002-03-19 International Business Machines Corporation ATM network available bit rate (ABR) explicit rate flow control system
US6122254A (en) * 1997-11-25 2000-09-19 International Business Machines Corporation Method and apparatus for network flow control with perceptive parameters
US6526060B1 (en) 1997-12-05 2003-02-25 Cisco Technology, Inc. Dynamic rate-based, weighted fair scheduler with explicit rate feedback option
US7570583B2 (en) * 1997-12-05 2009-08-04 Cisco Technology, Inc. Extending SONET/SDH automatic protection switching
WO1999030450A1 (en) 1997-12-08 1999-06-17 Packeteer, Inc. Method for data rate control for heterogenous or peer internetworking
JP3191754B2 (en) 1997-12-08 2001-07-23 日本電気株式会社 Congestion avoidance system
JP3206645B2 (en) * 1997-12-26 2001-09-10 日本電気株式会社 Broadcasting method
US6424649B1 (en) 1997-12-31 2002-07-23 Cisco Technology, Inc. Synchronous pipelined switch using serial transmission
US6111877A (en) 1997-12-31 2000-08-29 Cisco Technology, Inc. Load sharing across flows
US6263016B1 (en) * 1998-02-04 2001-07-17 3Com Corporation Methods for interfacing a subscriber link to digital networks
EP0935366A1 (en) * 1998-02-05 1999-08-11 Alcatel Cell tagging method for buffered and fragmented frames
DE69840321D1 (en) 1998-02-05 2009-01-22 Alcatel Lucent Cell elimination process
JP3075251B2 (en) * 1998-03-05 2000-08-14 日本電気株式会社 Virtual Path Bandwidth Distribution System in Asynchronous Transfer Mode Switching Network
US6115385A (en) 1998-03-11 2000-09-05 Cisco Technology, Inc. Method and system for subnetting in a switched IP network
US6208649B1 (en) 1998-03-11 2001-03-27 Cisco Technology, Inc. Derived VLAN mapping technique
US6104998A (en) * 1998-03-12 2000-08-15 International Business Machines Corporation System for coding voice signals to optimize bandwidth occupation in high speed packet switching networks
US6738814B1 (en) * 1998-03-18 2004-05-18 Cisco Technology, Inc. Method for blocking denial of service and address spoofing attacks on a private network
US6275797B1 (en) 1998-04-17 2001-08-14 Cisco Technology, Inc. Method and apparatus for measuring voice path quality by means of speech recognition
US6151633A (en) * 1998-04-20 2000-11-21 Sun Microsystems, Inc. Method and apparatus for routing and congestion control in multicast networks
US6529498B1 (en) 1998-04-28 2003-03-04 Cisco Technology, Inc. Routing support for point-to-multipoint connections
US6600724B1 (en) 1998-04-28 2003-07-29 Cisco Technology, Inc. Routing table structures
US6563798B1 (en) 1998-06-29 2003-05-13 Cisco Technology, Inc. Dynamically created service class-based routing tables
US6310877B1 (en) * 1998-04-30 2001-10-30 3Com Corporation Method of connectionless message transfer in an asynchronous transfer mode network
US6046979A (en) * 1998-05-04 2000-04-04 Cabletron Systems, Inc. Method and apparatus for controlling the flow of variable-length packets through a multiport switch
CA2237264A1 (en) 1998-05-08 1999-11-08 Northern Telecom Limited Receiver based congestion control
EP0955749A1 (en) * 1998-05-08 1999-11-10 Nortel Networks Corporation Receiver based congestion control and congestion notification from router
CA2237208A1 (en) 1998-05-08 1999-11-08 Northern Telecom Limited Congestion notification from router
GB2337429B (en) * 1998-05-15 2003-10-29 Northern Telecom Ltd Telecommunications system
JP2955561B1 (en) * 1998-05-29 1999-10-04 株式会社ディジタル・ビジョン・ラボラトリーズ Stream communication system and stream transfer control method
US6167028A (en) * 1998-06-01 2000-12-26 Motorola, Inc. Methods and apparatus for facilitating transmission of cells having multiple priorities in a cell relay network
US6385670B1 (en) * 1998-06-01 2002-05-07 Advanced Micro Devices, Inc. Data compression or decompressions during DMA transfer between a source and a destination by independently controlling the incrementing of a source and a destination address registers
US6463485B1 (en) 1998-06-03 2002-10-08 Cisco Technology, Inc. System for providing cell bus management in a switch platform including a write port cell count in each of a plurality of unidirectional FIFO for indicating which FIFO be able to accept more cell
US6438102B1 (en) 1998-06-03 2002-08-20 Cisco Technology, Inc. Method and apparatus for providing asynchronous memory functions for bi-directional traffic in a switch platform
US6512769B1 (en) * 1998-06-03 2003-01-28 Cisco Technology, Inc. Method and apparatus for rate-based cell traffic arbitration in a switch
US6483850B1 (en) 1998-06-03 2002-11-19 Cisco Technology, Inc. Method and apparatus for routing cells having different formats among service modules of a switch platform
US6370121B1 (en) 1998-06-29 2002-04-09 Cisco Technology, Inc. Method and system for shortcut trunking of LAN bridges
US6377577B1 (en) 1998-06-30 2002-04-23 Cisco Technology, Inc. Access control list processing in hardware
US6912637B1 (en) * 1998-07-08 2005-06-28 Broadcom Corporation Apparatus and method for managing memory in a network switch
US6308219B1 (en) 1998-07-31 2001-10-23 Cisco Technology, Inc. Routing table lookup implemented using M-trie having nodes duplicated in multiple memory banks
US6182147B1 (en) 1998-07-31 2001-01-30 Cisco Technology, Inc. Multicast group routing using unidirectional links
US6636512B1 (en) 1998-07-31 2003-10-21 International Business Machines Corporation System, method, and article of manufacture for increasing link bandwidth utilization in a high speed digital network
JP3145083B2 (en) * 1998-08-04 2001-03-12 松下電器産業株式会社 Transmission system, bandwidth management device, and bandwidth management method
US6389506B1 (en) 1998-08-07 2002-05-14 Cisco Technology, Inc. Block mask ternary cam
US6101115A (en) * 1998-08-07 2000-08-08 Cisco Technology, Inc. CAM match line precharge
US6535520B1 (en) 1998-08-14 2003-03-18 Cisco Technology, Inc. System and method of operation for managing data communication between physical layer devices and ATM layer devices
US6269096B1 (en) 1998-08-14 2001-07-31 Cisco Technology, Inc. Receive and transmit blocks for asynchronous transfer mode (ATM) cell delineation
US6826620B1 (en) * 1998-08-26 2004-11-30 Paradyne Corporation Network congestion control system and method
US6430153B1 (en) * 1998-09-04 2002-08-06 Cisco Technology, Inc. Trunk delay simulator
US6381245B1 (en) 1998-09-04 2002-04-30 Cisco Technology, Inc. Method and apparatus for generating parity for communication between a physical layer device and an ATM layer device
US6415326B1 (en) 1998-09-15 2002-07-02 Microsoft Corporation Timeline correlation between multiple timeline-altered media streams
US6622171B2 (en) * 1998-09-15 2003-09-16 Microsoft Corporation Multimedia timeline modification in networked client/server systems
US6611531B1 (en) * 1998-09-30 2003-08-26 Cisco Technology, Inc. Method and apparatus for routing integrated data, voice, and video traffic
US6717910B1 (en) * 1998-09-30 2004-04-06 Stmicroelectronics, Inc. Method and apparatus for controlling network data congestion
US6763017B1 (en) 1998-09-30 2004-07-13 Cisco Technology, Inc. Method and apparatus for voice port hunting of remote telephone extensions using voice over packet-data-network systems (VOPS)
US7009962B1 (en) 1998-09-30 2006-03-07 Cisco Technology, Inc. Method and apparatus for providing forwarding on ring-no-answer for remote telephone extensions using voice over packet-data-network systems (VOPS)
US6584108B1 (en) 1998-09-30 2003-06-24 Cisco Technology, Inc. Method and apparatus for dynamic allocation of multiple signal processing resources among multiple channels in voice over packet-data-network systems (VOPS)
US6535505B1 (en) 1998-09-30 2003-03-18 Cisco Technology, Inc. Method and apparatus for providing a time-division multiplexing (TDM) interface among a high-speed data stream and multiple processors
US7339924B1 (en) * 1998-09-30 2008-03-04 Cisco Technology, Inc. Method and apparatus for providing ringing timeout disconnect supervision in remote telephone extensions using voice over packet-data-network systems (VOPS)
US6785274B2 (en) 1998-10-07 2004-08-31 Cisco Technology, Inc. Efficient network multicast switching apparatus and methods
US6160793A (en) * 1998-10-13 2000-12-12 Nokia Telecommunications, Oy ECN-based approach for congestion management in hybrid IP-ATM networks
US6470013B1 (en) 1998-10-13 2002-10-22 Cisco Technology, Inc. Use of enhanced ethernet link—loop packets to automate configuration of intelligent linecards attached to a router
CA2349010A1 (en) 1998-10-27 2000-05-04 Fujitsu Network Communications, Inc. Event based rate policing with a jumping window
US6714517B1 (en) 1998-11-10 2004-03-30 Extreme Networks Method and apparatus for interconnection of packet switches with guaranteed bandwidth
US6654374B1 (en) 1998-11-10 2003-11-25 Extreme Networks Method and apparatus to reduce Jitter in packet switched networks
US6970424B2 (en) 1998-11-10 2005-11-29 Extreme Networks Method and apparatus to minimize congestion in a packet switched network
US6560196B1 (en) * 1998-11-19 2003-05-06 Cisco Technology, Inc. Method and apparatus for controlling the transmission of cells across a network
US6442165B1 (en) 1998-12-02 2002-08-27 Cisco Technology, Inc. Load balancing between service component instances
US7616640B1 (en) 1998-12-02 2009-11-10 Cisco Technology, Inc. Load balancing between service component instances
US6868061B1 (en) * 1998-12-10 2005-03-15 Nokia Corporation System and method for pre-filtering low priority packets at network nodes in a network service class utilizing a priority-based quality of service
US6700872B1 (en) 1998-12-11 2004-03-02 Cisco Technology, Inc. Method and system for testing a utopia network element
US6917617B2 (en) * 1998-12-16 2005-07-12 Cisco Technology, Inc. Use of precedence bits for quality of service
US6643260B1 (en) 1998-12-18 2003-11-04 Cisco Technology, Inc. Method and apparatus for implementing a quality of service policy in a data communications network
US6426944B1 (en) 1998-12-30 2002-07-30 At&T Corp Method and apparatus for controlling data messages across a fast packet network
US6463035B1 (en) 1998-12-30 2002-10-08 At&T Corp Method and apparatus for initiating an upward signaling control channel in a fast packet network
US6477145B1 (en) 1998-12-30 2002-11-05 At&T Corp. Method and apparatus for using video memory to buffer data packets for presentation to a data network
US6535511B1 (en) 1999-01-07 2003-03-18 Cisco Technology, Inc. Method and system for identifying embedded addressing information in a packet for translation between disparate addressing systems
US6453357B1 (en) * 1999-01-07 2002-09-17 Cisco Technology, Inc. Method and system for processing fragments and their out-of-order delivery during address translation
US6449655B1 (en) 1999-01-08 2002-09-10 Cisco Technology, Inc. Method and apparatus for communication between network devices operating at different frequencies
US6771642B1 (en) 1999-01-08 2004-08-03 Cisco Technology, Inc. Method and apparatus for scheduling packets in a packet switch
US6460085B1 (en) * 1999-02-02 2002-10-01 Mentat Inc. Method and system for managing memory in an internet over satellite connection
US6934255B1 (en) * 1999-02-02 2005-08-23 Packeteer, Inc. Internet over satellite apparatus
US6529477B1 (en) 1999-02-02 2003-03-04 Mentat Inc. Internet over satellite system
US7142506B1 (en) * 1999-02-02 2006-11-28 Vocaltec Communications Ltd. Method and apparatus for transmitting packets
US6654344B1 (en) 1999-02-02 2003-11-25 Mentat Inc. Method and system for controlling data flow in an internet over satellite connection
US6356629B1 (en) 1999-02-02 2002-03-12 Cisco Technology, Inc. Switched virtual circuit controller setup congestion management strategy
US6584083B1 (en) 1999-02-02 2003-06-24 Mentat Inc. Internet over satellite method
US6657970B1 (en) 1999-02-26 2003-12-02 Cisco Technology, Inc. Method and apparatus for link state determination in voice over frame-relay networks
US7068594B1 (en) 1999-02-26 2006-06-27 Cisco Technology, Inc. Method and apparatus for fault tolerant permanent voice calls in voice-over-packet systems
US7006493B1 (en) 1999-03-09 2006-02-28 Cisco Technology, Inc. Virtual voice port configured to connect a switched voice call to a permanent voice call
SE516571C2 (en) 1999-03-12 2002-01-29 Ericsson Telefon Ab L M Method for achieving improved transmission efficiency in a mobile packet data communication system
US6625119B1 (en) * 1999-03-17 2003-09-23 3Com Corporation Method and system for facilitating increased call traffic by switching to a low bandwidth encoder in a public emergency mode
US7102998B1 (en) * 1999-03-22 2006-09-05 Lucent Technologies Inc. Scaleable congestion control method for multicast communications over a data network
US7065762B1 (en) 1999-03-22 2006-06-20 Cisco Technology, Inc. Method, apparatus and computer program product for borrowed-virtual-time scheduling
AU2937599A (en) * 1999-03-24 2000-10-09 Nokia Networks Oy Cell transfer rate regulation using feedback in nominal bit rate (nbr) system
US6956850B1 (en) 1999-03-25 2005-10-18 Cisco Technology, Inc. Call record management for high capacity switched virtual circuits
US6757791B1 (en) 1999-03-30 2004-06-29 Cisco Technology, Inc. Method and apparatus for reordering packet data units in storage queues for reading and writing memory
US6603772B1 (en) 1999-03-31 2003-08-05 Cisco Technology, Inc. Multicast routing with multicast virtual output queues and shortest queue first allocation
US6760331B1 (en) 1999-03-31 2004-07-06 Cisco Technology, Inc. Multicast routing with nearest queue first allocation and dynamic and static vector quantization
US6567420B1 (en) * 1999-04-15 2003-05-20 Qualcomm, Incorporated Method and apparatus for high rate channel access control
US6449255B1 (en) * 1999-04-26 2002-09-10 Cisco Technology, Inc. Method and apparatus for managing packets using a real-time feedback signal
US6625121B1 (en) 1999-04-28 2003-09-23 Cisco Technology, Inc. Dynamically delisting and relisting multicast destinations in a network switching node
US6243359B1 (en) * 1999-04-29 2001-06-05 Transwitch Corp Methods and apparatus for managing traffic in an atm network
US6839348B2 (en) 1999-04-30 2005-01-04 Cisco Technology, Inc. System and method for distributing multicasts in virtual local area networks
US6553028B1 (en) 1999-04-30 2003-04-22 Cisco Technology, Inc. Method and apparatus for multicast switching using a centralized switching engine
US6993048B1 (en) * 2000-07-31 2006-01-31 Cisco Technology, Inc. ATM permanent virtual circuit and layer 3 auto-configuration for digital subscriber line customer premises equipment
US20020133620A1 (en) * 1999-05-24 2002-09-19 Krause Michael R. Access control in a network system
US7016971B1 (en) * 1999-05-24 2006-03-21 Hewlett-Packard Company Congestion management in a distributed computer system multiplying current variable injection rate with a constant to set new variable injection rate at source node
US6657955B1 (en) * 1999-05-27 2003-12-02 Alcatel Canada Inc. Buffering system employing per traffic flow accounting congestion control
US6778555B1 (en) 1999-05-28 2004-08-17 Cisco Technology, Inc. Voice over packet system configured to connect different facsimile transmission protocols
US6959006B1 (en) 1999-06-29 2005-10-25 Adc Telecommunications, Inc. Service delivery unit for an enterprise network
US6389493B1 (en) 1999-06-29 2002-05-14 Cisco Technology, Inc. System and method for dynamically allocating bandwidth to a plurality of slave cards coupled to a bus
US6847609B1 (en) 1999-06-29 2005-01-25 Adc Telecommunications, Inc. Shared management of a network entity
US6765864B1 (en) 1999-06-29 2004-07-20 Cisco Technology, Inc. Technique for providing dynamic modification of application specific policies in a feedback-based, adaptive data network
US7293280B1 (en) * 1999-07-08 2007-11-06 Microsoft Corporation Skimming continuous multimedia content
US7313808B1 (en) 1999-07-08 2007-12-25 Microsoft Corporation Browsing continuous multimedia content
EP1069801B1 (en) * 1999-07-13 2004-10-06 International Business Machines Corporation Connections bandwidth right sizing based on network resources occupancy monitoring
US6574193B1 (en) * 1999-07-28 2003-06-03 Veraz Networks Ltd. Congestion control using variable rate encoding based on queue fill
US6711125B1 (en) 1999-08-02 2004-03-23 Extreme Networks Provisioning networks for reliable quality of service
US6983350B1 (en) 1999-08-31 2006-01-03 Intel Corporation SDRAM controller for parallel processor architecture
JP3636947B2 (en) * 1999-08-31 2005-04-06 株式会社日立製作所 Information service system, service use client, and service regulation method
US6681364B1 (en) 1999-09-24 2004-01-20 International Business Machines Corporation Cyclic redundancy check for partitioned frames
US6674760B1 (en) 1999-09-28 2004-01-06 Extreme Networks, Inc. Method and system for implementing end-to-end QoS in packet-switched networks
US6535485B1 (en) * 1999-10-01 2003-03-18 Telcordia Technologies, Inc. Methods and systems for an improved reliability packet network
US6859435B1 (en) * 1999-10-13 2005-02-22 Lucent Technologies Inc. Prevention of deadlocks and livelocks in lossless, backpressured packet networks
US6977898B1 (en) 1999-10-15 2005-12-20 Cisco Technology, Inc. Method for supporting high priority calls on a congested WAN link
US7046665B1 (en) 1999-10-26 2006-05-16 Extreme Networks, Inc. Provisional IP-aware virtual paths over networks
US6820128B1 (en) * 1999-11-04 2004-11-16 Nortel Networks Limited Method and apparatus of processing packets having varying priorities by adjusting their drop functions according to a predefined fairness relationship
US6618373B1 (en) 1999-11-10 2003-09-09 Cisco Technology, Inc. Method and system for reliable in-order distribution of events
US6667976B1 (en) 1999-12-09 2003-12-23 Lucent Technologies Inc. Fuzzycast service in switches
EP1240753A1 (en) * 1999-12-13 2002-09-18 Nokia Corporation Congestion control method for a packet-switched network
US6928655B1 (en) * 1999-12-16 2005-08-09 Microsoft Corporation Live presentation searching
US7149359B1 (en) 1999-12-16 2006-12-12 Microsoft Corporation Searching and recording media streams
US6798746B1 (en) 1999-12-18 2004-09-28 Cisco Technology, Inc. Method and apparatus for implementing a quality of service policy in a data communications network
US7203171B1 (en) * 1999-12-20 2007-04-10 Cisco Technology, Inc. Ingress discard in output buffered switching devices
US6532509B1 (en) 1999-12-22 2003-03-11 Intel Corporation Arbitrating command requests in a parallel multi-threaded processing system
US6694380B1 (en) 1999-12-27 2004-02-17 Intel Corporation Mapping requests from a processing unit that uses memory-mapped input-output space
US7620702B1 (en) 1999-12-28 2009-11-17 Intel Corporation Providing real-time control data for a network processor
US6661794B1 (en) 1999-12-29 2003-12-09 Intel Corporation Method and apparatus for gigabit packet assignment for multithreaded packet processing
DE19963817C2 (en) * 1999-12-30 2002-09-26 Siemens Ag Method and device for converting a bidirectional data stream via an S2m interface for transmission over a low-voltage power grid
US6584522B1 (en) 1999-12-30 2003-06-24 Intel Corporation Communication between processors
US6952824B1 (en) 1999-12-30 2005-10-04 Intel Corporation Multi-threaded sequenced receive for fast network port stream of packets
US7480706B1 (en) 1999-12-30 2009-01-20 Intel Corporation Multi-threaded round-robin receive for fast network port
US6678244B1 (en) * 2000-01-06 2004-01-13 Cisco Technology, Inc. Congestion management system and method
US6775292B1 (en) 2000-01-24 2004-08-10 Cisco Technology, Inc. Method for servicing of multiple queues carrying voice over virtual circuits based on history
JP2001211207A (en) * 2000-01-26 2001-08-03 Sony Corp Packet transmission method, packet transmitter and band ensuring method
US6868440B1 (en) 2000-02-04 2005-03-15 Microsoft Corporation Multi-level skimming of multimedia content using playlists
US6687224B1 (en) * 2000-02-28 2004-02-03 Orckit Communications, Ltd. Bandwidth sharing method
US6970468B2 (en) * 2000-02-29 2005-11-29 International Business Machines Corporation Method and apparatus for hardware forwarding of LAN frames over ATM networks
US7016351B1 (en) 2000-02-29 2006-03-21 Cisco Technology, Inc. Small group multicast in a computer network
US6977895B1 (en) 2000-03-23 2005-12-20 Cisco Technology, Inc. Apparatus and method for rate-based polling of input interface queues in networking devices
US7237254B1 (en) 2000-03-29 2007-06-26 Microsoft Corporation Seamless switching between different playback speeds of time-scale modified data streams
US6985966B1 (en) 2000-03-29 2006-01-10 Microsoft Corporation Resynchronizing globally unsynchronized multimedia streams
US6954429B2 (en) * 2000-04-05 2005-10-11 Dyband Corporation Bandwidth control system
WO2001077850A1 (en) * 2000-04-06 2001-10-18 Rensselaer Polytechnic Institute System and method of source based multicast congestion control
US6850488B1 (en) * 2000-04-14 2005-02-01 Sun Microsystems, Inc. Method and apparatus for facilitating efficient flow control for multicast transmissions
US7142558B1 (en) 2000-04-17 2006-11-28 Cisco Technology, Inc. Dynamic queuing control for variable throughput communication channels
US7302490B1 (en) 2000-05-03 2007-11-27 Microsoft Corporation Media file format to support switching between multiple timeline-altered media streams
US7065079B1 (en) 2000-05-04 2006-06-20 Cisco Technology, Inc. VC sharing for multicast in a computer network
US6754216B1 (en) * 2000-05-08 2004-06-22 Nortel Networks Limited Method and apparatus for detecting congestion and controlling the transmission of cells across a data packet switch
US6741555B1 (en) * 2000-06-14 2004-05-25 Nokia Internet Communictions Inc. Enhancement of explicit congestion notification (ECN) for wireless network applications
US6925502B1 (en) 2000-06-20 2005-08-02 At&T Corp. Methods and systems for improving data transmission rates having adaptive protocols
US7307954B1 (en) 2000-06-23 2007-12-11 Nokia Corporation Differentiated service network and method of operating a differentiated service network
US7161938B1 (en) * 2000-07-26 2007-01-09 Infineon Technologies North America Corp. Network switch
WO2002019637A1 (en) * 2000-08-28 2002-03-07 Entridia Corporation Method and apparatus for queue administration in a packet processing device
US6920498B1 (en) 2000-08-31 2005-07-19 Cisco Technology, Inc. Phased learning approach to determining closest content serving sites
US6771665B1 (en) 2000-08-31 2004-08-03 Cisco Technology, Inc. Matching of RADIUS request and response packets during high traffic volume
US7411981B1 (en) 2000-08-31 2008-08-12 Cisco Technology, Inc. Matching of radius request and response packets during high traffic volume
US6564061B1 (en) * 2000-09-01 2003-05-13 Nokia Mobile Phones Ltd. Class based bandwidth scheduling for CDMA air interfaces
EP1187399B1 (en) * 2000-09-11 2008-12-10 Lucent Technologies Inc. Method and a device for controlling source specific data flow
US6820123B1 (en) 2000-09-28 2004-11-16 Cisco Technology, Inc. Method and apparatus for assigning hot objects to server load balancer
ATE326802T1 (en) * 2000-11-28 2006-06-15 Flash Networks Ltd SYSTEM AND METHOD FOR TRANSMISSION RATE CONTROL
US6765871B1 (en) * 2000-11-29 2004-07-20 Akara Corporation Fiber channel flow control method and apparatus for interface to metro area transport link
WO2002045322A2 (en) * 2000-11-29 2002-06-06 Pulsent Corporation Method and apparatus for combining dedicated and shared networks for efficient data transmission
US6947380B1 (en) * 2000-12-01 2005-09-20 Cisco Technology, Inc. Guaranteed bandwidth mechanism for a terabit multiservice switch
US7346698B2 (en) * 2000-12-20 2008-03-18 G. W. Hannaway & Associates Webcasting method and system for time-based synchronization of multiple, independent media streams
US7415025B1 (en) 2000-12-29 2008-08-19 Cisco Technology, Inc. Method and apparatus for clearing a large number of connections in an ATM network
US7130267B1 (en) 2000-12-29 2006-10-31 Cisco Technology, Inc. System and method for allocating bandwidth in a network node
US7382727B2 (en) * 2001-02-21 2008-06-03 Cisco Technology, Inc. System and method for asymmetrical bandwidth management
US20040037223A1 (en) * 2001-02-28 2004-02-26 David Harrison Edge-to-edge traffic control for the internet
JP3598985B2 (en) * 2001-03-21 2004-12-08 日本電気株式会社 Queue assignment system and queue assignment method for packet switch
US20020174362A1 (en) * 2001-03-29 2002-11-21 Ibm Corporation Method and system for network management capable of identifying sources of small packets
US7450510B1 (en) 2001-04-19 2008-11-11 Cisco Technology, Inc. System and method for distributing guaranteed bandwidth among service groups in a network node
US20020163883A1 (en) * 2001-05-03 2002-11-07 Price Jeffrey M. Methods and systems for providing call admission control in packetized voice networks
US6928471B2 (en) * 2001-05-07 2005-08-09 Quest Software, Inc. Method and apparatus for measurement, analysis, and optimization of content delivery
DE10122422A1 (en) 2001-05-09 2002-11-21 Siemens Ag Method for adjusting bandwidth in a connection between two communications terminals in a data network allocates a transmission channel to the connection for transmitting data.
US7050393B2 (en) * 2001-05-10 2006-05-23 International Business Machines Corporation Method, system, and product for alleviating router congestion
US7509671B1 (en) * 2001-06-20 2009-03-24 Microstrategy Incorporated Systems and methods for assigning priority to jobs in a reporting system
US7372811B1 (en) 2001-08-07 2008-05-13 Cisco Technology, Inc. Load balancing model for multilink frame relay
US7184402B1 (en) * 2001-08-30 2007-02-27 Cisco Technology, Inc Method for multi-link load balancing to improve sequenced delivery of frames at peer end
JP4032231B2 (en) * 2001-08-31 2008-01-16 日本電気株式会社 Data transmission method
US7215639B2 (en) * 2001-08-31 2007-05-08 4198638 Canada Inc. Congestion management for packet routers
IL160997A0 (en) * 2001-09-19 2004-08-31 Bay Microsystems Inc Vertical instruction and data processing in a network processor architecture
US7310348B2 (en) * 2001-09-19 2007-12-18 Bay Microsystems, Inc. Network processor architecture
US8125902B2 (en) * 2001-09-27 2012-02-28 Hyperchip Inc. Method and system for congestion avoidance in packet switching devices
US20030078001A1 (en) * 2001-10-11 2003-04-24 Steven Thompson System and method for managing congestion in a satellite communications network
US7054902B2 (en) * 2001-10-23 2006-05-30 Packeteer, Inc. Multicast delivery systems and methods
US7064740B2 (en) * 2001-11-09 2006-06-20 Sharp Laboratories Of America, Inc. Backlit display with improved dynamic range
US7480239B1 (en) 2001-11-27 2009-01-20 Cisco Technology, Inc. Method and apparatus for true priority based connection establishment within a PNNI ATM network
US7233571B1 (en) 2001-12-18 2007-06-19 Cisco Technology, Inc. Connection-based bandwidth utilization
US7076543B1 (en) 2002-02-13 2006-07-11 Cisco Technology, Inc. Method and apparatus for collecting, aggregating and monitoring network management information
US6837832B2 (en) * 2002-04-01 2005-01-04 Hanners Ernest M Orthopedic shoulder weight halter
US6959178B2 (en) * 2002-04-22 2005-10-25 Ipr Licensing Inc. Tunable upconverter mixer with image rejection
US7161904B2 (en) * 2002-06-04 2007-01-09 Fortinet, Inc. System and method for hierarchical metering in a virtual router based network switch
US7471688B2 (en) 2002-06-18 2008-12-30 Intel Corporation Scheduling system for transmission of cells to ATM virtual circuits and DSL ports
US7171482B2 (en) * 2002-07-12 2007-01-30 Ianywhere Solutions, Inc. System and method for managing bandwidth utilization
US7343398B1 (en) 2002-09-04 2008-03-11 Packeteer, Inc. Methods, apparatuses and systems for transparently intermediating network traffic over connection-based authentication protocols
US7352769B2 (en) 2002-09-12 2008-04-01 Intel Corporation Multiple calendar schedule reservation structure and method
US7130917B2 (en) * 2002-09-26 2006-10-31 Cisco Technology, Inc. Quality of service in a gateway
US8191136B2 (en) * 2002-11-04 2012-05-29 Riverbed Technology, Inc. Connection based denial of service detection
US7433307B2 (en) * 2002-11-05 2008-10-07 Intel Corporation Flow control in a network environment
KR100493234B1 (en) * 2002-11-25 2005-06-02 한국전자통신연구원 node system, dual ring communication system the using node system and method thereof
US7295565B2 (en) * 2002-12-24 2007-11-13 Sun Microsystems, Inc. System and method for sharing a resource among multiple queues
US7298753B1 (en) 2003-02-10 2007-11-20 Cisco Technology, Inc. Technique for managing heavy signaling traffic that is directed to a particular signaling control unit
JP2004342998A (en) * 2003-05-19 2004-12-02 Elpida Memory Inc Method for forming insulating film
US7443836B2 (en) 2003-06-16 2008-10-28 Intel Corporation Processing a data packet
US7132953B2 (en) * 2003-06-26 2006-11-07 Lear Corporation Spring sensor assembly for a vehicle seat cushion
US20050013589A1 (en) * 2003-07-14 2005-01-20 Microsoft Corporation Adding recording functionality to a media player
US7263067B2 (en) * 2003-07-15 2007-08-28 Nokia Siemans Networks Oy Method and apparatus for accelerating throughput in a wireless or other telecommunication system
US9247288B2 (en) 2003-08-12 2016-01-26 Time Warner Cable Enterprises Llc Technique for effectively delivering targeted advertisements through a communications network having limited bandwidth
US20050190779A1 (en) * 2004-03-01 2005-09-01 Cisco Technology, Inc., A California Corporation Scalable approach to large scale queuing through dynamic resource allocation
US20050047405A1 (en) * 2003-08-25 2005-03-03 International Business Machines Corporation Switching device for controlling data packet flow
US8199764B2 (en) * 2003-08-25 2012-06-12 Cisco Technology, Inc. Scalable approach to large scale queuing through dynamic resource allocation
US7317682B2 (en) * 2003-09-04 2008-01-08 Mitsubishi Electric Research Laboratories, Inc. Passive and distributed admission control method for ad hoc networks
US7532574B1 (en) 2003-10-02 2009-05-12 Cisco Technology, Inc. Method and apparatus for improved priority based connection establishment within a PNNI ATM network
US7355969B2 (en) * 2003-10-07 2008-04-08 Alcatel Line card port protection rate limiter circuitry
US20080052687A1 (en) * 2003-11-03 2008-02-28 Agustin Gonzales-Tuchmann Development environment for data transformation applications
WO2005052673A2 (en) 2003-11-21 2005-06-09 Sharp Laboratories Of America, Inc. Liquid crystal display with adaptive color
JP4349114B2 (en) * 2003-12-10 2009-10-21 ソニー株式会社 Transmission device and method, reception device and method, recording medium, and program
US7274663B2 (en) 2003-12-15 2007-09-25 International Business Machines Corporation System and method for testing differentiated services in a value add network service
CN1324862C (en) * 2004-01-06 2007-07-04 北京邮电大学 Method for regulating congest window in communication network
US7546082B2 (en) * 2004-03-02 2009-06-09 Telcordia Technologies, Inc. Application-layer multicast for mobile users in diverse networks
US7809007B2 (en) * 2004-03-05 2010-10-05 David Mayhew Low cost implementation for a device utilizing look ahead congestion management
US7612757B2 (en) * 2004-05-04 2009-11-03 Sharp Laboratories Of America, Inc. Liquid crystal display with modulated black point
US7532192B2 (en) * 2004-05-04 2009-05-12 Sharp Laboratories Of America, Inc. Liquid crystal display with filtered black point
US7777714B2 (en) * 2004-05-04 2010-08-17 Sharp Laboratories Of America, Inc. Liquid crystal display with adaptive width
US7872631B2 (en) * 2004-05-04 2011-01-18 Sharp Laboratories Of America, Inc. Liquid crystal display with temporal black point
US7505018B2 (en) * 2004-05-04 2009-03-17 Sharp Laboratories Of America, Inc. Liquid crystal display with reduced black level insertion
US8395577B2 (en) * 2004-05-04 2013-03-12 Sharp Laboratories Of America, Inc. Liquid crystal display with illumination control
US20050248553A1 (en) * 2004-05-04 2005-11-10 Sharp Laboratories Of America, Inc. Adaptive flicker and motion blur control
US7602369B2 (en) * 2004-05-04 2009-10-13 Sharp Laboratories Of America, Inc. Liquid crystal display with colored backlight
US7672243B2 (en) * 2004-06-04 2010-03-02 David Mayhew System and method to identify and communicate congested flows in a network fabric
US7023451B2 (en) * 2004-06-14 2006-04-04 Sharp Laboratories Of America, Inc. System for reducing crosstalk
US8843978B2 (en) 2004-06-29 2014-09-23 Time Warner Cable Enterprises Llc Method and apparatus for network bandwidth allocation
US7556836B2 (en) * 2004-09-03 2009-07-07 Solae, Llc High protein snack product
US7505405B2 (en) * 2004-10-08 2009-03-17 International Business Machines Corporation Method, apparatus, and computer program product for optimizing packet flow control through buffer status forwarding
US7898519B2 (en) * 2005-02-17 2011-03-01 Sharp Laboratories Of America, Inc. Method for overdriving a backlit display
US8050512B2 (en) * 2004-11-16 2011-11-01 Sharp Laboratories Of America, Inc. High dynamic range images from low dynamic range images
US8050511B2 (en) * 2004-11-16 2011-11-01 Sharp Laboratories Of America, Inc. High dynamic range images from low dynamic range images
US7525528B2 (en) * 2004-11-16 2009-04-28 Sharp Laboratories Of America, Inc. Technique that preserves specular highlights
US7567565B2 (en) 2005-02-01 2009-07-28 Time Warner Cable Inc. Method and apparatus for network bandwidth conservation
US9240834B2 (en) * 2005-03-22 2016-01-19 Hughes Network Systems, Llc Method and apparatus for providing open loop bandwidth allocation
US8612647B2 (en) * 2005-04-12 2013-12-17 Hewlett—Packard Development Company, L.P. Priority aware queue
US20070058559A1 (en) * 2005-09-15 2007-03-15 Sharp Laboratories Of America, Inc. Method and system of assigning priority to detection messages
US7961621B2 (en) * 2005-10-11 2011-06-14 Cisco Technology, Inc. Methods and devices for backward congestion notification
JP4904136B2 (en) * 2005-12-07 2012-03-28 韓國電子通信研究院 Single-port memory controller for bidirectional data communication and control method thereof
US8121401B2 (en) * 2006-01-24 2012-02-21 Sharp Labortories of America, Inc. Method for reducing enhancement of artifacts and noise in image color enhancement
US9143657B2 (en) * 2006-01-24 2015-09-22 Sharp Laboratories Of America, Inc. Color enhancement technique using skin color detection
US7826364B1 (en) * 2006-02-09 2010-11-02 Verizon Services Corp. Dynamic service-aware flow control in packet networks
US7751328B2 (en) * 2006-02-17 2010-07-06 Cisco Technology, Inc. System and method for efficient network flow control
US8170065B2 (en) 2006-02-27 2012-05-01 Time Warner Cable Inc. Methods and apparatus for selecting digital access technology for programming and data delivery
US8458753B2 (en) 2006-02-27 2013-06-04 Time Warner Cable Enterprises Llc Methods and apparatus for device capabilities discovery and utilization within a content-based network
US7872973B2 (en) * 2006-03-17 2011-01-18 Alcatel Lucent Method and system for using a queuing device as a lossless stage in a network device in a communications network
US7751339B2 (en) 2006-05-19 2010-07-06 Cisco Technology, Inc. Method and apparatus for simply configuring a subscriber appliance for performing a service controlled by a separate service provider
US8477614B2 (en) 2006-06-30 2013-07-02 Centurylink Intellectual Property Llc System and method for routing calls if potential call paths are impaired or congested
US9094257B2 (en) 2006-06-30 2015-07-28 Centurylink Intellectual Property Llc System and method for selecting a content delivery network
US8289965B2 (en) 2006-10-19 2012-10-16 Embarq Holdings Company, Llc System and method for establishing a communications session with an end-user based on the state of a network connection
US8488447B2 (en) 2006-06-30 2013-07-16 Centurylink Intellectual Property Llc System and method for adjusting code speed in a transmission path during call set-up due to reduced transmission performance
US8717911B2 (en) 2006-06-30 2014-05-06 Centurylink Intellectual Property Llc System and method for collecting network performance information
US8194643B2 (en) 2006-10-19 2012-06-05 Embarq Holdings Company, Llc System and method for monitoring the connection of an end-user to a remote network
US8130793B2 (en) 2006-08-22 2012-03-06 Embarq Holdings Company, Llc System and method for enabling reciprocal billing for different types of communications over a packet network
US8199653B2 (en) 2006-08-22 2012-06-12 Embarq Holdings Company, Llc System and method for communicating network performance information over a packet network
US8144587B2 (en) 2006-08-22 2012-03-27 Embarq Holdings Company, Llc System and method for load balancing network resources using a connection admission control engine
US8537695B2 (en) 2006-08-22 2013-09-17 Centurylink Intellectual Property Llc System and method for establishing a call being received by a trunk on a packet network
US8531954B2 (en) * 2006-08-22 2013-09-10 Centurylink Intellectual Property Llc System and method for handling reservation requests with a connection admission control engine
US8015294B2 (en) 2006-08-22 2011-09-06 Embarq Holdings Company, LP Pin-hole firewall for communicating data packets on a packet network
US8064391B2 (en) 2006-08-22 2011-11-22 Embarq Holdings Company, Llc System and method for monitoring and optimizing network performance to a wireless device
US8238253B2 (en) 2006-08-22 2012-08-07 Embarq Holdings Company, Llc System and method for monitoring interlayer devices and optimizing network performance
US8307065B2 (en) 2006-08-22 2012-11-06 Centurylink Intellectual Property Llc System and method for remotely controlling network operators
US8274905B2 (en) 2006-08-22 2012-09-25 Embarq Holdings Company, Llc System and method for displaying a graph representative of network performance over a time period
US8743703B2 (en) 2006-08-22 2014-06-03 Centurylink Intellectual Property Llc System and method for tracking application resource usage
US8619600B2 (en) 2006-08-22 2013-12-31 Centurylink Intellectual Property Llc System and method for establishing calls over a call path having best path metrics
US8189468B2 (en) 2006-10-25 2012-05-29 Embarq Holdings, Company, LLC System and method for regulating messages between networks
US8576722B2 (en) 2006-08-22 2013-11-05 Centurylink Intellectual Property Llc System and method for modifying connectivity fault management packets
US7684332B2 (en) 2006-08-22 2010-03-23 Embarq Holdings Company, Llc System and method for adjusting the window size of a TCP packet through network elements
US8223655B2 (en) 2006-08-22 2012-07-17 Embarq Holdings Company, Llc System and method for provisioning resources of a packet network based on collected network performance information
US7843831B2 (en) 2006-08-22 2010-11-30 Embarq Holdings Company Llc System and method for routing data on a packet network
US8224255B2 (en) 2006-08-22 2012-07-17 Embarq Holdings Company, Llc System and method for managing radio frequency windows
US8750158B2 (en) 2006-08-22 2014-06-10 Centurylink Intellectual Property Llc System and method for differentiated billing
US8407765B2 (en) 2006-08-22 2013-03-26 Centurylink Intellectual Property Llc System and method for restricting access to network performance information tables
US9479341B2 (en) 2006-08-22 2016-10-25 Centurylink Intellectual Property Llc System and method for initiating diagnostics on a packet network node
EP1892886A1 (en) * 2006-08-23 2008-02-27 Nokia Siemens Networks Gmbh & Co. Kg Method for controlling load adaptation in a mobile communications system
US7643504B2 (en) * 2006-11-29 2010-01-05 Applied Micro Circuits Corporation Credit based flow control in an asymmetric channel environment
US8941580B2 (en) * 2006-11-30 2015-01-27 Sharp Laboratories Of America, Inc. Liquid crystal display with area adaptive backlight
US20080235746A1 (en) 2007-03-20 2008-09-25 Michael James Peters Methods and apparatus for content delivery and replacement in a network
US20080298397A1 (en) * 2007-05-16 2008-12-04 Broadcom Corporation Communication fabric bandwidth management
US8140827B2 (en) * 2007-06-19 2012-03-20 Samsung Electronics Co., Ltd. System and method for efficient data transmission in a multi-processor environment
US7975027B2 (en) * 2007-08-06 2011-07-05 International Business Machines Corporation Credit depletion notification for transmitting frames between a port pair
US7787375B2 (en) * 2007-08-06 2010-08-31 International Business Machines Corporation Performing a recovery action in response to a credit depletion notification
US8798056B2 (en) 2007-09-24 2014-08-05 Intel Corporation Method and system for virtual port communications
US8561116B2 (en) 2007-09-26 2013-10-15 Charles A. Hasek Methods and apparatus for content caching in a video network
US9071859B2 (en) 2007-09-26 2015-06-30 Time Warner Cable Enterprises Llc Methods and apparatus for user-based targeted content delivery
US8099757B2 (en) 2007-10-15 2012-01-17 Time Warner Cable Inc. Methods and apparatus for revenue-optimized delivery of content in a network
US8171123B2 (en) * 2007-12-04 2012-05-01 Sony Computer Entertainment Inc. Network bandwidth detection and distribution
US7724652B2 (en) * 2008-01-08 2010-05-25 International Business Machines Corporation Method of reducing network congestion
US8813143B2 (en) 2008-02-26 2014-08-19 Time Warner Enterprises LLC Methods and apparatus for business-based network resource allocation
US20090225775A1 (en) * 2008-03-06 2009-09-10 Integrated Device Technology, Inc. Serial Buffer To Support Reliable Connection Between Rapid I/O End-Point And FPGA Lite-Weight Protocols
US8312241B2 (en) * 2008-03-06 2012-11-13 Integrated Device Technology, Inc. Serial buffer to support request packets with out of order response packets
US8625621B2 (en) * 2008-03-06 2014-01-07 Integrated Device Technology, Inc. Method to support flexible data transport on serial protocols
US8312190B2 (en) * 2008-03-06 2012-11-13 Integrated Device Technology, Inc. Protocol translation in a serial buffer
US8213448B2 (en) * 2008-03-06 2012-07-03 Integrated Device Technology, Inc. Method to support lossless real time data sampling and processing on rapid I/O end-point
US20090228733A1 (en) * 2008-03-06 2009-09-10 Integrated Device Technology, Inc. Power Management On sRIO Endpoint
US8068425B2 (en) 2008-04-09 2011-11-29 Embarq Holdings Company, Llc System and method for using network performance information to determine improved measures of path states
JP2009272874A (en) * 2008-05-07 2009-11-19 Sony Corp Communication apparatus, communicating method, program, and communicating system
US20100057860A1 (en) * 2008-08-29 2010-03-04 Fry Donna M Confirmation and acknowledgement of transmission reception
US8331231B2 (en) 2008-09-09 2012-12-11 Centurylink Intellectual Property Llc System and method for monitoring bursting traffic
US9866609B2 (en) 2009-06-08 2018-01-09 Time Warner Cable Enterprises Llc Methods and apparatus for premises content distribution
US8174974B2 (en) * 2009-11-12 2012-05-08 Yahoo! Inc. Voluntary admission control for traffic yield management
US8438300B2 (en) * 2010-01-27 2013-05-07 International Business Machines Corporation Achieving distributed flow control via data conflation
US8553540B2 (en) * 2010-03-05 2013-10-08 Microsoft Corporation Congestion control for delay sensitive applications
US20120008573A1 (en) 2010-07-08 2012-01-12 Apple Inc. Radio resource signaling during network congestion in a mobile wireless device
US8787256B2 (en) * 2010-12-03 2014-07-22 Motorola Solutions, Inc. Method and apparatus for ensuring transmission of critical data through a wireless adapter
KR101473317B1 (en) * 2010-12-23 2014-12-17 주식회사 케이티 Cloud computing system and traffic distributing and control method in cloud computing system
US9503223B2 (en) * 2011-03-04 2016-11-22 Blackberry Limited Controlling network device behavior
US9854280B2 (en) 2012-07-10 2017-12-26 Time Warner Cable Enterprises Llc Apparatus and methods for selective enforcement of secondary content viewing
US8862155B2 (en) 2012-08-30 2014-10-14 Time Warner Cable Enterprises Llc Apparatus and methods for enabling location-based services within a premises
US9215181B2 (en) 2012-11-06 2015-12-15 Comcast Cable Communications, Llc Systems and methods for managing a network
US9131283B2 (en) 2012-12-14 2015-09-08 Time Warner Cable Enterprises Llc Apparatus and methods for multimedia coordination
US8976664B2 (en) * 2013-03-11 2015-03-10 Broadcom Corporation Facilitating network flows
US10142236B2 (en) 2013-03-14 2018-11-27 Comcast Cable Communications, Llc Systems and methods for managing a packet network
US9066153B2 (en) 2013-03-15 2015-06-23 Time Warner Cable Enterprises Llc Apparatus and methods for multicast delivery of content in a content delivery network
US10368255B2 (en) 2017-07-25 2019-07-30 Time Warner Cable Enterprises Llc Methods and apparatus for client-based dynamic control of connections to co-existing radio access networks
US9166925B2 (en) * 2013-04-05 2015-10-20 International Business Machines Corporation Virtual quantized congestion notification
US9614794B2 (en) * 2013-07-11 2017-04-04 Apollo Education Group, Inc. Message consumer orchestration framework
US9313568B2 (en) 2013-07-23 2016-04-12 Chicago Custom Acoustics, Inc. Custom earphone with dome in the canal
US9288144B2 (en) * 2013-10-28 2016-03-15 Broadcom Corporation Scalable and predictive packet processing
US10122639B2 (en) * 2013-10-30 2018-11-06 Comcast Cable Communications, Llc Systems and methods for managing a network
US11540148B2 (en) 2014-06-11 2022-12-27 Time Warner Cable Enterprises Llc Methods and apparatus for access point location
US10028025B2 (en) 2014-09-29 2018-07-17 Time Warner Cable Enterprises Llc Apparatus and methods for enabling presence-based and use-based services
US9935833B2 (en) 2014-11-05 2018-04-03 Time Warner Cable Enterprises Llc Methods and apparatus for determining an optimized wireless interface installation configuration
US10225199B2 (en) * 2015-02-11 2019-03-05 Telefonaktiebolaget Lm Ericsson (Publ) Ethernet congestion control and prevention
US9986578B2 (en) 2015-12-04 2018-05-29 Time Warner Cable Enterprises Llc Apparatus and methods for selective data network access
US9918345B2 (en) 2016-01-20 2018-03-13 Time Warner Cable Enterprises Llc Apparatus and method for wireless network services in moving vehicles
US10492034B2 (en) 2016-03-07 2019-11-26 Time Warner Cable Enterprises Llc Apparatus and methods for dynamic open-access networks
US10586023B2 (en) 2016-04-21 2020-03-10 Time Warner Cable Enterprises Llc Methods and apparatus for secondary content management and fraud prevention
US10687115B2 (en) 2016-06-01 2020-06-16 Time Warner Cable Enterprises Llc Cloud-based digital content recorder apparatus and methods
US10164858B2 (en) 2016-06-15 2018-12-25 Time Warner Cable Enterprises Llc Apparatus and methods for monitoring and diagnosing a wireless network
US10911794B2 (en) 2016-11-09 2021-02-02 Charter Communications Operating, Llc Apparatus and methods for selective secondary content insertion in a digital network
US10645547B2 (en) 2017-06-02 2020-05-05 Charter Communications Operating, Llc Apparatus and methods for providing wireless service in a venue
US10638361B2 (en) 2017-06-06 2020-04-28 Charter Communications Operating, Llc Methods and apparatus for dynamic control of connections to co-existing radio access networks
US11109290B2 (en) 2017-08-04 2021-08-31 Charter Communications Operating, Llc Switching connections over frequency bands of a wireless network
US10939142B2 (en) 2018-02-27 2021-03-02 Charter Communications Operating, Llc Apparatus and methods for content storage, distribution and security within a content distribution network
US11240176B2 (en) 2018-08-21 2022-02-01 At&T Intellectual Property I, L.P. Time-spaced messaging for facilitating network communications

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0374928A2 (en) * 1988-12-23 1990-06-27 Hitachi, Ltd. Packet congestion control method and packet switching equipment

Family Cites Families (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CH577253A5 (en) * 1974-05-17 1976-06-30 Ibm
JPS585541B2 (en) * 1978-05-08 1983-01-31 富士通株式会社 Packet transmission method
JPS57136833A (en) * 1981-02-17 1982-08-24 Sony Corp Time-division multiplex data transmitting method
US4475192A (en) * 1982-02-16 1984-10-02 At&T Bell Laboratories Data packet flow control scheme for switching networks
US4506358A (en) * 1982-06-25 1985-03-19 At&T Bell Laboratories Time stamping for a packet switching system
US4494230A (en) * 1982-06-25 1985-01-15 At&T Bell Laboratories Fast packet switching system
US4499576A (en) * 1982-08-13 1985-02-12 At&T Bell Laboratories Multiplexed first-in, first-out queues
US4561090A (en) * 1983-05-18 1985-12-24 At&T Bell Laboratories Integrated self-checking packet switch node
US4819228A (en) * 1984-10-29 1989-04-04 Stratacom Inc. Synchronous packet voice/data communication system
US4903261A (en) * 1984-10-29 1990-02-20 Stratacom, Inc. Synchronous packet voice/data communication system
US4771425A (en) * 1984-10-29 1988-09-13 Stratacom, Inc. Synchoronous packet voice/data communication system
US4644533A (en) * 1985-05-06 1987-02-17 American Telephone & Telegraph Company Packet switch trunk circuit queueing arrangement
JPH0831876B2 (en) * 1985-09-20 1996-03-27 株式会社日立製作所 Routing control method in packet switching network
US4703475A (en) * 1985-12-04 1987-10-27 American Telephone And Telegraph Company At&T Bell Laboratories Data communication method and apparatus using multiple physical data links
US4703477A (en) * 1986-02-28 1987-10-27 American Telephone And Telegraph Company At&T Bell Laboratories Packet information field data format
US4757529A (en) * 1986-02-28 1988-07-12 American Telephone And Telegraph Company, At&T Bell Laboratories Call distribution arrangement
DE3685217D1 (en) * 1986-08-27 1992-06-11 Ibm DEVICE FOR OPTIMIZING BANDWIDTH ASSIGNMENT BETWEEN CALL-SWITCHED TIME SLOTS AND PACKET BIT CURRENT IN A COMMUNICATION NETWORK.
US4839891A (en) * 1987-07-24 1989-06-13 Nec Corporation Method for controlling data flow
US4905233A (en) * 1987-11-23 1990-02-27 Harris Corporation Multiple path routing mechanism for packet communications network
US5088032A (en) * 1988-01-29 1992-02-11 Cisco Systems, Inc. Method and apparatus for routing communications among computer networks
FR2630606B1 (en) * 1988-04-20 1990-07-27 Servel Michel ROUTING METHOD AND PACKET SWITCHING NETWORKS IN PARTICULAR FOR BUSINESS NETWORKS AND USER FACILITIES IN A BROADBAND PUBLIC NETWORK
US4914650A (en) * 1988-12-06 1990-04-03 American Telephone And Telegraph Company Bandwidth allocation and congestion control scheme for an integrated voice and data network
US5020058A (en) * 1989-01-23 1991-05-28 Stratacom, Inc. Packet voice/data communication system having protocol independent repetitive packet suppression
US4974224A (en) * 1989-11-07 1990-11-27 Harris Corporation Distributed split flow routing mechanism for multi-node packet switching communication network
US5072449A (en) * 1989-12-21 1991-12-10 Stratacom, Inc. Packet framing using cyclic redundancy checking
US5115431A (en) * 1990-09-28 1992-05-19 Stratacom, Inc. Method and apparatus for packet communications signaling

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0374928A2 (en) * 1988-12-23 1990-06-27 Hitachi, Ltd. Packet congestion control method and packet switching equipment

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
COMPUTER NETWORKING SYMPOSIUM April 1988, WASHINGTON US pages 134 - 143 R. JAIN ET AL. 'CONGESTION AVOIDANCE IN COMPUTER NETWORKS WITH A CONNECTIONLESS NETWORK LAYER: CONCEPTS, GOALS AND METHODOLOGY' *
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATION vol. 9, no. 7, September 1991, NEW YORK US pages 1119 - 1130 A. GERSHT ET AL. 'A CONGESTION CONTROL FRAMEWORK FOR ATM NETWORKS' *
ITC-13 PROCEEDINGS OF THE THIRTEENTH INTERNATIONAL TELETRAFFIC CONGRESS June 1991, COPENHAGEN DK pages 853 - 858 O. ABOUL-MAGD ET AL. 'FLOW AND CONGESTION CONTROL FOR BROADBAND PACKET NETWORKS' *

Cited By (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2281005A (en) * 1993-06-21 1995-02-15 Plessey Telecomm Network management to avoid congestion
EP0658999A2 (en) * 1993-12-15 1995-06-21 Nec Corporation Traffic control for ATM networks
EP0658999A3 (en) * 1993-12-15 1998-05-27 Nec Corporation Traffic control for ATM networks
EP0706297A1 (en) * 1994-10-07 1996-04-10 International Business Machines Corporation Method for operating traffic congestion control in a data communication network and system for implementing said method
US5790522A (en) * 1994-10-07 1998-08-04 International Business Machines Corporation Method and system for performing traffic congestion control in a data communication network
US5633861A (en) * 1994-12-19 1997-05-27 Alcatel Data Networks Inc. Traffic management and congestion control for packet-based networks
US5726977A (en) * 1994-12-22 1998-03-10 Korea Telecommunication Authority Apparatus and method for determining a network node congestion state in order to control the congestion in an ATM network
WO1997009814A1 (en) * 1995-09-07 1997-03-13 Ericsson Australia Pty. Ltd. Controlling traffic congestion in intelligent electronic networks
AU706714B2 (en) * 1995-09-07 1999-06-24 Ericsson Australia Pty Ltd Controlling traffic congestion in intelligent electronic networks
US6532214B1 (en) 1995-09-07 2003-03-11 Ericsson Australia Pty Ltd. Controlling traffic congestion in intelligent electronic networks
US8161182B1 (en) 2000-01-26 2012-04-17 Cisco Technology, Inc. Managing network congestion using dynamically advertised congestion status
US7570584B1 (en) 2002-03-29 2009-08-04 Cisco Technology, Inc. Network-wide congestion control of SPVC signaling messages
WO2004002085A1 (en) * 2002-06-20 2003-12-31 Telefonaktiebolaget Lm Ericsson (Publ) Apparatus and method for resource allocation
US7969979B2 (en) 2003-09-29 2011-06-28 Runcom Technologies Ltd. Distribution of multicast data to users
US8909807B2 (en) 2005-04-07 2014-12-09 Opanga Networks, Inc. System and method for progressive download using surplus network capacity
US9065595B2 (en) 2005-04-07 2015-06-23 Opanga Networks, Inc. System and method for peak flow detection in a communication network
US10396913B2 (en) 2005-04-07 2019-08-27 Opanga Networks, Inc. System and method for peak flow detection in a communication network
US11258531B2 (en) 2005-04-07 2022-02-22 Opanga Networks, Inc. System and method for peak flow detection in a communication network
WO2008005696A1 (en) * 2006-06-30 2008-01-10 Intel Corporation End-point based tamper resistant congestion management
US9143341B2 (en) 2008-11-07 2015-09-22 Opanga Networks, Inc. Systems and methods for portable data storage devices that automatically initiate data transfers utilizing host devices
WO2011022095A1 (en) * 2009-08-19 2011-02-24 Opanga Networks, Inc Enhanced data delivery based on real time analysis of network communications quality and traffic
KR20120089467A (en) * 2009-08-19 2012-08-10 오팡가 네트웍스, 인크. Enhanced data delivery based on real time analysis of network communications quality and traffic
US8886790B2 (en) 2009-08-19 2014-11-11 Opanga Networks, Inc. Systems and methods for optimizing channel resources by coordinating data transfers based on data type and traffic
KR101689778B1 (en) 2009-08-19 2016-12-27 오팡가 네트웍스, 인크. Enhanced data delivery based on real time analysis of network communications quality and traffic
WO2018004412A1 (en) * 2016-07-01 2018-01-04 Telefonaktiebolaget Lm Ericsson (Publ) Round trip time skew control methods and arrangements
US10944678B2 (en) 2016-07-01 2021-03-09 Telefonaktiebolaget Lm Ericsson (Publ) Round trip time skew control methods and arrangements

Also Published As

Publication number Publication date
AU3797093A (en) 1993-11-08
US5313454A (en) 1994-05-17
JPH07508378A (en) 1995-09-14
EP0634068A1 (en) 1995-01-18

Similar Documents

Publication Publication Date Title
US5313454A (en) Congestion control for cell networks
US5224099A (en) Circuitry and method for fair queuing and servicing cell traffic using hopcounts and traffic classes
Sriram Methodologies for bandwidth allocation, transmission scheduling, and congestion avoidance in broadband ATM networks
AU602379B2 (en) Packet switching system arranged for congestion control through bandwidth management
US6108307A (en) Frame relay priority queses to offer multiple service classes
CA2214838C (en) Broadband switching system
US5339332A (en) Arrangement for monitoring the bit rate in ATM networks
US6526023B1 (en) Traffic control apparatus
EP0275679A1 (en) Packet switching system arranged for congestion control
WO1995003657A1 (en) Atm exchange
EP1014747A2 (en) Method and apparatus for managing communications traffic
JPH10135975A (en) Feedback controller in cell exchange and cell-scheduling equipment
JP4652494B2 (en) Flow control method in ATM switch of distributed configuration
US6621792B1 (en) Computationally-efficient traffic shaper
US6587436B1 (en) Method and apparatus for allocation of available bandwidth
JP2001060952A (en) Traffic shaper for maintenance cell without causing either of jitter or delay
EP1090529B1 (en) Method and system for a loop back connection using a priority ubr and adsl modem
Kalampoukas et al. Dynamics of an Explicit Rate Allocation Algorithm for Available Bit-Rate (ABR) Service in ATM Networks
KR100319457B1 (en) Traffic control method for ATM switches
Chang IP layer per-flow queueing and credit flow control
O'Neill A method for congestion control in atm networks using peak rate throttling
JP3833664B2 (en) Feedback control device and cell scheduling device in cell switch
Zervanos et al. Design and implementation of an ABR server in a shared-bus ATM switch
Bhardwaj et al. ISDN frame relay and its congestion control
JPH09261231A (en) Transmitting speed controller

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AT AU BB BG BR CA CH CZ DE DK ES FI GB HU JP KP KR KZ LK LU MG MN MW NL NO NZ PL PT RO RU SD SE SK UA

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN ML MR SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
WWE Wipo information: entry into national phase

Ref document number: 1993907328

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 1993907328

Country of ref document: EP

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

NENP Non-entry into the national phase

Ref country code: CA

WWW Wipo information: withdrawn in national office

Ref document number: 1993907328

Country of ref document: EP