+

US20130322254A1 - System and method for managing mpls-te overload - Google Patents

System and method for managing mpls-te overload Download PDF

Info

Publication number
US20130322254A1
US20130322254A1 US13/838,673 US201313838673A US2013322254A1 US 20130322254 A1 US20130322254 A1 US 20130322254A1 US 201313838673 A US201313838673 A US 201313838673A US 2013322254 A1 US2013322254 A1 US 2013322254A1
Authority
US
United States
Prior art keywords
igp
tlv
overload condition
overload
router
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/838,673
Inventor
Pradeep G. Jain
Kanwar D. Singh
Jaishal Shah
Srikrishnan Venkataraman
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Alcatel Lucent SAS
Original Assignee
Alcatel Lucent USA Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Alcatel Lucent USA Inc filed Critical Alcatel Lucent USA Inc
Priority to US13/838,673 priority Critical patent/US20130322254A1/en
Assigned to ALCATEL-LUCENT USA INC. reassignment ALCATEL-LUCENT USA INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SHAH, Jaishal, VENKATARAMAN, SRIKRISHNAN, JAIN, PRADEEP G, SINGH, KANWAR D
Priority to PCT/US2013/042831 priority patent/WO2013181130A1/en
Priority to EP13728880.9A priority patent/EP2856715B1/en
Priority to JP2015515107A priority patent/JP2015518358A/en
Priority to KR20147033610A priority patent/KR20150010968A/en
Priority to CN201380028224.7A priority patent/CN104521194B/en
Publication of US20130322254A1 publication Critical patent/US20130322254A1/en
Assigned to ALCATEL LUCENT reassignment ALCATEL LUCENT ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ALCATEL-LUCENT USA INC.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/16Multipoint routing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/16Threshold monitoring
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/02Topology update or discovery
    • H04L45/04Interdomain routing, e.g. hierarchical routing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/12Shortest path evaluation
    • H04L45/124Shortest path evaluation using a combination of metrics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/12Shortest path evaluation
    • H04L45/125Shortest path evaluation based on throughput or bandwidth
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/50Routing or path finding of packets in data switching networks using label swapping, e.g. multi-protocol label switch [MPLS]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/12Avoiding congestion; Recovering from congestion
    • H04L47/125Avoiding congestion; Recovering from congestion by balancing the load, e.g. traffic engineering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/35Flow control; Congestion control by embedding flow control information in regular packets, e.g. piggybacking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/70Admission control; Resource allocation
    • H04L47/72Admission control; Resource allocation using reservation actions during connection setup
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/70Admission control; Resource allocation
    • H04L47/72Admission control; Resource allocation using reservation actions during connection setup
    • H04L47/724Admission control; Resource allocation using reservation actions during connection setup at intermediate nodes, e.g. resource reservation protocol [RSVP]
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02BCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO BUILDINGS, e.g. HOUSING, HOUSE APPLIANCES OR RELATED END-USER APPLICATIONS
    • Y02B90/00Enabling technologies or technologies with a potential or indirect contribution to GHG emissions mitigation
    • Y02B90/20Smart grids as enabling technology in buildings sector

Definitions

  • the invention relates to the field of communication networks such as multi-protocol label switching (MPLS) networks and, more particularly but not exclusively, to resource overload detection and management mechanisms.
  • MPLS multi-protocol label switching
  • Multiprotocol Label Switching enables efficient delivery of a wide variety of differentiated, end-to-end services.
  • Multiprotocol Label Switching (MPLS) traffic engineering (TE) provides a mechanism for selecting efficient paths across an MPLS network based on bandwidth considerations and administrative rules.
  • Each label switching router maintains a TE link state database with a current network topology. Once a path is computed, TE is used to maintain a forwarding state along that path.
  • RSVP Resource Reservation Protocol
  • a router or other network element or node may experience a resource overutilization condition (i.e., insufficient memory, processor, input/output or other resources) in response to receiving a large number of RSVP Packets.
  • a resource overutilization condition i.e., insufficient memory, processor, input/output or other resources
  • Such a condition may result in the RSVP/MPLS process temporarily dropping RSVP packets to conserve resources. If the condition persists, then the node may start tearing down existing MPLS-TE LSPs to release resources, which in turn may lead to service interruption in a Service Provider Network.
  • Various deficiencies in the prior art are addressed by systems, methods, apparatus, mechanism, telecom network elements and the like for managing MPLS-TE loading, such as by detecting, responding to, and otherwise managing MPLS-TE loading conditions a manner adapted to minimize service impact.
  • Various embodiments provide a mechanism of alerting other routers, network elements or nodes in an MPLS-TE Domain so that they may avoid using an overloaded node in subsequent new MPLS-TE LSP path computations.
  • Various embodiments are directed toward propagating information indicative of a MPLS TE-Overload condition.
  • various MPLS/RSVP tasks inform a routing protocol (e.g., OSPF, IS-IS and the like) about this state.
  • the routing protocol in turn communicates the overload condition to the nodes in the MPLS TE routing domain by inserting a new flag or bit value in an OSPF Router Information Capability TLV (if using OSPF) or an IS-IS Router Capability TLV (if using IS-IS).
  • a method for managing MPLS-TE loading comprises: monitoring a utilization level of a label switch router (LSR) associated with one or more label switched path (LSPs); and in response to a said utilization level being indicative of a MPLS-TE overload condition, transmitting an overload message toward an Interior Gateway Protocol (IGP), said overload message adapted to cause said IGP to advertise said overload condition.
  • LSR label switch router
  • IGP Interior Gateway Protocol
  • the utilization level may be associated with one or more of a memory utilization level, a central processing unit (CPU) utilization level, an input/output utilization level, a number of received RSVP packets, a rate of RSVP packet reception, a number of dropped RSVP packets and a rate of dropped RSVP packets.
  • FIG. 1 depicts a high-level block diagram of a system benefiting from various embodiments
  • FIG. 2 depicts a flow diagram of a method according to one embodiment
  • FIG. 3 depicts a high-level block diagram of a computing device suitable for use in performing functions described herein.
  • Various embodiments provide systems, methods and/or apparatus for detecting, responding to, and otherwise managing MPLS-TE Overload conditions a manner adapted to minimize service impact.
  • various embodiments are directed toward propagating information indicative of a MPLS TE-Overload condition.
  • various MPLS/RSVP tasks inform a routing protocol (e.g., OSPF, IS-IS and the like) about this state.
  • the routing protocol in turn communicates the overload condition to the nodes in the MPLS TE routing domain by inserting new flag or bit value in an OSPF Router Information Capability TLV (if using OSPF) or an IS-IS Router Capability TLV (if using IS-IS).
  • FIG. 1 depicts a high-level block diagram of a communication network benefiting from various embodiments.
  • the network 100 of FIG. 1 provides a Multi-Protocol Label Switching (MPLS) network supporting Resource Reservation Protocol (RSVP).
  • MPLS Multi-Protocol Label Switching
  • RSVP Resource Reservation Protocol
  • the network may be modified by those skilled in the art to use other MPLS related protocols rather that the exemplary protocol discussed herein.
  • exemplary network 100 includes a plurality of nodes 110 1 - 110 7 (collectively, nodes 110 ) that are interconnected via a plurality of communication links 120 (collectively, communication links 120 ).
  • the network 100 is managed by a management system 130 , which may provide any suitable management functions for the network 100 . While the network 100 may compr9sebe any suitable type of network and, thus, the nodes 110 may be any suitable types of nodes.
  • the network 102 may be an MPLS network in which nodes 110 are label switching routers (LSRs).
  • the nodes 110 are configured for transporting traffic within the network 102 .
  • the nodes 110 may transport traffic within network 102 using any suitable protocols (e.g., Internet Protocol (IP), MPLS, and the like, as well as various combinations thereof).
  • IP Internet Protocol
  • MPLS MPLS
  • the nodes 110 are configured to collect link state information associated with the communication link(s) 120 to which each node 110 is connected.
  • the nodes 110 are further configured to flood the collected link state information within network 102 .
  • each node 110 receives link state information associated with network 102 and, thus, each node 110 is able to maintain a database including information suitable for use in computing paths (e.g., network topology information, link state information, and the like).
  • This type of database is typically referred to as a Traffic Engineering (TE) database.
  • the nodes 110 also may be configured to store link constraints for use in computing paths for network 102 .
  • the link constraints may include any suitable link constraints which may be evaluated within the context of path computation.
  • the link constraints may include one or more of a link utilization for the link, a minimum link capacity required for a link, a maximum link bandwidth allowed for a link, a link cost associated with a link, an administrative constraint associated with the link, and the like, as well as various combinations thereof.
  • the link constraints may be configured on the nodes 110 in any suitable manner.
  • the link constraints may be pre-configured on the nodes 110 (e.g., automatically and/or by administrators), specified when requesting path computation or establishment, and the like, as well as various combinations thereof.
  • the link constraints may be provided to the nodes 110 , for storage on the nodes 110 , from any suitable source(s) of link constraints (e.g., a management system such as MS 130 , or any other suitable source).
  • link constraints may not be stored on the nodes 110 .
  • link constraints may only be available to the device(s) computing the paths.
  • At least a portion of the nodes 110 may be configured to operate as ingress nodes into network 102 and, similarly, at least a portion of the nodes 110 may be configured to operate as egress nodes from network 102 .
  • node 110 1 operates as an ingress node for the path
  • node 110 7 operates as an egress node for the path.
  • each of the nodes 110 may operate as an ingress node only, an egress node only, or both an ingress and egress node (e.g., for different traffic flows).
  • each of the nodes 110 may be configured to operate as an ingress node and/or as an egress node
  • each node 110 configured to operate as an ingress node may be referred to as an ingress node 110 and each node 110 configured to operate as an egress node may be referred to as an egress node 110 .
  • the ingress nodes 110 each are configured for computing paths to egress nodes 110 , thereby enabling establishment of connections, from the ingress nodes 110 to the egress nodes 110 , configured for transporting traffic via the network 102 .
  • the ingress nodes 110 in response to path computation requests, compute the requested paths based on the network information (e.g., network topology, link state, and the like, which may be available in a TE database and/or any other suitable database or databases) and link constraints available to the ingress nodes 110 , respectively.
  • the ingress nodes 110 upon computation of paths, may then initiate establishment of connections using the computed paths.
  • the ingress nodes 110 may then transmit information to the egress nodes 110 via the established connections, at which point the egress nodes 110 may then forward the information to other networks and devices.
  • MS 130 is configured for computing paths from ingress nodes 110 to egress nodes 110 , thereby enabling establishing of connections, from the ingress nodes 110 to the egress nodes 110 , configured for transporting traffic via the network 102 .
  • the MS 130 in response to path computation requests, computes the requested paths based on the network information (e.g., network topology, link state, and the like, which may be available in a TE database and/or any other suitable database or databases) and link constraints available to MS 130 .
  • the MS 130 upon computing a path, transmits path configuration information for the computed path to the relevant nodes 110 , where the path configuration information may be used to establish a connection via the computed path within network 102 .
  • the ingress node 110 of the computed path may then transmit information to the egress node 110 via the connection, at which point the egress node 110 may then forward the information to other networks and devices.
  • the network 102 comprises an MPLS network in which nodes 110 are label switching routers (LSRs) operating according to Multi-Protocol Label Switching (MPLS) Label Distribution Protocol (LDP).
  • LSRs label switching routers
  • MPLS Multi-Protocol Label Switching
  • LDP Label Distribution Protocol
  • FIG. 2 depicts a flow diagram of a method according to one embodiment. Specifically, FIG. 2 depicts a flow diagram of a method for managing MPLS-TE Overload conditions a manner adapted to minimize service impact.
  • the method 200 of FIG. 2 contemplates that some or all of a plurality of label switching routers (LSRs) associated with various label switched paths (LSPs) through a MPLS network operate to monitor various operating parameters to determine thereby whether a MPLS-TE overload condition exists or is imminent.
  • LSRs label switching routers
  • LSPs label switched paths
  • a LSP is established between an ingress node and an egress node.
  • the established LSP further supports upstream and downstream messages between the various LSRs.
  • RSVP path message is propagated downstream toward the egress node while RSVP Resv messages are propagated upstream toward the ingress node.
  • resource utilization is monitored to determine if a MPLS-TE overload condition exists or is imminent. Referring to box 225 , this determination may be made with respect to memory, CPU, input/output or other resources, a number of received RSVP packets, a rate of RSVP packet reception, a number of dropped RSVP packet, the rate at which RSVP packets are dropped, one or more resource utilization threshold levels and/or other mechanisms. For example, a MPLS/RSVP task processing mechanism at the node or LSR continues polling is system resource utilization and/or RSVP packet reception statistics.
  • the MPLS/RSVP task processing mechanism (or other mechanism) at the node or LSR informs the IGP of the overload condition.
  • the IGP advertises the MPLS-TE overload condition to routers within the MPLS domain.
  • such advertising is performed via a new or predefined flag or bit setting within a IGP router capability TLV or sub-TLV, such as a Open Shortest Path First (OSPF) routing protocol, Intermediate System To Intermediate System (IS-IS) routing protocol and the like.
  • OSPF Open Shortest Path First
  • IS-IS Intermediate System To Intermediate System
  • Other IGP advertising mechanisms may also be used. Further, other types of IGP may also be used.
  • IS-IS CAPABILITY TLV is composed of 1 octet specifying the number of bytes in the value field, and a variable length value field, starting with 4 octets of Router ID, indicating the source of the TLV, and followed by 1 octet of flags. A set of optional sub-TLVs may follow the flag field.
  • Sub-TLVs are formatted as described in IETF Request for Comment (RFC) 3784.
  • RRC Request for Comment
  • Router Informational Capabilities TLV includes a “value” field comprising a variable length sequence of capability bits rounded to a multiple of 4 octets padded with undefined bits.
  • Various embodiments use assigned or unassigned bits or flags within the value field (or other fields) to indicate an overload condition.
  • a MPLS/RSVP task processing mechanism at the node or LSR may inform IGP of the overload condition.
  • the IGP in turn advertises this condition by adapting or setting to a first state a flag or bit setting of a OSPF router info capability TLV, IS-IS router info capability TLV, other TLV, existing LSP attribute and the like.
  • Various embodiments are adapted to propagating information indicative of a MPLS-TE overload condition upstream to a head-end router (such as an ingress LSP, ABR and the like) adapted to cause the head end router to initiate or trigger a reroute (if desired) of one or more LSPs supported by transit or egress LSRs.
  • a head-end router receiving information indicative of a downstream MPLS-TE overload condition may request re-routing for any existing TE-LSP transiting the overloaded egress (or transit) nodes or LSRs forming the TE-LSP.
  • Suitable mechanisms for requesting rerouting exist, including those described in more detail in various Internet Engineering Task Force (IETF) Request for Comment (RFC), such as RFC5710 (PathErr Message Triggered MPLS and GMPLS LSP Reroutes).
  • IETF Internet Engineering Task Force
  • RFC5710 PathErr Message Triggered MPLS and GMPLS LSP Reroutes.
  • the head-end router when performing a path computation for any new TE-LSP, the head-end router should avoid a router advertising an MPLS-TE overload condition if possible. In this manner, for existing or new MPLS-TE LSPs associated with an overloaded router, one or more head-end routers operate to reduce the RSVP load (resource load) associated with the overloaded router. In this manner, the resources of the overloaded router are conserved such that existing RSVP sessions may quickly return to a normal working state.
  • RSVP load resource load
  • an IGP advertised overload condition operates to inhibit other LSRs from routing new LSPs through an overloaded LSR.
  • an IGP advertised non-overload condition operates to enable other LSRs to route existing and new LSPs to the non-overloaded LSR.
  • the MPLS/RSVP task processing mechanism (or other mechanism) at the node or LSR informs the IGP of the non-overload condition.
  • the IGP advertises the MPLS-TE non-overload condition to routers within the MPLS domain in a manner similar to that described above with respect to step 240 .
  • a MPLS/RSVP task processing mechanism at the node or LSR may inform IGP of the non-overload condition.
  • the IGP in turn advertises this condition by adapting or setting to a second state a flag or bit setting of a OSPF router info capability TLV, IS-IS router info capability TLV, other TLV, existing LSP attribute and the like.
  • a node entering a MPLS-TE Overloaded state informs the IGP of this state such that the IGP advertises the overload state to all the nodes in the in the MPLS-TE domain by, illustratively, setting a MPLS-TE overload flag or bit in a corresponding TLV or sub-TLV.
  • a node exiting a MPLS-TE Overloaded state informs the IGP of this state such that the IGP advertises the normal state to all the nodes in the in the MPLS-TE domain by, illustratively, resetting a MPLS-TE overload flag or bit in a corresponding TLV or sub-TLV.
  • the IGP advertised overload condition is adapted to cause other LSRs to reroute existing LSPs around an overloaded LSR and/or routing new LSPs through an overloaded LSR.
  • an IGP advertised non-overload condition is adapted to cause other LSRs to again route existing or new LSPs through a previously overloaded LSR is such routing is appropriate in terms of cost constraints, path management criteria and so on.
  • FIG. 3 depicts a high-level block diagram of a computing device, such as a processor in a telecom network element, suitable for use in performing functions described herein, such as the various network management functions, LSR functions, encapsulation functions, routing/path functions and so on associated with the various elements described above with respect to the figures.
  • a computing device such as a processor in a telecom network element
  • functions described herein such as the various network management functions, LSR functions, encapsulation functions, routing/path functions and so on associated with the various elements described above with respect to the figures.
  • computing device 300 includes a processor element 303 (e.g., a central processing unit (CPU) and/or other suitable processor(s)), a memory 304 (e.g., random access memory (RAM), read only memory (ROM), and the like), a cooperating module/process 305 , and various input/output devices 306 (e.g., a user input device (such as a keyboard, a keypad, a mouse, and the like), a user output device (such as a display, a speaker, and the like), an input port, an output port, a receiver, a transmitter, and storage devices (e.g., a persistent solid state drive, a hard disk drive, a compact disk drive, and the like)).
  • processor element 303 e.g., a central processing unit (CPU) and/or other suitable processor(s)
  • memory 304 e.g., random access memory (RAM), read only memory (ROM), and the like
  • cooperating module/process 305 e.g.,
  • cooperating process 305 can be loaded into memory 304 and executed by processor 303 to implement the functions as discussed herein.
  • cooperating process 305 (including associated data structures) can be stored on a computer readable storage medium, e.g., RAM memory, magnetic or optical drive or diskette, and the like.
  • computing device 300 depicted in FIG. 3 provides a general architecture and functionality suitable for implementing functional elements described herein or portions of the functional elements described herein.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

A system, method and apparatus for detecting MPLS-TE overload conditions and informing a IGP routing protocol, wherein the IGP routing protocol communicates the overload condition to the nodes in the MPLS TE routing domain by inserting new flag or bit value in an OSPF Router Information Capability TLV or an IS-IS Router Capability TLV.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application claims the benefit of U.S. Provisional Patent Application Ser. No. 61/653,219, filed May 30, 2012, entitled TE-LSP SYSTEMS AND METHODS (Attorney Docket No. 811458-PSP) which application is incorporated herein by reference in its entirety.
  • FIELD OF THE INVENTION
  • The invention relates to the field of communication networks such as multi-protocol label switching (MPLS) networks and, more particularly but not exclusively, to resource overload detection and management mechanisms.
  • BACKGROUND
  • Multiprotocol Label Switching (MPLS) enables efficient delivery of a wide variety of differentiated, end-to-end services. Multiprotocol Label Switching (MPLS) traffic engineering (TE) provides a mechanism for selecting efficient paths across an MPLS network based on bandwidth considerations and administrative rules. Each label switching router maintains a TE link state database with a current network topology. Once a path is computed, TE is used to maintain a forwarding state along that path.
  • In the case of Resource Reservation Protocol (RSVP) Inter-Domain TE-LSPs, a router or other network element or node may experience a resource overutilization condition (i.e., insufficient memory, processor, input/output or other resources) in response to receiving a large number of RSVP Packets. Such a condition may result in the RSVP/MPLS process temporarily dropping RSVP packets to conserve resources. If the condition persists, then the node may start tearing down existing MPLS-TE LSPs to release resources, which in turn may lead to service interruption in a Service Provider Network.
  • SUMMARY
  • Various deficiencies in the prior art are addressed by systems, methods, apparatus, mechanism, telecom network elements and the like for managing MPLS-TE loading, such as by detecting, responding to, and otherwise managing MPLS-TE loading conditions a manner adapted to minimize service impact. Various embodiments provide a mechanism of alerting other routers, network elements or nodes in an MPLS-TE Domain so that they may avoid using an overloaded node in subsequent new MPLS-TE LSP path computations.
  • Various embodiments are directed toward propagating information indicative of a MPLS TE-Overload condition. In particular, upon detecting such a condition, various MPLS/RSVP tasks inform a routing protocol (e.g., OSPF, IS-IS and the like) about this state. The routing protocol in turn communicates the overload condition to the nodes in the MPLS TE routing domain by inserting a new flag or bit value in an OSPF Router Information Capability TLV (if using OSPF) or an IS-IS Router Capability TLV (if using IS-IS).
  • A method for managing MPLS-TE loading according to one embodiment comprises: monitoring a utilization level of a label switch router (LSR) associated with one or more label switched path (LSPs); and in response to a said utilization level being indicative of a MPLS-TE overload condition, transmitting an overload message toward an Interior Gateway Protocol (IGP), said overload message adapted to cause said IGP to advertise said overload condition.
  • The utilization level may be associated with one or more of a memory utilization level, a central processing unit (CPU) utilization level, an input/output utilization level, a number of received RSVP packets, a rate of RSVP packet reception, a number of dropped RSVP packets and a rate of dropped RSVP packets.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The teachings herein can be readily understood by considering the following detailed description in conjunction with the accompanying drawings, in which:
  • FIG. 1 depicts a high-level block diagram of a system benefiting from various embodiments;
  • FIG. 2 depicts a flow diagram of a method according to one embodiment;
  • FIG. 3 depicts a high-level block diagram of a computing device suitable for use in performing functions described herein.
  • To facilitate understanding, identical reference numerals have been used, where possible, to designate identical elements that are common to the figures.
  • DETAILED DESCRIPTION OF THE INVENTION
  • Various embodiments provide systems, methods and/or apparatus for detecting, responding to, and otherwise managing MPLS-TE Overload conditions a manner adapted to minimize service impact.
  • Generally speaking, various embodiments are directed toward propagating information indicative of a MPLS TE-Overload condition. In particular, upon detecting such a condition, various MPLS/RSVP tasks inform a routing protocol (e.g., OSPF, IS-IS and the like) about this state. The routing protocol in turn communicates the overload condition to the nodes in the MPLS TE routing domain by inserting new flag or bit value in an OSPF Router Information Capability TLV (if using OSPF) or an IS-IS Router Capability TLV (if using IS-IS).
  • FIG. 1 depicts a high-level block diagram of a communication network benefiting from various embodiments. Specifically, the network 100 of FIG. 1 provides a Multi-Protocol Label Switching (MPLS) network supporting Resource Reservation Protocol (RSVP). The network may be modified by those skilled in the art to use other MPLS related protocols rather that the exemplary protocol discussed herein.
  • As depicted in FIG. 1, exemplary network 100 includes a plurality of nodes 110 1-110 7 (collectively, nodes 110) that are interconnected via a plurality of communication links 120 (collectively, communication links 120). The network 100 is managed by a management system 130, which may provide any suitable management functions for the network 100. While the network 100 may compr9sebe any suitable type of network and, thus, the nodes 110 may be any suitable types of nodes. For example, the network 102 may be an MPLS network in which nodes 110 are label switching routers (LSRs).
  • The nodes 110 are configured for transporting traffic within the network 102. The nodes 110 may transport traffic within network 102 using any suitable protocols (e.g., Internet Protocol (IP), MPLS, and the like, as well as various combinations thereof).
  • The nodes 110 are configured to collect link state information associated with the communication link(s) 120 to which each node 110 is connected. The nodes 110 are further configured to flood the collected link state information within network 102.
  • In one embodiment, the collection and flooding of link state information is performed using an Interior Gateway Protocol (IGP) supporting link-state, such as Open Shortest Path First (OSPF), Intermediate System to Intermediate System (IS-IS), or any other suitable protocol. In this manner, each node 110 receives link state information associated with network 102 and, thus, each node 110 is able to maintain a database including information suitable for use in computing paths (e.g., network topology information, link state information, and the like). This type of database is typically referred to as a Traffic Engineering (TE) database. The nodes 110 also may be configured to store link constraints for use in computing paths for network 102.
  • The link constraints may include any suitable link constraints which may be evaluated within the context of path computation. For example, the link constraints may include one or more of a link utilization for the link, a minimum link capacity required for a link, a maximum link bandwidth allowed for a link, a link cost associated with a link, an administrative constraint associated with the link, and the like, as well as various combinations thereof.
  • The link constraints may be configured on the nodes 110 in any suitable manner. For example, the link constraints may be pre-configured on the nodes 110 (e.g., automatically and/or by administrators), specified when requesting path computation or establishment, and the like, as well as various combinations thereof. In such embodiments, the link constraints may be provided to the nodes 110, for storage on the nodes 110, from any suitable source(s) of link constraints (e.g., a management system such as MS 130, or any other suitable source).
  • Although primarily depicted and described herein with respect to embodiments in which link constraints are configured on the nodes 110, in other embodiments the link constraints may not be stored on the nodes 110. For example, in embodiments in which path computation is performed by a device or devices other than nodes 110 (e.g., by a management system, such as MS 130), link constraints may only be available to the device(s) computing the paths.
  • In network 102, at least a portion of the nodes 110 may be configured to operate as ingress nodes into network 102 and, similarly, at least a portion of the nodes 110 may be configured to operate as egress nodes from network 102. In FIG. 1, for example, for a given path between node 110 1 and node 110 7, node 110 1 operates as an ingress node for the path and node 110 7 operates as an egress node for the path. It will be appreciated that each of the nodes 110 may operate as an ingress node only, an egress node only, or both an ingress and egress node (e.g., for different traffic flows).
  • As each of the nodes 110 may be configured to operate as an ingress node and/or as an egress node, each node 110 configured to operate as an ingress node may be referred to as an ingress node 110 and each node 110 configured to operate as an egress node may be referred to as an egress node 110.
  • In one embodiment, the ingress nodes 110 each are configured for computing paths to egress nodes 110, thereby enabling establishment of connections, from the ingress nodes 110 to the egress nodes 110, configured for transporting traffic via the network 102. The ingress nodes 110, in response to path computation requests, compute the requested paths based on the network information (e.g., network topology, link state, and the like, which may be available in a TE database and/or any other suitable database or databases) and link constraints available to the ingress nodes 110, respectively. The ingress nodes 110, upon computation of paths, may then initiate establishment of connections using the computed paths. The ingress nodes 110 may then transmit information to the egress nodes 110 via the established connections, at which point the egress nodes 110 may then forward the information to other networks and devices.
  • In one embodiment, MS 130 is configured for computing paths from ingress nodes 110 to egress nodes 110, thereby enabling establishing of connections, from the ingress nodes 110 to the egress nodes 110, configured for transporting traffic via the network 102. The MS 130, in response to path computation requests, computes the requested paths based on the network information (e.g., network topology, link state, and the like, which may be available in a TE database and/or any other suitable database or databases) and link constraints available to MS 130. The MS 130, upon computing a path, transmits path configuration information for the computed path to the relevant nodes 110, where the path configuration information may be used to establish a connection via the computed path within network 102. The ingress node 110 of the computed path may then transmit information to the egress node 110 via the connection, at which point the egress node 110 may then forward the information to other networks and devices.
  • In various embodiments, the network 102 comprises an MPLS network in which nodes 110 are label switching routers (LSRs) operating according to Multi-Protocol Label Switching (MPLS) Label Distribution Protocol (LDP).
  • FIG. 2 depicts a flow diagram of a method according to one embodiment. Specifically, FIG. 2 depicts a flow diagram of a method for managing MPLS-TE Overload conditions a manner adapted to minimize service impact. The method 200 of FIG. 2 contemplates that some or all of a plurality of label switching routers (LSRs) associated with various label switched paths (LSPs) through a MPLS network operate to monitor various operating parameters to determine thereby whether a MPLS-TE overload condition exists or is imminent.
  • At step 210, a LSP is established between an ingress node and an egress node. Referring to box 215, the established LSP further supports upstream and downstream messages between the various LSRs. In particular, RSVP path message is propagated downstream toward the egress node while RSVP Resv messages are propagated upstream toward the ingress node.
  • At step 220, at one or more of the egress (or transit) nodes or LSRs forming the LSP, resource utilization is monitored to determine if a MPLS-TE overload condition exists or is imminent. Referring to box 225, this determination may be made with respect to memory, CPU, input/output or other resources, a number of received RSVP packets, a rate of RSVP packet reception, a number of dropped RSVP packet, the rate at which RSVP packets are dropped, one or more resource utilization threshold levels and/or other mechanisms. For example, a MPLS/RSVP task processing mechanism at the node or LSR continues polling is system resource utilization and/or RSVP packet reception statistics.
  • At step 230, in response to a determination that a MPLS-TE overload condition exists or is imminent at a particular node or LSR, the MPLS/RSVP task processing mechanism (or other mechanism) at the node or LSR informs the IGP of the overload condition.
  • At step 240, the IGP advertises the MPLS-TE overload condition to routers within the MPLS domain. Referring to box 245, such advertising is performed via a new or predefined flag or bit setting within a IGP router capability TLV or sub-TLV, such as a Open Shortest Path First (OSPF) routing protocol, Intermediate System To Intermediate System (IS-IS) routing protocol and the like. Other IGP advertising mechanisms may also be used. Further, other types of IGP may also be used.
  • Various embodiments described herein utilize IGP advertising mechanisms conforming to IS-IS CAPABILITY TLVs and sub-TLVs such as described in more detail in Internet Engineering Task Force (IETF) document “IS-IS Extensions for Advertising Router Information.” As defined therein, the IS-IS router CAPABILITY TLV is composed of 1 octet specifying the number of bytes in the value field, and a variable length value field, starting with 4 octets of Router ID, indicating the source of the TLV, and followed by 1 octet of flags. A set of optional sub-TLVs may follow the flag field. Sub-TLVs are formatted as described in IETF Request for Comment (RFC) 3784. Various embodiments use assigned or unassigned bits or flags within a value field (or other fields) to indicate an overload condition.
  • Various embodiments described herein utilize IGP advertising mechanisms conforming to OSPF CAPABILITY TLVs and sub-TLVs such as described in more detail in IETF RFC 4970. The format of the Router Informational Capabilities TLV includes a “value” field comprising a variable length sequence of capability bits rounded to a multiple of 4 octets padded with undefined bits. Various embodiments use assigned or unassigned bits or flags within the value field (or other fields) to indicate an overload condition.
  • For example, upon determining that a MPLS-TE overload condition exists or is imminent, a MPLS/RSVP task processing mechanism at the node or LSR may inform IGP of the overload condition. The IGP in turn advertises this condition by adapting or setting to a first state a flag or bit setting of a OSPF router info capability TLV, IS-IS router info capability TLV, other TLV, existing LSP attribute and the like.
  • Various embodiments are adapted to propagating information indicative of a MPLS-TE overload condition upstream to a head-end router (such as an ingress LSP, ABR and the like) adapted to cause the head end router to initiate or trigger a reroute (if desired) of one or more LSPs supported by transit or egress LSRs. A head-end router receiving information indicative of a downstream MPLS-TE overload condition may request re-routing for any existing TE-LSP transiting the overloaded egress (or transit) nodes or LSRs forming the TE-LSP. Suitable mechanisms for requesting rerouting exist, including those described in more detail in various Internet Engineering Task Force (IETF) Request for Comment (RFC), such as RFC5710 (PathErr Message Triggered MPLS and GMPLS LSP Reroutes).
  • Generally speaking, when performing a path computation for any new TE-LSP, the head-end router should avoid a router advertising an MPLS-TE overload condition if possible. In this manner, for existing or new MPLS-TE LSPs associated with an overloaded router, one or more head-end routers operate to reduce the RSVP load (resource load) associated with the overloaded router. In this manner, the resources of the overloaded router are conserved such that existing RSVP sessions may quickly return to a normal working state.
  • Thus, in various embodiments, an IGP advertised overload condition operates to inhibit other LSRs from routing new LSPs through an overloaded LSR. Similarly, an IGP advertised non-overload condition operates to enable other LSRs to route existing and new LSPs to the non-overloaded LSR.
  • At step 250, in response to a determination that a MPLS-TE overload condition no longer exists or is imminent at a particular node or LSR, the MPLS/RSVP task processing mechanism (or other mechanism) at the node or LSR informs the IGP of the non-overload condition. The IGP advertises the MPLS-TE non-overload condition to routers within the MPLS domain in a manner similar to that described above with respect to step 240.
  • For example, upon determining that a MPLS-TE overload condition no longer exists or is no longer imminent, a MPLS/RSVP task processing mechanism at the node or LSR may inform IGP of the non-overload condition. The IGP in turn advertises this condition by adapting or setting to a second state a flag or bit setting of a OSPF router info capability TLV, IS-IS router info capability TLV, other TLV, existing LSP attribute and the like.
  • Thus, in one embodiment, a node entering a MPLS-TE Overloaded state informs the IGP of this state such that the IGP advertises the overload state to all the nodes in the in the MPLS-TE domain by, illustratively, setting a MPLS-TE overload flag or bit in a corresponding TLV or sub-TLV. Similarly, a node exiting a MPLS-TE Overloaded state (i.e., returning to a normal state) informs the IGP of this state such that the IGP advertises the normal state to all the nodes in the in the MPLS-TE domain by, illustratively, resetting a MPLS-TE overload flag or bit in a corresponding TLV or sub-TLV.
  • Thus, the IGP advertised overload condition is adapted to cause other LSRs to reroute existing LSPs around an overloaded LSR and/or routing new LSPs through an overloaded LSR. Similarly, an IGP advertised non-overload condition is adapted to cause other LSRs to again route existing or new LSPs through a previously overloaded LSR is such routing is appropriate in terms of cost constraints, path management criteria and so on.
  • FIG. 3 depicts a high-level block diagram of a computing device, such as a processor in a telecom network element, suitable for use in performing functions described herein, such as the various network management functions, LSR functions, encapsulation functions, routing/path functions and so on associated with the various elements described above with respect to the figures.
  • As depicted in FIG. 3, computing device 300 includes a processor element 303 (e.g., a central processing unit (CPU) and/or other suitable processor(s)), a memory 304 (e.g., random access memory (RAM), read only memory (ROM), and the like), a cooperating module/process 305, and various input/output devices 306 (e.g., a user input device (such as a keyboard, a keypad, a mouse, and the like), a user output device (such as a display, a speaker, and the like), an input port, an output port, a receiver, a transmitter, and storage devices (e.g., a persistent solid state drive, a hard disk drive, a compact disk drive, and the like)).
  • It will be appreciated that the functions depicted and described herein may be implemented in software and/or in a combination of software and hardware, e.g., using a general purpose computer, one or more application specific integrated circuits (ASIC), and/or any other hardware equivalents. In one embodiment, the cooperating process 305 can be loaded into memory 304 and executed by processor 303 to implement the functions as discussed herein. Thus, cooperating process 305 (including associated data structures) can be stored on a computer readable storage medium, e.g., RAM memory, magnetic or optical drive or diskette, and the like.
  • It will be appreciated that computing device 300 depicted in FIG. 3 provides a general architecture and functionality suitable for implementing functional elements described herein or portions of the functional elements described herein.
  • It is contemplated that some of the steps discussed herein as software methods may be implemented within hardware, for example, as circuitry that cooperates with the processor to perform various method steps. Portions of the functions/elements described herein may be implemented as a computer program product wherein computer instructions, when processed by a computing device, adapt the operation of the computing device such that the methods and/or techniques described herein are invoked or otherwise provided. Instructions for invoking the inventive methods may be stored in tangible and non-transitory computer readable medium such as fixed or removable media or memory, transmitted via a tangible or intangible data stream in a broadcast or other signal bearing medium, and/or stored within a memory within a computing device operating according to the instructions.
  • Although various embodiments which incorporate the teachings of the present invention have been shown and described in detail herein, those skilled in the art can readily devise many other varied embodiments that still incorporate these teachings. Thus, while the foregoing is directed to various embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof. As such, the appropriate scope of the invention is to be determined according to the claims.

Claims (20)

What is claimed is:
1. A method for managing MPLS-TE loading, comprising:
monitoring a utilization level of a label switch router (LSR) associated with one or more label switched path (LSPs); and
in response to a determination of said utilization level being indicative of a MPLS-TE overload condition, transmitting an overload message toward an Interior Gateway Protocol (IGP) router, said overload message adapted to cause IGP router to advertise said overload condition. 10
2. The method of claim 1, wherein said utilization level is associated with one or more of a memory utilization level, a central processing unit (CPU) utilization level and an input/output utilization level.
3. The method of claim 1, wherein said utilization level is associated with one or more of a number of received RSVP packets, a rate of RSVP packet reception, a number of dropped RSVP packets and a rate of dropped RSVP packets.
4. The method of claim 1, wherein IGP is adapted to advertise said overload condition to routers within the IGP domain via a flag setting or bit state within a Router Information Capability TLV or sub-TLV.
5. The method of claim 4, wherein said IGP comprises an Open Shortest Path First (OSPF) protocol.
6. The method of claim 5, wherein said OSPF protocol advertises said overload condition via a flag setting or bit having a first state within a OSPF Router Information Capability Type-Length-Value (TLV) or sub-TLV. 30
7. The method of claim 1, wherein said IGP comprises an Intermediate System to Intermediate System (IS-IS) protocol.
8. The method of claim 7, wherein said IS-IS protocol advertises said overload condition via a flag setting or bit having a first state within a IS-IS Router Information Capability TLV or sub-TLV.
9. The method of claim 1, further comprising:
in response to a determination that said utilization level is no longer indicative of a MPLS-TE overload condition, transmitting an non-overload message toward an Interior Gateway Protocol (IGP) router, said non-overload message adapted to cause IGP to advertise said non-overload condition.
10. The method of claim 9, wherein said utilization level is associated with one or more of a memory utilization level, a central processing unit (CPU) utilization level, an input/output utilization level, a number of received RSVP packets, a rate of RSVP packet reception, a number of dropped RSVP packets and a rate of dropped RSVP packets.
11. The method of claim 5, wherein said OSPF protocol advertises a non-overload condition via said flag setting or bit having a second state within a OSPF Router Information Capability Type-Length-Value (TLV) or sub-TLV.
12. The method of claim 7, wherein said IS-IS protocol advertises a non-overload condition via said flag setting or bit having a second state within a IS-IS Router Information Capability TLV or sub-TLV.
13. The method of claim 9, wherein:
an OSPF protocol advertises an overload condition via a flag setting or bit having a first state within a OSPF Router Information Capability Type-Length-Value (TLV) or sub-TLV; and
said OSPF protocol advertises a non-overload condition via said flag setting or bit having a second state within said OSPF Router Information Capability Type-Length-Value (TLV) or sub-TLV.
14. The method of claim 9, wherein:
an IS-IS protocol advertises an overload condition via a flag setting or bit having a first state within a IS-IS Router Information Capability Type-Length-Value (TLV) or sub-TLV; and
said IS-IS protocol advertises a non-overload condition via said flag setting or bit having a second state within said IS-IS Router Information Capability Type-Length-Value (TLV) or sub-TLV.
15. The method of claim 1, wherein said IGP advertised overload condition is adapted to cause other LSRs to reroute existing LSPs around an overloaded LSR.
16. The method of claim 1, wherein said IGP advertised overload condition inhibiting other LSRs from routing new LSPs through said overloaded LSR.
17. The method of claim 9, wherein said IGP advertised non-overload enabling other LSRs to route existing and new LSPs to the non-overloaded LSR.
18. A telecom network element for managing MPLS-TE loading, comprising a processor configured for:
monitoring a utilization level of a label switch router (LSR) associated with one or more label switched path (LSPs); and
in response to a determination of said utilization level being indicative of a MPLS-TE overload condition, transmitting an overload message toward an Interior Gateway Protocol (IGP) router, said overload message adapted to cause IGP to advertise said overload condition.
19. A computer program product wherein computer instructions, when executed by a processor in a telecom network element, adapt the operation of the telecom network element to perform a method for managing MPLS-TE loading, the method comprising:
monitoring a utilization level of a label switch router (LSR) associated with one or more label switched path (LSPs); and
in response to a determination of said utilization level being indicative of a MPLS-TE overload condition, transmitting an overload message toward an Interior Gateway Protocol (IGP) router, said overload message adapted to cause IGP to advertise said overload condition.
20. A tangible and non-transient computer readable storage medium storing instructions which, when executed by a computer, adapt the operation of the computer to provide a method for managing MPLS-TE loading, the method comprising:
monitoring a utilization level of a label switch router (LSR) associated with one or more label switched path (LSPs); and
in response to a determination of said utilization level being indicative of a MPLS-TE overload condition, transmitting an overload message toward an Interior Gateway Protocol (IGP) router, said overload message adapted to cause IGP to advertise said overload condition.
US13/838,673 2012-05-30 2013-03-15 System and method for managing mpls-te overload Abandoned US20130322254A1 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
US13/838,673 US20130322254A1 (en) 2012-05-30 2013-03-15 System and method for managing mpls-te overload
PCT/US2013/042831 WO2013181130A1 (en) 2012-05-30 2013-05-28 System and method for managing mpls-te overload
EP13728880.9A EP2856715B1 (en) 2012-05-30 2013-05-28 System and method for managing mpls-te overload
JP2015515107A JP2015518358A (en) 2012-05-30 2013-05-28 System and method for managing MPLS-TE overload
KR20147033610A KR20150010968A (en) 2012-05-30 2013-05-28 System and method for managing mpls-te overload
CN201380028224.7A CN104521194B (en) 2012-05-30 2013-05-28 The system and method for managing MPLS TE overloads

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201261653219P 2012-05-30 2012-05-30
US13/838,673 US20130322254A1 (en) 2012-05-30 2013-03-15 System and method for managing mpls-te overload

Publications (1)

Publication Number Publication Date
US20130322254A1 true US20130322254A1 (en) 2013-12-05

Family

ID=49670125

Family Applications (4)

Application Number Title Priority Date Filing Date
US13/623,172 Active 2033-01-13 US8948051B2 (en) 2012-05-30 2012-09-20 System and method for efficient MVPN source redundancy with S-PMSI
US13/837,531 Expired - Fee Related US9571381B2 (en) 2012-05-30 2013-03-15 System and method for inter-domain RSVP-TE LSP load balancing
US13/838,673 Abandoned US20130322254A1 (en) 2012-05-30 2013-03-15 System and method for managing mpls-te overload
US14/041,851 Expired - Fee Related US9148364B2 (en) 2012-05-30 2013-09-30 System, method and apparatus for efficient management of S-PMSI resource in RSVP P2MP multicast networks

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US13/623,172 Active 2033-01-13 US8948051B2 (en) 2012-05-30 2012-09-20 System and method for efficient MVPN source redundancy with S-PMSI
US13/837,531 Expired - Fee Related US9571381B2 (en) 2012-05-30 2013-03-15 System and method for inter-domain RSVP-TE LSP load balancing

Family Applications After (1)

Application Number Title Priority Date Filing Date
US14/041,851 Expired - Fee Related US9148364B2 (en) 2012-05-30 2013-09-30 System, method and apparatus for efficient management of S-PMSI resource in RSVP P2MP multicast networks

Country Status (6)

Country Link
US (4) US8948051B2 (en)
EP (1) EP2856715B1 (en)
JP (1) JP2015518358A (en)
KR (1) KR20150010968A (en)
CN (1) CN104521194B (en)
WO (1) WO2013181130A1 (en)

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8948051B2 (en) * 2012-05-30 2015-02-03 Alcatel Lucent System and method for efficient MVPN source redundancy with S-PMSI
US9118564B2 (en) * 2012-06-29 2015-08-25 Futurewei Technologies, Inc. Providing PIM-SM support for mRSVP-TE based multicast virtual private networks
US9106556B2 (en) * 2013-02-11 2015-08-11 Avaya Inc. Method to achieve the use of an external metric as the primary tie-breaker in intermediate system to intermediate system (ISIS) route selections
CN106330703A (en) * 2015-06-16 2017-01-11 中兴通讯股份有限公司 Route selection method and device, routing bridge and network
US9973389B2 (en) * 2015-12-21 2018-05-15 Juniper Networks, Inc. Propagating leaf count information to facilitate switching between a segmented tunnel and a non-segmented tunnel
CN107181689B (en) * 2016-03-10 2021-09-14 中兴通讯股份有限公司 Message interaction method and device between routers
US10104139B2 (en) 2016-03-31 2018-10-16 Juniper Networks, Inc. Selectively signaling selective tunnels in multicast VPNs
US10608931B1 (en) * 2016-11-28 2020-03-31 Juniper Networks, Inc. Selective multicasting of traffic flows in multiprotocol label switching virtual private networks
JP6996615B2 (en) * 2018-03-16 2022-01-17 日本電気株式会社 Multicast controller, multicast control method, and program
CN110535784B (en) * 2018-05-23 2021-01-15 北京三快在线科技有限公司 Traffic management method and device based on confidence interval and computing equipment
FR3081645A1 (en) 2018-06-28 2019-11-29 Orange COMMUNICATION METHOD IMPLEMENTED BY A FIRST ROUTER OF AN AUTONOMOUS SYSTEM USING AN INTERNAL ROUTING PROTOCOL
CN112104539B (en) * 2019-06-18 2022-01-11 华为技术有限公司 Information interaction method, device and system in segmented NGMVPN
CN114900462A (en) * 2020-02-28 2022-08-12 华为技术有限公司 A route matching method, information sending method and device
WO2022135217A1 (en) * 2020-12-21 2022-06-30 华为技术有限公司 Load sharing method and system, root node device, and leaf node device
US20240015095A1 (en) * 2022-07-11 2024-01-11 Juniper Networks, Inc. Designating a primary multicast flow and a backup multicast flow for multicast traffic

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060198308A1 (en) * 2005-03-02 2006-09-07 Jean-Philippe Vasseur Technique for selecting a path computation element based on response time delay
US20090238074A1 (en) * 2008-03-18 2009-09-24 Jean-Philippe Marcel Vasseur Dynamic reroute of network traffic

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7319700B1 (en) * 2000-12-29 2008-01-15 Juniper Networks, Inc. Communicating constraint information for determining a path subject to such constraints
JP3715596B2 (en) * 2002-07-11 2005-11-09 富士通株式会社 Wide area load balancing control system
US8717899B2 (en) * 2004-10-13 2014-05-06 Cisco Technology, Inc. System and method for reporting out-of-resources (OOR) conditions in a data network
US9306831B2 (en) 2005-02-14 2016-04-05 Cisco Technology, Inc. Technique for efficient load balancing of TE-LSPs
US8014291B2 (en) * 2006-11-28 2011-09-06 Cisco Technology, Inc. Relaxed constrained shortest path first (R-CSPF)
US8411691B2 (en) * 2009-01-12 2013-04-02 Juniper Networks, Inc. Transfer of mobile subscriber context in cellular networks using extended routing protocol
US8218553B2 (en) * 2009-02-25 2012-07-10 Juniper Networks, Inc. Load balancing network traffic on a label switched path using resource reservation protocol with traffic engineering
US8711863B2 (en) * 2009-04-27 2014-04-29 Ciena Corporation Virtual links in a routed ethernet mesh network
CN101656679B (en) * 2009-09-25 2012-04-04 华为技术有限公司 A multicast fast convergence method, router and communication system
US8422364B2 (en) * 2010-05-17 2013-04-16 Cisco Technology, Inc. Multicast label distribution protocol node protection
US9240942B2 (en) * 2011-02-28 2016-01-19 Cisco Technology, Inc. Bandwidth utilization for equal cost multiple paths
US8861342B2 (en) * 2011-10-28 2014-10-14 Cisco Technology, Inc. Multicast-only fast re-route processing for point-to-multipoint pseudowire
US8948051B2 (en) * 2012-05-30 2015-02-03 Alcatel Lucent System and method for efficient MVPN source redundancy with S-PMSI
US9001672B2 (en) * 2012-07-27 2015-04-07 Alcatel Lucent System, method and apparatus conforming path cost criteria across multiple ABRs
US20140269296A1 (en) * 2013-03-13 2014-09-18 Cox Communications, Inc. Systems and Methods of Bundled Label Switch Path for Load Splitting

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060198308A1 (en) * 2005-03-02 2006-09-07 Jean-Philippe Vasseur Technique for selecting a path computation element based on response time delay
US20090238074A1 (en) * 2008-03-18 2009-09-24 Jean-Philippe Marcel Vasseur Dynamic reroute of network traffic

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
LE ROUX (EDITOR) FRANCE TELECOM J P VASSEUR (EDITOR) CISCO SYSTEM INC YUICHI IKEJIRI NTT COMMUNICATIONS RAYMOND ZHANG BT INFONET J: "IG protocol extensions for Path Computation Element (PCE) Discovery; draft-ietf-pce-disco-proto-igp-02.txt", 20060601, vol. pce, no. 2, 1 June 2006 (2006-06-01), XP015045054, ISSN: 0000-0004 *

Also Published As

Publication number Publication date
CN104521194A (en) 2015-04-15
EP2856715A1 (en) 2015-04-08
US9148364B2 (en) 2015-09-29
EP2856715B1 (en) 2019-07-03
US9571381B2 (en) 2017-02-14
JP2015518358A (en) 2015-06-25
KR20150010968A (en) 2015-01-29
US20150092569A1 (en) 2015-04-02
WO2013181130A1 (en) 2013-12-05
CN104521194B (en) 2017-11-10
US20130322253A1 (en) 2013-12-05
US20130322291A1 (en) 2013-12-05
US8948051B2 (en) 2015-02-03

Similar Documents

Publication Publication Date Title
EP2856715B1 (en) System and method for managing mpls-te overload
US10721156B2 (en) Technique for selecting a path computation element based on response time delay
US9491046B2 (en) System and method for switching traffic from sub-optimal primary P2MP to standby P2MP
KR101576412B1 (en) System and method for implementing label switch router (lsr) overload protection
CN107566274B (en) Method, router and controller for bandwidth management
US7302494B2 (en) Traffic engineering method and node apparatus using traffic engineering method
CN102771096B (en) For calculating the system and method for the backup ingress of point-to-multipoint label switch paths
US9065750B2 (en) Congestion-based notification during fast reroute operations in stateful path computation element environments
US8259584B2 (en) Dynamic reroute of network traffic
CN101099351A (en) Used to trigger the packaging of path calculation requests
US8913490B1 (en) Selective notification for label switched path re-routing
US20140269737A1 (en) System, method and apparatus for lsp setup using inter-domain abr indication
CN103117949B (en) A kind of message transmitting method based on lsp tunnel and equipment
US20160315854A1 (en) System and method of signaling protection path bindings

Legal Events

Date Code Title Description
AS Assignment

Owner name: ALCATEL-LUCENT USA INC., NEW JERSEY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JAIN, PRADEEP G;SINGH, KANWAR D;SHAH, JAISHAL;AND OTHERS;SIGNING DATES FROM 20130315 TO 20130320;REEL/FRAME:030382/0212

AS Assignment

Owner name: ALCATEL LUCENT, FRANCE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ALCATEL-LUCENT USA INC.;REEL/FRAME:032743/0222

Effective date: 20140422

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载