IP - MPLS Design - v1.5
IP - MPLS Design - v1.5
IP - MPLS Design - v1.5
Wateen Telecom
Converged Multi-Service Network - IP/MPLS Low Level Design
Corporate Headquarters
Cisco Systems, Inc.
170 West Tasman Drive
San Jose, CA 95134-1706
USA
http://www.cisco.com
Tel: 408 526-4000
800 553-NETS (6387)
Fax: 408 526-4100
THE SPECIFICATIONS AND INFORMATION REGARDING THE PRODUCTS IN THIS MANUAL ARE SUBJECT TO CHANGE WITHOUT NOTICE. ALL STATEMENTS, INFORMATION, AND
RECOMMENDATIONS IN THIS MANUAL ARE BELIEVED TO BE ACCURATE BUT ARE PRESENTED WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED. USERS MUST TAKE FULL
RESPONSIBILITY FOR THEIR APPLICATION OF ANY PRODUCTS.
THE SOFTWARE LICENSE AND LIMITED WARRANTY FOR THE ACCOMPANYING PRODUCT ARE SET FORTH IN THE INFORMATION PACKET THAT SHIPPED WITH THE PRODUCT AND ARE
INCORPORATED HEREIN BY THIS REFERENCE. IF YOU ARE UNABLE TO LOCATE THE SOFTWARE LICENSE OR LIMITED WARRANTY, CONTACT YOUR CISCO REPRESENTATIVE FOR A
COPY.
The following information is for FCC compliance of Class A devices: This equipment has been tested and found to comply with the limits for a Class A digital device, pursuant to part 15 of the FCC rules. These limits
are designed to provide reasonable protection against harmful interference when the equipment is operated in a commercial environment. This equipment generates, uses, and can radiate radio-frequency energy and, if
not installed and used in accordance with the instruction manual, may cause harmful interference to radio communications. Operation of this equipment in a residential area is likely to cause harmful interference, in
which case users will be required to correct the interference at their own expense.
The following information is for FCC compliance of Class B devices: The equipment described in this manual generates and may radiate radio-frequency energy. If it is not installed in accordance with Cisco’s
installation instructions, it may cause interference with radio and television reception. This equipment has been tested and found to comply with the limits for a Class B digital device in accordance with the
specifications in part 15 of the FCC rules. These specifications are designed to provide reasonable protection against such interference in a residential installation. However, there is no guarantee that interference will
not occur in a particular installation.
You can determine whether your equipment is causing interference by turning it off. If the interference stops, it was probably caused by the Cisco equipment or one of its peripheral devices. If the equipment causes
interference to radio or television reception, try to correct the interference by using one or more of the following measures:
Move the equipment to one side or the other of the television or radio.
Plug the equipment into an outlet that is on a different circuit from the television or radio. (That is, make certain the equipment and the television or radio are on circuits controlled by different circuit breakers or fuses.)
Modifications to this product not authorized by Cisco Systems, Inc. could void the FCC approval and negate your authority to operate the product.
The following third-party software may be included with your product and will be subject to the software license agreement:
CiscoWorks software and documentation are based in part on HP OpenView under license from the Hewlett-Packard Company. HP OpenView is a trademark of the Hewlett-Packard Company. Copyright © 1992, 1993
Hewlett-Packard Company.
The Cisco implementation of TCP header compression is an adaptation of a program developed by the University of California, Berkeley (UCB) as part of UCB’s public domain version of the UNIX operating system.
All rights reserved. Copyright © 1981, Regents of the University of California.
Network Time Protocol (NTP). Copyright © 1992, David L. Mills. The University of Delaware makes no representations about the suitability of this software for any purpose.
Point-to-Point Protocol. Copyright © 1989, Carnegie-Mellon University. All rights reserved. The name of the University may not be used to endorse or promote products derived from this software without specific
prior written permission.
The Cisco implementation of TN3270 is an adaptation of the TN3270, curses, and termcap programs developed by the University of California, Berkeley (UCB) as part of the UCB’s public domain version of the UNIX
operating system. All rights reserved. Copyright © 1981-1988, Regents of the University of California.
Cisco incorporates Fastmac and TrueView software and the RingRunner chip in some Token Ring products. Fastmac software is licensed to Cisco by Madge Networks Limited, and the RingRunner chip is licensed to
Cisco by Madge NV. Fastmac, RingRunner, and TrueView are trademarks and in some jurisdictions registered trademarks of Madge Networks Limited. Copyright © 1995, Madge Networks Limited. All rights
reserved.
Xremote is a trademark of Network Computing Devices, Inc. Copyright © 1989, Network Computing Devices, Inc., Mountain View, California. NCD makes no representations about the suitability of this software for
any purpose.
The X Window System is a trademark of the X Consortium, Cambridge, Massachusetts. All rights reserved.
NOTWITHSTANDING ANY OTHER WARRANTY HEREIN, ALL DOCUMENT FILES AND SOFTWARE OF THESE SUPPLIERS ARE PROVIDED “AS IS” WITH ALL FAULTS. CISCO AND THE
ABOVE-NAMED SUPPLIERS DISCLAIM ALL WARRANTIES, EXPRESSED OR IMPLIED, INCLUDING, WITHOUT LIMITATION, THOSE OF MERCHANTABILITY, FITNESS FOR A PRACTICAL
PURPOSE AND NONINFRINGEMENT OR ARISING FROM A COURSE OF DEALING, USAGE, OR TRADE PRACTICE.
IN NO EVENT SHALL CISCO OR ITS SUPPLIERS BE LIABLE FOR ANY INDIRECT, SPECIAL, CONSEQUENTIAL, OR INCIDENTAL DAMAGES, INCLUDING, WITHOUT LIMITATION, LOST
PROFITS OR LOSS OR DAMAGE TO DATA ARISING OUT OF THE USE OR INABILITY TO USE THIS MANUAL, EVEN IF CISCO OR ITS SUPPLIERS HAVE BEEN ADVISED OF THE POSSIBILITY
OF SUCH DAMAGES.
AccessPath, AtmDirector, Browse with Me, CCIP, CCSI, CD-PAC, CiscoLink, the Cisco Powered Network logo, Cisco Systems Networking Academy, the Cisco Systems Networking Academy logo, Cisco Unity, Fast
Step, Follow Me Browsing, FormShare, FrameShare, IGX, Internet Quotient, IP/VC, iQ Breakthrough, iQ Expertise, iQ FastTrack, the iQ logo, iQ Net Readiness Scorecard, MGX, the Networkers logo, ScriptBuilder,
ScriptShare, SMARTnet, TransPath, Voice LAN, Wavelength Router, and WebViewer, Aironet, ASIST, BPX, Catalyst, CCDA, CCDP, CCIE, CCNA, CCNP, Cisco, the Cisco Certified Internetwork Expert logo,
Cisco IOS, the Cisco IOS logo, Cisco Press, Cisco Systems, Cisco Systems Capital, the Cisco Systems logo, Empowering the Internet Generation, Enterprise/Solver, EtherChannel, EtherSwitch, FastHub, FastSwitch,
GigaStack, IOS, IP/TV, LightStream, MICA, Network Registrar, Packet, PIX, Post-Routing, Pre-Routing, RateMUX, Registrar, SlideCast, StrataView Plus, Stratm, SwitchProbe, TeleRouter, and VCO are trademarks
or registered trademarks of Cisco Systems, Inc. and/or its affiliates in the U.S. and certain other countries.
All other trademarks mentioned in this document or Web site are the property of their respective owners. The use of the word partner does not imply a partnership relationship between Cisco and any other company.
(0110R).
Please refer to http://www.cisco.com/logo/ for the latest information on Cisco logos, branding and trademarks.
THIS DOCUMENT CONTAINS VALUABLE TRADE SECRETS AND CONFIDENTIAL INFORMATION OF CISCO SYSTEMS, INC. AND IT’S SUPPLIERS, AND SHALL NOT BE DISCLOSED TO ANY
PERSON, ORGANIZATION, OR ENTITY UNLESS SUCH DISCLOSURE IS SUBJECT TO THE PROVISIONS OF A WRITTEN NON-DISCLOSURE AND PROPRIETARY RIGHTS AGREEMENT OR
INTELLECTUAL PROPERTY LICENSE AGREEMENT APPROVED BY CISCO SYSTEMS, INC. THE DISTRIBUTION OF THIS DOCUMENT DOES NOT GRANT ANY LICENSE IN OR RIGHTS, IN
WHOLE OR IN PART, TO THE CONTENT, THE PRODUCT(S), TECHNOLOGY OF INTELLECTUAL PROPERTY DESCRIBED HEREIN.
Contents......................................................................................................................................... 3
Figures ......................................................................................................................................... 11
History..................................................................................................................................... 21
Review..................................................................................................................................... 22
Conclusion................................................................................................................................. 295
Figure 68 - Hub and Spoke Model with Connectivity between Spoke Sites 123
Figure 76 Configuring static routes redistribution into vrf context under BGP 130
Figure 124 Sample ingress ACL Configuration for Internet Gateway 202
Figure 125 Sample egress ACL Configuration for Internet Gateway 202
Figure 148 Example Configuration for LLQ & Explicit Policing 230
History
Table 1 Revision History
1.2 August 31st, Draft Added NAS/LAC and VHG tablular details to Dial Services
2006 section.
1.3 September 28, Draft Incorporated customer feedback for MPLS Core, updated
2006 IOS versions
1.4 October 1st, Draft Incorporated customer feedback for Dial Services,
2006 realigned Dial Service section topics to improve readability
1.5 October 2nd, Draft Modified SFP module number per 7613 Main PoP router as
2006 per BoQ
Document Purpose
The goal of this document is to provide Wateen Telecom with detailed design and configuration template
material explaining how different services like internet, MPLS VPN and Ethernet over MPLS can run on
their new ip core network. This network (Converged Multi-service Network as referenced by the customer)
consists of various sites in Pakistan. Wateen Telecom is expected to use WiMAX technology as last mile
access to the end customers. WiMax as well as overall network services design including various voice
related services as well as data services is led by Motorola. It is expected that this Low Level Design
Document will be used as a basis for implementation for this ip core network deployment across all sites of
Wateen Telecom in various cities in Pakistan. This document is part of a larger network design which deals
with Data Center architecture, Network Security services, Network Management solution, Public Wireless
LAN, Metro Ethernet and Dialup services. Other relevant sections of various technologies encompassing
this network and previously mentioned are discussed in separate low level design documents.
Intended Audience
Wateen Telecom Architecture & Engineering Team
Wateen Network Support & Operations Team
Motorola Engineering Team
Cisco Systems Wateen Telecom Account Team and Cisco Systems Advance Services Team
Scope
The scope of the project is to provide an IP/MPLS network to satisfy the requirements laid out within this
document. The document only pertains to a low level design, required configuration & implementation of
an ip/MPLS core network deployed from ground up.
Related Documents
Advanced Services Statement of work Wateen_X97743_SOW
Wateen CMSN Technical Proposal Release(x) <……..>
Finalized Bill of Quantity <……..>
Wateen Data Center and Security LLD Wateen_X97743_LLD_DCN.doc
Wateen NMS LLD Wateen_X97743_LLD_NMS.doc
Wateen Metro LLD Wateen_X97743_LLD_METRO.doc
Wateen PWLAN LLD Wateen_X97743_LLD_PWLAN.doc
References
Catalyst 6500 Series and Cisco 7600 Series Supervisor Engine 720 Compact Flash Install Note
http://www.cisco.com/univercd/cc/td/doc/product/lan/cat6000/cfgnotes/78_15537.htm
Release Notes for Cisco IOS Release 12.2SX on the Supervisor Engine 720, Supervisor Engine 32, and
Supervisor Engine 2
http://www.cisco.com/univercd/cc/td/doc/product/lan/cat6000/122sx/ol_4164.htm
Cisco 12010, Cisco 12410, and Cisco 12810 Router Installation and Configuration Guide
http://www.cisco.com/en/US/products/hw/routers/ps167/products_installation_and_configuration_guide_b
ook09186a00801fc679.html
Cisco XR 12000 and 12000 Series SPA Interface Processor-600 – Data Sheet
http://www.cisco.com/en/US/partner/products/hw/routers/ps167/products_data_sheet0900aecd8028085a.ht
ml
Cisco 5-Port, 8-Port, and 10-Port Gigabit Ethernet Shared Port Adapter – Data Sheet
http://www.cisco.com/en/US/partner/products/ps6267/products_data_sheet0900aecd8027cb99.html
Customer Wateen Telecom, a sister company of Warid Telecom has been established with the goal of
becoming a telecommunications provider in Pakistan Wateen Telecom is planning to install a major
network to achieve these objectives.
Warid Telecom (Sister Company) has a nation-wide cellular network launched in May of 2005 in 28 cities
and the major highways of Pakistan to provide high quality cellular services to Pakistan’s urban, suburban
and rural populations.
As part of the network buildout process, Wateen Telecom, would like to build an IP/MPLS core network
which can be used to offer several services to residential and business customers. This is a network being
built from scratch (green field) with last mile access technology in form of WiMAX which will be
provided by Cisco partner Motorola. The WiMAX infrastructure will consist of WiMAX radio network,
relevant CPE devices and Microwave devices (as a backhaul mechanism).
Wateen Telecom will use Cisco 12800 series routers and 7600 series routers to form the core IP/MPLS
network. The network will provide necessary infrastructure for offering services like internet, MPLS vpn
(L3vpn), Ethernet over MPLS (L2vpn). The converged multiservice network (CMSN) as referred by the
customer will also consist of:
• Metro rings. Islamabad, Karachi and Lahore are going to be covered by Ethernet metro rings. There
will be one metro Ethernet ring in Lahore, while two in Karachi and Islamabad. These rings will
provide above mentioned services while using Cisco 10720 devices and RPR (resilient packet ring)
technology.
• Data Center and Security Services: A data center will be built in Lahore which will provide data center
services like hosting services etc. The Data center will incoroporate security, content switching and
loadbalancing services for Wateen Telecom and its customers. Data Center will also host relevant
NMS elements as they pertain to Wateen Telecom infrastructure.
• PWLAN: Wateen Telecom intends to deploy public wireless LAN services in various hot spots, these
services will use IP/MPLS core as a transport mechanism.
• Voice Solution: Wateen Telecom along with Cisco partner Motorola intends to build a packetized
voice infrastructure to cater to business and residential voice requirements. IP/MPLS core provided by
Cisco systems will serve as the transport infrastructure for various voice solutions that Wateen
Telecom intends to deploy.
• Internet Peering Services: Wateen Telecom would like provide internet peering services to its
downstream customers and hence be a transit service provider.
• Content Caching & URL filtering Services: content caching services for an improved user experience
are also targeted at internet edge as well as URL filtering services.
While IP/MPLS and internet connectivity services are covered in this low level design document, please
refer to other low level design documents for other specific technologies and related low level design.
Network Layout:
Overall layout of the network as envisioned for connectivity and relevant transport technologies is given
below.
It is understood that infrastructure used to provide various services consists of 7609 and 7613 chassis using
Supervisor720 and Supervisor 32 as P/PE devices. Core of the network will be formed by four 12816
routers which will be located in four core sites namely Lahore, Karachi, Islamabad and Faisalabad. These
four devices will act as P nodes within in IP/MPLS core and will have no customer facing ports on them.
Of the four core sites, Lahore will also host a Data center as well as initial internet gateway link. The data
center will provide hosting services as well as various customer portals. Some infrastructure related
network management systems will also be located in this Data center including centralized DNS, DHCP
and radius servers. The core sites are referred to as Main PoP by Motorola, the two terms are used
interchangeably in rest of the document. All other sites connecting to core sites (Main PoP) are referred to
as remote sites. Customer refers to these sites as City PoPs. There are few remote sites which connect to a
remote site (City PoP) rather than core site (Main PoP), these sites are referred to satellite sites. Customer
refers these sites to as remote PoP.
Metro rings with Cisco 10720 and RPR technology will be deployed in Lahore, Karachi and Islamabad.
Lahore will be covered by a single metro Ethernet ring whereas Karachi and Islamabad will be covered by
two metro rings. Islamabad site will also provide coverage to its sister city Rawalpindi. Similarly
Sheikhupura will be covered by Lahore where microwave links will be used to backhaul traffic to nearest
Lahore PoP. For details on Metro rings please reference a separate design document.
One dial concentrator AS5400 will be deployed in Lahore, Karachi and Islamabad. All other sites will
deploy a single AS5350XM as dial concentrator. Dial concentrators will be deployed to provide dial up
internet access as well as dial access into MPLS VPN boxes.
Customer will also deploy Public Wireless Lan services using WiFi mesh. These services will be deployed
by the customer in core sites (Lahore, Karachi, Islamabad and Faislabad) at various hotspots.
The underlying transport for this IP/MPLS core is being built on SDH rings. These rings are being
deployed on a country wide basis which will ultimately enable required transports being called out in this
design.
A more detailed view of Lahore PoP, covering ip connectivity is shown below.
The diagram does not cover variouis services elements deployed in data center as well as at the internet
border router. These details can be referenced in other low level design documents. Specifically low level
design document covering data center and security elements.
Design Considerations
It is desired to build this network to provide ip connectivity to interconnect various sites of the customers.
This network should be able to carry voip traffic (customer intends to offer hosted voip solution) as well as
provide basic data connectivity in form of L3 vpn (MPLS vpn), L2 vpn (Ethernet over MPLS) and internet.
With MPLS vpn services, any to any model is desired for most customers however information on other
services model is also requested by the customer as forward looking option in form of design templates.
Key factors in design include the ability to provide all aforementioned services to various customers, have
a redundant and fault tolerant network with ability to react to various changes (link failure, node failure
etc.) in the network as quickly as possible.
Following two diagrams are included in this document for reference purpose only and are included on “as
is” basis to show physical transport and fiber connectivity. The diagrams are provided by the customer and
are meant to serve as a reference for physical topology perspective. IP/MPLS core topology is built on top
of the SDH network which is shown below.
Peshawar
Islamabad
Bannu
Kharian
Okara
Metro Fiber ring in 3 Cities
Quetta Multan
MNKA2084
Shikarpur
Larkana MDKA2103 MNKA2024
Kar-2Fiber
Hyderabad MSC-01
Karachi
MDKA2041
58
Ahmadi Banda is
Km
Replaced by Kark
Mach
98
Sp
Km
ur 1
01 15
Sibbi
mK
Dera Murad
50
Km
Jamali
Jacobabad Kohat
Bannu Peshawar
Kark 2nd pair 0 3
39
Larkana Shahbaz Mardan ur
Shikarpur Layyah Noshera Sp m
Km
Dadu Khel Abbotabad
Kandkot K
70 Km D.I.Khan 76
K m Kot Addu m 64 Km 60 Km 60 K
Kalari 126 66 K Rojhan
53 km 80 K m 65 K m 16 K
m 96 Taxila
m Km 71 K
Km 95 Km
m
32
95 K m Fazilpur 72 K Islamabad
Client #1 m
Jamshoro m
90
Bhakkar Rawalpindi
Km
Km
D.G. Khan 31
25
Km
82
Km
57
85
129 km Gujar Khan
57
Km 3
Mianwali 2F DWDM STM-64 (1 ?) 2nd pair
Km58 K
Reg
m
7
Qureshi 2nd pair Synchronization
Km
58
m
Standby Mode
Client #4 (Central-II)
Khushab Reg 23 Km
m
Km
m
m
Client #3 22 Km
29
49 Km
m
108 Km Faisalabad
4K
Thatta NMS Wagha 2F DWDM STM-64 (1 ?)
95
Khanewal
11
Standby
26 04
K
Km
m
Lodhran 83 Km
ur
43 Km 39 Pattoki Km
Sp
Km 67 58 Gujrat
Km
62 KmRaiwind Lahore
m
Hyderabad
11
K
24
70 K
m Mailsi Spencer
5
Sahiwal 26 K
m
Km
Km
Chichawatni
Km
43 48 km
20
83 Km
99
Km
6K m
Nawab Shah m 7K MS-SPring Lahore
10
32
33
109 K m Km Km Gujranwala
m 88 K 35 Thokar
5
101 Km Liaqatpur 43 Km Kasur
0
Vehari
ur
Amirabad
R.Y. Khan
m
Arifwala
Sp
Burewala
17 K
Sukkur Dherki
Client #2
LEGEND To Ganda Singh
NMS
Synchronization
Equipment
Active Mode
Wala
Active
Following describes configuration of a 12816. 12816 will be a P device in Wateen Telecom network. One
12816 will be deployed in each core network.
• 12816/1280-DC Cisco 12816 Internet Router, Sixteen-Slot System, 4 DC
• PRP-2 Performance Router Processor 2 (PRP-2)
• PRP-2/R Performance Router Processor 2(PRP-2) Redundant
• MEM-PRP2-1G*2 Cisco 12000 PRP-2 1Gig DRAM Default Option
• MEM-12KRP-FD128M*2 Cisco 12000 Series 128MB PCMCIA ATA Flash Disk
++
• 4OC12X/POS-I-SC-B*3 4 PORT OC-12 POS B
• 12000-SIP-600*2 10 Gbps IP Services Engine (modular)
• SPA-2XOC48POS/RPR*2 2-port OC48/STM16 POS/RPR Shared Port Adapters
• SPA-5X1GE*2 5-port Gigabit Ethernet Shared Port Adapter
• SFP-GE-S*10 1000BASE-SX SFP (DOM)
• SFP-OC48-IR1*4 OC-48c/STM-16c IR optics
Each core PoP will also have a single Dial concentrator. Lahore, Karachi and Islamabad will have dial
concentrator with the following configuration:
• AS54XM-16E1-480-D AS5400XM Data; 16E1, 492 DSPs, AC RPS, IP+ IOS
• AS54XM-DC-RPS AS5400XM DC Redundant Power Supply
• MEM-1024M-AS5XM AS5350XM and AS5400XM 1024MB Main SDRAM
• MEM-128CF-AS5XM AS5350XM and AS5400XM 128M Compact Flash
Faisalabad (and all other city PoPs) will have a single dial concentrator which will have the following
configuration:
• AS535XM-4E1-120-D AS5350XM Data; 4E1, 120 DSPs, Single AC, IP+ IOS
• AS535XM-DC-RPS AS5350XM DC Redundant Power Supply
• MEM-1024M-AS5XM AS5350XM and AS5400XM 1GB Main SDRAM
All four core PoPs have single Virtual Home Gateway (VHG) PE to allow dial in users to map to their
respective VRF. VHG has following configuration:
There will also be a data center in Lahore, which will consist of two 6513 along with various services
modules, 3750 access switches, IPS 4255 appliance sensors etc.
There are also two 7609 which will act as internet gateway peering to external service providers. These
7609 will have WS-SVC-ADM-1-K9 (Anomaly Detection Module) and WS-SVC-AGM-1-K9 (Anomaly
Guard Module). These routers will also connect with WAE-7326-K9 (Wide Area Application Engine
7326) and SCE2020-4XGBE-MM (Service Control Engine, Multi Mode, 4-port GE).
Data Center equipment and other devices pertaining to different technologies like security, network
management, content switching etc. are mentioned here for completeness sake, these are discussed in more
detail in other low level design documents.
City PoPs:
All city PoP sites (excluding remote city PoPs which do not connect directly to 12816 P device) have
following equipment for MPLS/VPN deployement:
City PoPs will also have a single dial concentrator which will have the following configuration:
• AS535XM-4E1-120-D AS5350XM Data; 4E1, 120 DSPs, Single AC, IP+ IOS
• AS535XM-DC-RPS AS5350XM DC Redundant Power Supply
• MEM-1024M-AS5XM AS5350XM and AS5400XM 1GB Main SDRAM
Four sites (namely Bahawalpur , Okara, R.Y. Khan and Gujrat) which do not connect directly to a 12816 P
device have slightly different configuration. These sites have 7609 chassis but have different supervisor
engine as well as different linecards.
Satellite (Remote) PoPs will also have a single dial concentrator which will have the following
configuration:
• AS535XM-4E1-120-D AS5350XM Data; 4E1, 120 DSPs, Single AC, IP+ IOS
• AS535XM-DC-RPS AS5350XM DC Redundant Power Supply
• MEM-1024M-AS5XM AS5350XM and AS5400XM 1GB Main SDRAM
Please note above is not comprehensive list of equipment but a summary of equipment pertaining to IP/
MPL core infrastructure. Kindly refer to latest version of BoQ for complete details of all the equipment for
this project.
• Modules A1 and B1 provide redundant power for system load zone 1 (the upper blower module and the
upper card cage).
• Modules A2 and B2 provide redundant power for system load zone 2 (the switch fabric card cage, the lower
card cage, and the lower blower module).
Caution A router configured for source DC operation must be operated with 4 DC-input PEMs installed at
all times for electromagnetic compatibility (EMC).
Note The lower card cage is an inverted, or head-down, copy of the upper card cage, which means that
cards are installed in an inverted or head-down orientation. The orientation of the slots is opposite that of
the upper card cage.
• Slot 8—The far left slot is reserved for an optional redundant RP.
Note This slot may be used for a line card if you are not using an redundant RP.
• Slots 9 through 15—Can be populated with any line cards supported by the router.
• Alarm—The far right slot is a dedicated slot for an alarm card.
Note 10-Gbps and 40-Gbps switch fabrics do not operate in 1/4-bandwidth mode as they did in some
earlier models of the Cisco 12000 series routers. You must have at least one CSC and three SFCs for the
system to function. You can add an additional CSC for redundancy.
The combination of CSCs and SFCs make up the 2.5-Gbps, 10-Gbps, or 40-Gbps per-slot switch fabric.
Routers are identified by the switch fabrics they use:
• Cisco 12016: 2.5-Gbps switch fabric
• Cisco 12416: 10-Gbps switch fabric
• Cisco 12816: 40-Gbps switch fabric
Each SFC or CSC provides a 2.5-Gbps, 10-Gbps, or 40-Gbps full-duplex connection to each line card in
the system. For example, in a Cisco 12416 router with 16 line cards, each with 2 x 10 Gbps capacity (full
duplex), the system switching bandwidth is 16x 20 Gbps = 320 Gbps.
Note The Cisco 12000 series router supports online insertion and removal (OIR), allowing you to remove
and replace a card while the router remains powered on.
Caution Do not remove the blank filler panel unless instructed to do so by a Cisco support representative.
PRP-2 Components
The PRP-2 contains the following components:
• PowerPC processor—Motorola PowerPC 7450 central processing unit (CPU). The CPU runs at an
external bus clock speed of 133 MHz and an internal clock speed of 667 MHz.
• SRAM—2 megabytes (MB) of static random-access memory (SRAM) for secondary CPU cache
memory functions. SRAM is not user configurable or field replaceable.
• NVRAM—2 MB of nonvolatile RAM (NVRAM). NVRAM is not user configurable or field
replaceable.
• Memory—Additional memory components include onboard Flash memory and up to two Flash disks.
• Sensors—Air-temperature sensors for environmental monitoring.
The PRP-2 contains the following additional components:
• SDRAM—Up to 4 GB of Cisco-approved synchronous dynamic random-access memory (SDRAM)
on two dual in-line memory modules (DIMMs). 1 GB of SDRAM is the default shipping
configuration. SDRAM is field replaceable only when using Cisco-approved DIMMs.
Note Software releases prior to 12.0(30)S do not recognize more than 2 GB of SDRAM and will only use
the first 2 GB of the installed memory. This does not affect the functioning of the PRP-2, but commands
such as show version will indicate that only 2 GB of SDRAM are installed.
• Hard disk drive—40-GB hard disk drive can be optionally installed on the PRP-2 board.
• CF—1-GB compact flash disk can be optionally installed on the PRP-2 board.
Note The PRP only supports +5.2 VDC flash memory devices. It does not support +3.3 VDC PCMCIA
devices.
Status LEDs (Slot-0/Slot-1) indicate when the flash memory card in that slot is accessed, each slot has an
eject button (located behind the cover) to remove a flash card from the slot.
Caution The reset switch is not a mechanism for resetting the PRP and reloading the Cisco IOS image. It is
intended for software development use only. To prevent system problems or loss of data, use the reset switch only
on the advice of Cisco service personnel.
Pressing the reset switch causes a nonmaskable interrupt (NMI) and places the PRP in ROM monitor
mode. When the PRP enters ROM monitor mode, its behavior depends on the setting of the PRP software
configuration register. For example, if the boot field of the software configuration register is set to:
• 0x0—The PRP remains at the ROM monitor prompt (rommon>) and waits for a user command to boot
the system manually.
• 0x1—The system automatically boots the first Cisco IOS image found in flash memory on the PRP.
The alphanumeric message displays show router status messages during the boot process, and after the
boot process is complete.
• During the boot process, the message displays are controlled directly by the MBus module.
• After the boot process, the message displays are controlled by Cisco IOS software (through the
MBus).
The alphanumeric message displays also provide information about different levels of system operation,
including the status of the GRP, router error messages, and user-defined status and error messages
Note A list of all system and error messages appears in the Cisco IOS System Error Messages publication.
FAN
STATUS
9 8 7 6 5 4 3 2 1
0
0
A/L
C/A
A/L
C/A
K
K
LIN
LIN
K
K
LIN
LIN
SPA-1XOC12-POS
SPA-1XOC12-POS
K
K
LIN
LIN
STA
STA
T
TUS
TUS
SE
SE
RE
RE
0
A/L
C/A
A/L
C/A
SPA-1XOC12-POS
SPA-1XOC12-POS
STA
STA
TUS
TUS
TEM TIV PW MT
TIV PW MT
MG
MG
R
R
E
E
AC
AC
EM
ST
SYS
SY
S
S
TU
TU
STA
STA
L L
AL AL
ST ST
IN UN IN UN
R R
OSR-7609
Above diagram represents a chassis where both STM-4 and STM-1 SPAs are populated in a each SIP-400.
In some instances, there is only single STM-4 SPA in each SIP-400. Please note that Slots 5 and 6 can
support primary and redundant Supervisor Engine 720s. A Supervisor Engine 720 with two GBIC uplink
ports and one 10/100/1000 Tx port. Only two of the three ports can be active at any one time.
City PoP:
Please see below for slot allocation for 7609:
* only used in Multan, Sahilwal, Sukkhur and Gujranwala to terminate satellite-remote PoPs.
Following is a representation of 7613 chassis populated with various cards and redundant supervisor
engines.
WS-X6724-SFP
24 PORT GIGABIT ETHERNET SFP
STATUS
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
WS-X6724-SFP
24 PORT GIGABIT ETHERNET SFP
STATUS
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24
WS-X6748-GE-TX 1 11 13 23 25 35 37 47
2 12 14 24 26 36 38 48
WS-X6748-GE-TX 1 11 13 23 25 35 37 47
2 12 14 24 26 36 38 48
WS-X6748-GE-TX 1 11 13 23 25 35 37 47
2 12 14 24 26 36 38 48
ALL A LL
ST ST
IN IN
N
N
U
U
R
For 7613 Series router, a Supervisor Engine 720 can fit in slot 7 and an optional redundant Supervisor
Engine 720 in slot 8. Each supervisor engine provides switching, local and remote management. The
Supervisor Enigne 720 has two GBIC uplink ports and one 10/100/1000 Tx port. Only two of the three
ports can be active at any one time.
13
Above figure shows fabric channel allocation for different 7600 Series Router. Please note that Sup32 is a
bus based supervisor only. In 13 slots chassis, following modules MUST be inserted in slot 9 to 13:
WS-X6748-GE-TX, WS-X6748.
(Slot and port numbering start at 1 (top/left) , 7609 with vertical line card ,NEB chassis, it starts from right)
Redundancy
The Cisco 7609 and 7613 series routers have these redundancy features:
• Ability to house two hot-swappable supervisor engines
• Ability to house two fully redundant, AC-input or DC-input, load-sharing power supplies
• Redundant hot-swappable fan assemblies containing multiple fans
Fan Assembly
The system fan assembly, located in the chassis, provides cooling air for the supervisor engine and
the switching modules
Following figure shows Cisco 7609 router show the direction of airflow into and out of the router. Sensors
on the supervisor engine monitor the internal air temperatures. If the air temperature exceeds a preset
threshold, the environmental monitor displays warning messages.
If an individual fan within the assembly fails, the FAN STATUS LED turns red.
Reset Button
The Reset button allows you to restart the system.
Supervisor Engine 32
3 SIP subslot 2
BGP AS Number
Wateen Telecom has applied to APNIC for publicly announceable AS number. In the interim a
private AS number from the range of 64512 to 65535 will be chosen. Customer has advised to use
65501 AS number until assignment of public AS number.
Naming Conventions
Customer has advised to use the following naming convention.
LHE-7600-PE1-GE0/0
Where LHE represents airport code of Lahore, 7600 is the platform name and PE01 is device role.
It has been advised to use airport code for cities where available to abbreviate a city’s name.
Devices included in this design and their hostnames are listed below.
Gatewy
6513 DC Lhe-6513-DC1
Switch
6513 DC Lhe-6513-DC2
Switch
2 Sialkot 7609 PE SKT Skt-7609-PE1
3 Gujranwala 7609 P/PE GUJ Guj-7609-PE1
4 Gujrat 7604 PE GJT Gjt-7609-PE1
IP Addressing Scheme
For infrastructure, publicly assigned ip addressing will be used. Customer has advised that APNIC
has allocated following public ip address space: 58.27.128.0 - 58.27.255.255 (/17). Please see
subsequent section for IP address allocation of core mpls infrastructure.
Cisco 7600 Series run 12.2S series software image specifically built for 7600/Catalyst 6500 series
platform. The current software image has various releases which are indicated by 12.2(18)SX. A
maintenance build of a particular release is noted as 12.2(18)SX. The latest release for 12.2(18)SX
at the time of writing this document is 12.2(18)SXF and is required for 7609 with Sup32 chassis.
Currently 12.2SXF has a maintenance build of 12.2(18)SXF5 and is recommended release for all
7609/7613 series chassis.
There are some features which are only available in a recently released train called 12.2(33)SRA
internally code named Cascades. While the release is available, it is suggested that either customer
should test this release extensively before deployment or wait for some field exposure of this
train.
On 12816, latest version on 12.0S train is 12.0(32)SY at the time of writing this document,
however it is suggested that customer deploy with 12.0(32)S4 as there are no features which are
required on customer network and only available in 12.0(32)SY.
http://www.cisco.com/en/US/products/hw/switches/ps708/prod_release_note09186a00801c8339.h
tml
For 12.0(32)S4
Software Feature Set: SERVICE PROVIDER/SECURED SHELL 3DES
Minimum Memory needed: 512
Minimum Flash Size 64
Date Released: 18-AUG-2006
Platform: 12000-PRP
The image can be downloaded here:
http://cco/cgi-bin/Software/Iosplanner/Planner-
tool/iosresult.cgi?get_crypto=&data_from=&hardware_name=12000-
PRP&software_name=SERVICE%20PROVIDER%2FSECURED%20SHELL%203DES&release
_name=12.0.32S3&majorRel=12.0&state=:HW:RL:SW&type=Early%20Deployment
IGP Routing:
After several discussions, customer has chosen OSPF as routing protocol of choice as IGP for the
core network. The decision is driven primarily by folks responsible for supporting the network
feel more comfortable with OSPF. Further details about OSPF are given under MPLS Core
Design section.
Route Reflectors:
The two PE routers i.e. 7613 in Lahore and Karachi will be configured as redundant route
reflectors for vpnv4 traffic in absence of dedicated route reflectors as an interim measure:
The second pair of P routers i.e.12816 in Lahore and Faisalabad will be configured as redundant
route reflectors for ipv4 traffic in absence of dedicated route reflectors as an interim measure.
Resilience
Resilience is achieved by using redundant route processors and redundant power supplies at the
device level. At the network level this is achieved by providing backup paths in case of a link
failure in the core network consisting of 12816 devices. On the edge, in core PoPs redundant
Gigabit Ethernet links are used to attach to 12816 devices. All remote PoP 7609 routers
connecting to 12816 have APS protection to ADM (add drop Mux). Same protection is used at
other side of the link, where 12816 link has APS protection to ADM. Similarly remote PoP 7609
connecting via STM-1 also have APS protection to ADM. While APS provides some protection
loca to the PoP, it would be much preferred to have redundant uplinks to core P (12816) device. It
is understood that due to underlying SDH ring and related optical transport constraint, it is not
possible to have redundant uplinks at this stage.
IGP will be used to reroute traffic in case of a link failure where a redundant link is available.
It is understood that in current Wateen Telecom network design, single P routers (12816s) are
single point of failure, e.g. in case if a 12816 router in Lahore goes down, it will isolate
connectivity to Lahore and all the infrastructure residing in Lahore. Same is true for all other core
(Main) PoPs. It is understood that this is a design choice Wateen Telecom made in early design
cycles, it is understood that possible impact of P routers going down is well understood by the
customer.
Scalability
Scalability of solution is restricted by various factors in the network. These include forwarding
capabilities of each device with various features turned on as required in our solution. It also has to do
with maximum number of interfaces and vlans that can be terminated on each element, number of
pseudowires that can be terminated, number of vpn routes that PE and Route Reflectors can carry as
well as maximum number of vpns that can be defined on a given platform, maximum number of
interfaces available on a platform, maximum memory supported etc.
This is something to be considered once network has achieved substantial growth. It is advised that as a
precautionary measure routes accepted in a vpn be limited to a specific number to help scale mpls vpn
service and avoid any potential misconfiguration issues. This is shown in the example below:
ip vrf cust1
rd 65501:10
route-target export 65501:10
route-target import 65501:10
maximum routes 500 warning-only #if number of routes exceeed 500 in this
vpn, issue a warning
It is understood that Wateen Telecom will use maximum 500 prefixes as initial number for limiting
number of prefixes received in a customer vpn.
There may also be some platform specific information that must be taken into consideration for overall
scalability as well. This information will vary from platform to platform. Following section describes
some factors to be taken into account for 7600 series routers.
7600 Restrictions
The Cisco 7609/7613 Series routers based on supervisor 720 are able to support 1024 MPLS L3 VPN
Virtual Routing and Forwarding (VRF) instances. The first 511 VRFs are supported with optimal
forwarding performance, the rest may experience some performance degredation due to recirculation in
the Cisco Supervisor Engine 720-3BXL.
Layer 3 LAN ports, WAN interfaces and subinterfaces, and some software features use internal
VLANs in the extended range. It is not possible to use an extended range VLAN that has been
allocated for internal use. Internal VLANs and user-configured VLANs share the 1006 to 4094 VLAN
spaces. A "first come, first served" policy is used in allocating these spaces.
By default internal vlan allocation is done in an ascending manner i.e.internal VLANs are allocated
from 1006 to 4094. A knob has been added to set the vlan internal allocation policy. The command
allows one to configure the allocation direction of the internal VLAN in descending order i.e internal
VLAN allocation is done from 4094 to 1006. The global config command is “vlan internal allocation
policy descending”. Internal vlan allocation can be verified by the command “show vlan internal
usage”.
For more details, please see: http://www.cisco.com/en/US/products/hw/routers/ps368/products_configuration_guide_chapter09186a0080160ed0.html
Wateen Telecom will use global command: “vlan internal allocation policy descending” to ensure
that internal vlan allocation is done in descending order. A system reload is required for change to take
effect. Vlan allocation as planned by the customer will leave a vlan range of 3500-4094 set aside for
internal usage initially.
These factors should be considered when designing the scaling factors for the maximum number of
VPNs supported in the system.
In certain cases, the PFC3BXL or PFC3B provides the capability to recirculate the packets.
Recirculation can be used to perform additional lookups in the ACL or QoS TCAMs, the Netflow
table, or the FIB TCAM table. Recirculation is necessary in these situations:
• To push more than three labels on imposition
• To pop more than two labels on disposition
• To pop an explicit null top label
• When the VPN Routing and Forwarding (VRF) number is more than 511
• For IP ACL on the egress interface (for nonaggregate (per-prefix) labels only)
Packet recirculation occurs only on a particular packet flow; other packet flows are not affected.The
rewrite of the packet occurs on the modules; the packets are then forwarded back to the PFC3BXL or
PFC3B for additional processing.
Please see below for further details:
http://www.cisco.com/univercd/cc/td/doc/product/core/cis7600/cfgnotes/optical/122sx/mpls.htm#wp1327282
It must also be noted that it is recommend to configure no more than 2,000 Layer 3 VLAN interfaces
on 7600 platform.
Please se below for further details: http://www.cisco.com/univercd/cc/td/doc/product/core/cis7600/software/122sx/swcg/layer3.htm
While it is understood that Wateen Telecom anticipates to have a mix of L3 and L2 vpn as well as
internet (ipv4 only) services on 7600 platform, above limits should be kept in mind when considering
overall capacity of the platform.
At the edge of the MPLS network there will be access switches (or RPR ring(s) populated with
10720 devices). Access switches will connect CE and perform a layer 2 function. In our case 7600
series routers will act as P/PE routers, providing customer facing trunk ports or access ports.
The PEs can classify traffic for different customers and apply the appropriate labels in a process
known as label imposition. The labels applied by the PE indicate the destination PE for each
packet, and also identify the VPN to which the device is connected or subscribes to (for example,
VPN1).
Two labels are applied to the IP packet, the first level, or top label defines the destination PE,
whilst the second level label defines the interface on the destination PE that the packet will be
forwarded to.
As the MPLS network is frame based (carries IP packets not ATM cells), the first level label and
second level labels are imposed as part of the frame that carries the IP packet. This is done by
inserting or stacking the labels between the layer 2 (link layer) and layer 3 (network layer)
headers. The following figure illustrates this.
Destination Customer
PE Destination
IP Packet
Frame
In terms of details of the MPLS header, it is located between the Layer 3 (IP) header and Layer 2
header. The EXP bits and the TTL field of the MPLS header can be copied from the IP header.
The S bit indicates whether there is more than one MPLS label in this frame.
The labelled frames are usually then forwarded to either a P device (12816 in the core) or a PE
(7613/7609) device depending on the routing metric and the shortest path as determined via the IP
routing table. In our case traffic will almost always pass through one of the core routers to reach
destination PE (depending on the traffic flow).
Once the labelled frames have left the PE routers, they are switched through the P network to the
destination PE. It is important to understand that P routers have no knowledge of layer 3 traffic. It
only looks at the first level label.
As the labelled frames are forwarded to the appropriate egress PE, each P router in the path will
examine the incoming label and replace it with an outgoing label to direct it to the appropriate
egress interface. This process is referred to as label swapping.
The last router in the path to the PE,, known as the penultimate hop (the hop before the last), will
remove the level 1 label. This is known as penultimate hop popping. Therefore when the frame
arrives at the destination PE, only the level 2 (or bottom of the stack) label remains.
The level 2 label is then examined to determine the egress interface on the PE, which will be the
link to the customer site. The label is removed from the frame and the original IP packet is then
forwarded to the destination customer network.
The underlying protocols to allow this process to operate: OSPF; Label Distribution Protocol; and
MP-BGP; are explained in further sections.
Network Design
The network design consists of a core consisting of point to point STM-16/STM-4 links between
12816 devices. These devices will be “P” devices and will not have any customer facing
links/ports. As described earlier these core sites are Lahore, Karachi, Islamabad and Faisalabad.
All four sites are cross connected to each other. Connectivity between Lahore, Karachi and
Islamabad consists of STM-16, while Faisalabad is connected via STM-4 link to all other devices.
It has been advised that connectivity to Faislabad had to be reduced to STM-4 (from STM-16, for
a symmetrical core) due to facilities reasons.
Gigabit Ethernet link exist between 7613 PEs in core sites and 12816 routers. Two 7613 PEs are
deployed in each core PoP. All remote PoPs connect with 12816 routers via STM-4 links. These
links are APS protected to respective ADMs. Some remote (satellite) sites connect via STM-1 link
to remote PoPs, these too are protected with ADM links. Only core sites have redundant PEs and
redundant links (albeit to same device) to 12816 P device.
All customer facing ports are Ethernet ports, these ports will connect to Microwave/WiMax
equipment which will provide last mile access to the customer via Ethernet. All customer facing
Ethernet ports on PEs are on LAN based cards (67xx or 65xx series cards).
There are also metro rings (layer 3 based RPR rings with 10720 as PE devices) which will be
deployed in three core sites, Lahore, Karachi and Islamabad. The interconnect between metro
rings and core city PEs (7613) is via point to point GE links. These GE links are cross connected
(i.e. between two 10720 and two 7613s). Further details can be found in separate low level design
which covers metro ring design.
Lahore site will also host a data center, the data center will have two 6513 chassis populated with
various services modules to provide content switching, security services and among others will
also host NMS systems for IP/MPLS infrastructure core.
Internet connectivity will be available in Lahore via two 7609 series routers. These routers will
connect via point to point GE links to data center switches as well as two 7613 core PE devices in
Lahore. Internet connectivity will be achieved by peering with different service providers and
learning full internet routing table via BGP. These routes will be carried in the core via IBGP and
offered to potential clients who are interested in multi homed internet connectivity to multiple
service providers. This service will establish Wateen Telecom as transit service provider in
addition to other services being offered to the customers.
Sites Connectivity
All sites are interconnected via either point to point Gigabit Ethernet links (core PEs) or STM-16
(between three core 12816s), STM-4 (remote PoPs and 12816 in Faislabad) and STM-1 (satellite
remote PoPs) links. .
58.27.191.209 -- 58.27.191.210
AS Number Notes
65501 For use in MP-iBGP peerings, Route Distinguishers, Route Targets
Backbone Protocols
This section deals with the protocols that the Label Switch Routers require to provide an MPLS
backbone and associated IP services, for example, Virtual Private Networks (VPNs). The
following routing protocols will operate in the MPLS backbone:
• OSPF will function as the Interior Gateway routing Protocol (IGP) providing IP connectivity
amongst all Label Switch Routers (P & PE).
• LDP (Label Distribution Protocol) is responsible for distributing label information for IP
destinations within the IGP routing table only. That is, P and PE addresses not part of a VPN,
such as link and loopback addresses.
• MP-BGP4 (Multi-protocol Border Gateway Protocol V4) is used to distribute VPN routing
information amongst the Edge LSRs (PE). It also distributes VPN-related labels to the PEs. It
will also carry internet ipv4 prefixes.
To provide MPLS VPN service, Wateen Telecom will need to implement MP-BGP between the
PE routers to exchange the VPN prefixes. An Interior Gateway Routing Protocol (IGP) must be
operating amongst the P and PEs before the label distribution (LDP) and VPN routing distribution
can execute.
OSPF is one of the most highly scalable and widely deployed link state protocol. OSPF has been
chosen by the customer as preferred IGP protocol primarily due to familiarity with the protocol in
comparison to ISIS. It will be used to build the internal routing table, which MPLS then uses to
build the Label Forwarding Information Base, LFIB table.
LDP assigns and distributes labels only for the IGP learnt routes. Therefore IGP must be running
for MPLS to be successfully enabled in the network.
Because the current roll out of Wateen Telecom’s IP/MPLS core network is not very large in
terms of the number of devices & links; a single area design is recommended. This allows for
simple management and operations support and allows for area definition depending on growth of
the network.
Area 1
Area 2 Area 3
Area 0
Core
Area 4 Area 5
With an OSPF design as shown above it is crucial to remember that any future network links may
only be within in area, or interconnecting an area to the backbone Area 0.
If a connection from Area 2 to Area 1 was required then either the area would need to be merged
into a single area; or a virtual link would be required. Virtual links are to be avoided if at all
possible.
For Wateen’s MPLS/IP network, a single area design will allow for maximum flexibility and will
scale for current set of requirements for IP/MPLS Core.*
As with other routing protocols, enabling OSPF requires that you create an OSPF routing instance
by creating a OSPF process, specify the range of IP addresses to be associated with the routing
process, and assign area IDs to be associated with that range of IP addresses. By default, OSPF
process ID is chosen as the OSPF domain ID. Since, any process number can be chosen. e.g. 1
router ospf 1
network <address> <wildcard-mask> area 0
passive-Interface loopback 0
log-adjacency-changes
Figure 23 OSPF Router process
There are multiple ways to define OSPF network statements to activate OSPF on an interface or
interfaces. By specifying a subnet to encompass multiple interfaces, keep in mind that other
interfaces will build neighbor adjacencies if the interface is active with a valid IP address
configured. This can be avoided by specifying individual host or subnets specific to an interface.
“Passive-Interface” disables the ability to create OSPF adjacencies over specified interfaces.
Clearly the loopback interface is not required to establish adjacencies.
The “log-adjacency-changes” command provides a high level view of adjacency state specifically
neighbors going up or down events. Include the detail keyword to cause syslog messages of all
state changes. This can assist in troubleshooting network issues in the future.
OSPF Router ID
Every router running OSPF within a network must have a unique router ID. The router ID is used
by the OSPF link-state database (LSDB) as a method of tracking each router within the AS and
the links associated with it. To assign the router ID, by default OSPF uses the highest IP address
of one of the router's active interfaces. If a loopback interface (loopback0) is configured with an
IP address, the Cisco IOS software will use this IP address as its router ID, even if other interfaces
have larger IP addresses. Since loopback interfaces are not affected by circuit failure, greater
stability in the routing table is achieved. Should there be other loopback addresses defined on the
router it is recommended to specifically set the router-id to loopback0 to remove any confusion as
to the ip address to used as the router id.
Cisco IOS provides an option of manually setting the OSPF router ID with the “ip ospf router-id”
router configuration command. We recommend setting the router-id manually as shown in Figure
24.
router ospf 1
router-id <address>
Figure 24 OSPF router-id
The OSPF metric is calculated as ref-bw divided by bandwidth, with ref-bw equal to 108 by
default, and bandwidth determined by the bandwidth command. The calculation gives FDDI a
metric of 1.
Granted that Wateen Telecom has multiple interface with high bandwidth (i.e. STM-16, one
gigabit Ethernet etc), one must use a larger number to differentiate the cost on those links vs other
links (if and when such links should be available, e.g. ten gigabit Ethernet or instances where Fast
Ethernet may be deemed sufficient).
If the reference bandwidth is left at its default, then the cost for a Fast Ethernet, Gigabit Ethernet
and ten gigabit Ethernet links would be 1, providing no differentiation to OSPF for the various
link speeds.
To overcome this, the reference bandwidth should be changed on all routers to the fastest link that
might exist in the network. Setting reference bandwidth to 10 Gigabit Ethernet should be
sufficient for foreseeable future, therefore the reference bandwidth should be changed using the
following command:
router ospf 1
auto-cost reference-bandwidth 10000
Figure 25 OSPF Reference Bandwidth
This will set all links with a cost relative to the 10 Gigabit Ethernet as shown in the following
table.
The value set by the ip ospf cost command overrides the cost resulting from the ospf auto-cost
command.
622 STM-4 16
1000 Gigabit Ethernet 10
2488 STM-16 4
10000 Ten Gigabit Ethernet 1
The reference bandwidth must be set on all routers participating in OSPF to provide consistent
cost allocation across the network. Also it should be assured that correct bandwidth of interface is
reflected for the physical interface.
set as point-to-point network type as long as there are only two routers on a single link as shown
in Figure 26.
OSPF Convergence
Resiliency and redundancy to circuit failure is provided by the convergence capabilities of OSPF
at layer 3. There are two components to OSPF routing convergence; detection of topology
changes and recalculation of routes.
Detection of topology changes are supported in two ways by OSPF. The first, and quickest, is a
failure or change of status on the physical interface, such as Loss of Carrier. The second is a
timeout of the OSPF hello timer. An OSPF neighbor is deemed to have failed if the time to wait
for a hello packet exceeds the dead timer, which defaults to four times the value of the hello timer.
On a Serial, Fast Ethernet or Gigabit Ethernet interface, the default hello timer is set to 10
seconds, therefore the dead timer is 40 seconds
Recalculation of routes is done by each router after a failure has been detected. A link-state
advertisement (LSA) is sent to all routers in the OSPF area to signal a change in topology. This
causes all routers to recalculate all of their routes using the Djikstra (SPF) algorithm. This is a
CPU intensive task, and a large network, with unreliable links, could cause a CPU overload.
A summary of OSPF protocol timers is given below. They are further discussed in detail in the
following section. It is suggested that in the initial deployment of the Wateen Telecom network,
initial timers should be left to default values as shown in table below.
These values can then be further adjusted and behavior of the network monitored depending on
results required. Section below discusses various options and suggested values that can be used
for fast convergence. These values can be tuned further based on the convergence requirements
and overall stability of the network. Some values have been suggested in subsequent sections.
Each of these timers will affect the performance of OSPF and can be tuned to effect both
convergence time and network resource utilization. While it is possible to use rather very small
values for hello and dead intervals (subseconds), these values need to be evaluated against overall
network stability and overall resource utilization (please see below for futher details) of the
network.
It is also worth noting that tuning hello and dead interval for the edge sites has no substantial
value as they are connected via POS links and are singly attached to the core. In case of quickly
reacting to the changes in network may mean quick reaction to bad news but no alternate path for
traffic.
For Wateen Telecom it is understood that core links between all 12816 P nodes (STM-16 and
STM-4) links carrier delay should be set to zero. Similarly carrier-delay should be set to zero for
all Gigabit Ethernet interfaces interconnecting 7613 PE devices to 12816 P nodes as well as 7613
PE devices to 10720 metro devices.
It is suggested that this value should be left unchanged for STM-4 and STM-1 links to remote PE
devices. This is primarily due to single point of connectivity for these devices. A quick reaction to
failed link will not lead to an alternate path. However for consistency reasons carrier delay may be
set to zero on all interfaces.
B. Auto Negotation
To deal with with unidirectional failures it is recommended that all Gigabit Ethernet interfaces be
configured with auto-negotiation, for all supported Gigabit Ethernet interfaces on all devices
within IP/MPLS core. The router detecting the LoS will signal this fact to the other router thereby
quickly bringing the link to the down state.
While it is possible to set hello timer value in msec in recent releases of IOS it is recommended
that the minimum value of Hello should be set to no less than 1 sec and therefore the minimum
value of RouterDeadInterval should be set to 4 sec. The following commands can be configured
on the link
!
interface GigabitEthernet1/2
ip address <ip address> <subnet mask>
ip ospf hello-interval 1
ip ospf dead-interval 4
POS interfaces provide for timely (sub 10ms) and reliable link failure management. Therefore the
tuning of OSPF hello timers is not necessary and should remain at their default settings.
For all the core links and core PEs it is suggested to rely on alternate mechanism like BFD
(discussed later) for dead peer detection as a worst case. BFD is very fast and allows for a failure
detection in 150 msec. Fore all remote PE devices (city and satellite PoP), it is suggested that BFD
should not be used (as all these devices have only single uplink). However should it be deemed
necessary, it is suggested to use above mentioned OSPF hello and dead interval values.
!
interface GigabitEthernet1/2
dampening
For Wateen Telecom, it is recommended that interface dampening be enabled on all interfaces
which have redundant uplink paths. This will include all core links on 12816 P devices as well
as Gigabit Ethernet links to core PEs, and interconnect of core 7613 PEs with 10720 metro
devices.
Interface dampening should not be enabled for remote PE links where only single connection
exists to the core network.
the SPF tree and is a function of the number of nodes and number of links in the overall network
topology.
Incremental SPF optimises the SPF re-computation by only considering part of the Shortest Path
Tree (SPT) that has changed rather than a Full SPF computation from scratch.
iSPF decreases the SPF run time and leads to a faster convergence. The more nodes and links are
there in the network, the better results of the iSPF computation in terms of noticeable
improvement.
!
router ospf 1
ispf
For Wateen Telecom it is recommended to enable incremental SPF on all routers participating in
OSPF backbone area.
LSA generation can be tuned in order to react quickly to a change and delay LSA generation in
the event of network instability. Following values are most aggressive in terms of
recommendation.
Router ospf 1
timers throttle lsa all 1 200 5000
We will generate a LSA immediately and exponentially delay by 200msec until reaching 5 sec
maximum value. Please note that MinLSArrival is by default 1 sec therefore it needs to be
changed on all routers otherwise any LSA generation before 1 sec will be ignored. The “timers
lsa arrival” command controls the minimum interval for accepting the same LSA. If an instance
of the same LSA arrives sooner than the interval that is set, the LSA is dropped. It is
recommended that the arrival interval be less than or equal to the hold-time interval of the timers
throttle lsa all command.
For Wateen Telecom it is recommended to leave the default values initially. Throttle and LSA
arrival timers can be tuned to above recommended vaules. For a network with lot of churn can
cause excessive CPU burden on the router, hence these values should be tuned keeping overall
stability of the network in mind. It is suggested that these values be kept consistent across the
entire OSPF backbone area and hence should be applied to all routers participating in backbone
OSPF area.
For fast convergence, it is possible to tune the SPF timers to increase the convergence time. Hold
time SPF is the time which controls the SPF in case of instability in the network (link flap)
therefore initial SPF time can be set to a low value.
SPT throttling allows SPF execution to be schedule in milliseconds intervals. At the same time it
can allow SPF execution to be delayed during network churn. SPF calculations occur at the
interval set by the “timers throttle spf” command.
SPF Start interval : Initial SPF schedule delay. By default initial SPF schedule delay is 5
seconds (5000 msec).
SPF Hold Time: (i.e. wait interval): The amount of time to wait until the next SPF calculation
occurs. Each wait interval after the “wait interval” calculation is twice as long as the previous one
until maximum wait time is reached. Default value for spf hold time is set to 10 seconds (10000
msec).
SPF maximum wait : maximum wait time: once the maximum wait time is reached, the wait
interval remains the same until the topology stablizes and no event is received in that interval.
Default value for maximum hold time is set to 10 seconds (10000 msec).
SPF computation can be tuned in order to react fast to a topology change and delay SPF
computation in the event of network instability. Following values are suggested:
router ospf 1
timers throttle spf 50 200 5000
The first computation will be delayed by 50msec in order to flood all changes to the neighbours
before running the SPF, next SPF will be delayed exponentially by 200 msec until reaching the
maximum value of 5 sec.
Please note in earlier versions of IOS use “timers spf” command to tune SPF values (allowed
two arguments, initial value and hold value). This command has been replaced by “timers
throttle spf” command.
For Wateen Telecom it is recommended to leave the default values initially. SPF throttling can be
tuned to above vaules which are most aggressive values suggested. For a network with lot of
churn can cause excessive CPU burden on the router, hence these values should be tuned keeping
overall stability of the network in mind. It is suggested that these values be kept consistent across
the entire OSPF backbone area and hence should be applied to all routers participating in the
backbone OSPF area.
Please note that any changes made should be configured across the entire backbone area not only
limited to 7600 and 12816 devices. i.e. any device (e.g. 10720) participating in backbone area
should have consistent configuration as noted above to achieve deterministic behaviour.
The fast detection of a failure provides immediate reaction in the event of a failed link or node,
whilst link down detection is normally used to determine this level of failure, BFD has the
additional advantage of being able to detect forwarding plane failures that may not be seen by the
control plane.
BFD packets use a UDP Datagram with a destination port of 3784 and a source port between
49252 and 65535, because this is designed as a rapid response protocol the output features are
bypassed for locally generated BFD control packets.
The main applications for BFD are to detect that the forwarding plane is operational
BFD uses two timers that are negotiated between the adjacent nodes, a transmit timer and a
receive timer the system that reports the slower times is used to determine the transmission rate.
As can be seen above, the router on the left has TX/RX rates of 100/50ms and the remote site has
TX/RX Rates of 40/60ms.
The left router will pick the slowest out of his transmit rate and the other side receive rate, and the
same vice versa.
The right node here is transmitting at 50ms (even though the desired transmit rate was set to
40ms), this is because the desired receive rate on the adjacent node was set for 50ms.
BFD has two main modes of operation: Asynchronous and Echo mode:
In Asynchronous mode BFD sends control packets in each direction informing the adjacent node
that it is still alive.
These control packets are sent on a periodic basis, if a number of packets (default 3) are not
received by the adjacent node then it assumes there is a link / forwarding plane issues and the
session is declared down. This is the only mode available today
In Echo mode BFD sends packets to it’s own IP address, the packets are sent to the adjacent node
and looped back.
As before control packets are sent on a periodic basis and a number of lost echo packets will cause
the session to be declared down, echo mode also has the benefit of testing the remote forwarding
plane as well.
In both modes BFD will immediately notify routing process that interface is down and thus
‘shortcut’ the legacy, ‘hello/holdtime’ timers.
interface GigabitEthernet1/1/0
ip address <ip address> <subnetmask>
ip ospf bfd disable !selectively disable bfd on an interface
interface GigabitEthernet1/1/1
ip address <ip address> <subnetmask>
bfd interval 50 min_rx 50 multiplier 3
!
router ospf 1
bfd all-interfaces
!
!on PE devices where there are many interfaces and bfd is not required on
most it is prefferable to enable BFD on a per interface basis
!
interface GigabitEthernet1/1/1
ip address <ip address> <subnetmask>
no ip directed-broadcast
ip ospf bfd
negotiation auto
bfd interval 50 min_rx 50 multiplier 3
Following table shows the interfaces for each device where BFD is recommended to be enabled in
Wateen Telecom’s network.
As mentioned earlier, BFD can be enabled on all interfaces for consistency reasons, in case of
stable network this will not pose any problems. However as precaustionary measure it is
suggested to follow above table as a guideline.
OSPF Security
It is recommended to authenticate the OSPF packets such that routers can participate in routing
domains based on predefined passwords. By default, a router uses a Null authentication, which
means that routing exchanges over a network are not authenticated. Two other authentication
methods exist: Simple password authentication and Message Digest authentication (md5).
Authentication does not need to be set, but it is strongly recommended for security purposes. And
it is recommended that MD5 be used as the authentication method since it provides higher
security than the plain text authentication method.
Message Digest Authentication is a cryptographic authentication. A key (password) and key-id are
configured on each router. The router uses an algorithm based on the OSPF packet, the key, and
the key-id to generate a “message digest” that gets appended to the packet. Unlike the simple
authentication, the key is not exchanged over the wire. A non-decreasing sequence number is also
included in each OSPF packet to protect against replay attacks.
This authentication not only provides for protection against Denial of Service attacks; but it also
assists with prevention of downtime by preventing routes advertised from a misconfigured node
entering the routing table of core devices.
It is possible to configure a unique password for every subnet in the network; however this may be
difficult to administer so a single password can be used for all interfaces in the core network.
router ospf 1
area 0 authentication message-digest
Figure 38 Configuring OSPF Authentication
OSPF Configuration
The following OSPF configuration is an example for a PE/P router in the Wateen Telecom’s
network. The same configuration can be used for the other PE routers – just the router-id needs to
be changed along with identifying correct network interfaces. SPF timers are shown as default,
these may be tuned to suggested values documented earlier.
interface Loopback0
ip address <ip-address> 255.255.255.255
no ip directed-broadcast
interface Ethernet0/0
bandwidth 1000000
ip address <ip-address> 255.255.255.252
no ip directed-broadcast
ip ospf message-digest-key 1 md5 7 110A1016141D
!
router ospf 1
router-id <loopback-ip-address>
log-adjacency-changes
auto-cost reference-bandwidth 10000
bfd all-interfaces
area 0 authentication message-digest
network <network address> <wild-card mask> area 0
network <network address> <wild-card mask> area 0
network <network address> <wild-card mask> area 0
network <network address> <wild-card mask> area 0
network <network address> <wild-card mask> area 0
Cisco Express Forwarding (CEF) is advanced Layer 3 IP switching technology. CEF optimizes
network performance and scalability for networks with large and dynamic traffic patterns by
essentially distilling the routing information into a forwarding database known as the FIB,
Forwarding Information Database. Cisco Express Forwarding or CEF switching is a pre-requisite
for MPLS to function properly. Therefore CEF must be configured on all the PE and P devices in
the Wateen Telecom’s network.
This is default switching mode for all Cisco 7600 Series routers. It is possible that due to memory
or some other software related issues, CEF may get disabled. In that scenario, CEF can manually
be enabled as shown in Figure 40.
ip cef [distributed]
* show ip cef and show cef line can be used to verify if CEF is operational
cef forwarding should not be turned off even if a platform allows for turning it off.
bytes. Cisco routers with PoS interfaces can receive and send appropriate protection signals to
connecting ADMs. The Cisco PoS protection scheme can be set up for situations where protect
and working interfaces are different ports on the same router or even on the same line card in the
same router. These scenarios, however, provide protection for only router interface or link failure.
APS refers to the mechanism of using a "protect" POS interface in the SONET network as the
backup for "working" POS interface. When the working interface fails, the protect interface
quickly assumes its traffic load. On the protect circuit, the K1 and K2 bytes from the line
overhead (LOH) of the SONET frame indicate the current status of the APS connection and
convey any requests for action. This signalling channel is used by the two ends of the connection
to maintain synchronization.
The working and protect circuits themselves, within the router or routers in which they terminate,
are synchronized over an independent communication channel, not involving the working and
protect circuits. This independent channel may be a different SONET connection, or a lower-
bandwidth connection. In a router configured for APS, the configuration for the protect interface
includes the IP address of the router (normally its loopback address) that has the working
interface.
Cisco routers also use a proprietary protocol, known as the Protect Group Protocol, between
working and protect routers to complement SONET/SDH protection signaling that occur with
ADMs. The Protect Group Protocol is IP-based and uses User Datagram Protocol (UDP) transport
(UDP port 172). Using this protocol, the process controlling the protect circuit directs the process
containing the working circuit whether to activate or deactivate the working circuit, in the case of
degradation or loss of channel signal, or manual intervention. If communication between the two
processes is lost, the working router assumes full control of the working circuit as if no protect
circuit existed.
1 + 1 linear APS and MSP implementation requires the ADMs to forward the same data signal to
both protect and working circuits. In the 1 + 1 protection scheme, the working PoS interface
selects the data signal when there are no failures or severe errors. Signal selection is at a protected
interface when there are problems on the working circuit.
There are two modes of operation, unidirectional and bidirectional. In bidirectional mode, the
receive and transmit channels are switched as a pair. In unidirectional mode, the transmit and
receive channels are switched independently. For example, in bidirectional mode, if the receive
channel on the working interface has a loss of channel signal, both the receive and transmit
channels are switched.
Cisco 12000 and 7600 sereies routers do not provide true unidirectional mode of operation, where
the transmit interfaces are bridged and the receive interface is chosen independently. Instead, Tx
and Rx are tied to the same interface (the Active one). The router uses alarms like LAIS to force
ADM switchover to the appropriate interface. It is a hack for ADMs which do not provide
bidirectional support. It is recommended to configure the ADM in bidirectional mode.
The W (working) and P (protect) interfaces receive the same payload from the ADM - but only
one is selected, or currently active. Only the selected interface actually processes the payload. The
deselected interface is held in a "line protocol is down" state and cannot participate in routes or
adjacencies. That is, the currently-deselected interface is completely removed from the layer 3
picture. A protection switch forces the APS routers to remove adjacencies and routes involving
the now-deselected interface, and form new adjacencies over the now-selected interface. In other
words, IP traffic begins to flow on the new W interface only after routing-protocol convergence.
Thus, although the APS switch itself requires less than 50 ms to complete, as required, all this
means is that the choice of which interface is to be selected is changed, which affects at most two
routers (W and P). Full restoration of IP traffic via the newly-selected interface requires that new
adjacencies be formed between the newly-selected interface and the remote router, and that the
resulting routes be disseminated to all the routers directly connected to either W or P.
Reflector mode provides improved routing convergence. Reflector mode enhances routing
convergence because the remote router now has early notice of a switch between W to P, and can
tear down its now-outdated adjacency with the now-deselected system, and need not wait for a
timeout. The convergence enhancement does not depend on whether the aps reflector command is
configured. The W(working), P(protect), and R(remote) routers must support reflector mode
requirements. Reflector mode requires an interface capable of reflector mode at the remote end of
the SONET path .
Open Shortest Path First (OSPF) supports APS reflector mode via DDTS CSCdr57673.
A value of "(null)" in the Remote APS configuration field of the show controller pos command
indicates that the local end has not received reflector channel information from the remote PTE. If
the remote PTE supports the reflector channel capability, a problem probably exists between the
remote PTE and remote ADM.
As noted above APS protection is always configured between a router and ADM, in Wateen
Telecom’s deployment both links (working and protect) terminate on same router.It is
recommended that these links be placed on separate linecards. Remote PEs (7609) connect to
their respective core P (12816) device, connecting via POS. These links are expected to be APS
protected to their respective ADM device on both ends. Following diagram represent the
connectivity between different routers and ADMs.
interface Loopback0
ip address 100.1.1.1 255.255.255.255
no ip directed-broadcast
!
interface POS1/0
ip address 10.1.1.1 255.255.255.0
no ip directed-broadcast
no keepalive
crc 32
pos framing sdh
pos report rdool
pos report lais
pos report lrdi
pos report pais
pos report prdi
pos report sd-ber
pos flag s1s0 2
aps group 10
aps working 1
pos ais-shut
!
interface POS1/1
ip address 10.1.1.3 255.255.255.0
no ip directed-broadcast
carrier-delay ms 0
no keepalive
crc 32
no keepalive
pos framing sdh
pos report rdool
pos report lais
pos report lrdi
pos report pais
pos report prdi
pos report sd-ber
pos flag s1s0 2
aps group 10 !allows multiple protect/working int on one Router
aps protect 1 100.1.1.1
pos ais-shut
* Permitted group numbers are 0-255. A sequential group number approach is suggested.
* aps revert 1 command allows for revertive aps i.e. it enables automatic switchover (1 minute in
this case) from the protect interface to the working interface after the working interface becomes
available. It is recommended that revertive APS should not be used as for every n failures one
gets 2n switches and therefore 2n times to reconverge.
It is also possible to use same ip address space for both working and protect interfaces. Using the
same IP address obviously conserves IP address space. Using different IP addresses provides with
a bit more granularity while testing - i.e by pinging different IP addresses one can confirm which
link (working or protect) is currently active. Alternatively this info can also be obtained via "sh
aps pos x/y" command as well. It is suggested that different IP addresses be used for ease of
troubleshooting unless insufficient IP addresses are available.
The pos report commands as shown above are optional and allow reporting of POS alarms to the
console.
The pos ais-shut is an important command which triggers an Alarm Indication Signal (AIS) if the
interface is placed in administrative shutdown state.
While the configuration on a 7609 PE device would look like the following:
interface Loopback0
ip address 200.1.1.1 255.255.255.255
!
interface POS3/0
ip address 10.1.1.2 255.255.255.0
no ip directed-broadcast
no keepalive
crc 32
pos framing sdh
pos report rdool
pos report lais
pos report lrdi
pos report pais
pos report prdi
pos report sd-ber
pos flag s1s0 2
aps group 10
aps working 1
pos ais-shut
!
interface POS3/1
ip address 10.1.1.4 255.255.255.0
no ip directed-broadcast
no keepalive
crc 32
pos framing sdh
pos report rdool
pos report lais
pos report lrdi
pos report pais
pos report prdi
pos report sd-ber
aps group 10
aps protect 1 200.1.1.1
pos ais-shut
!
The pos delay triggers commands “pos delay triggers line” and “pos delay triggers path”
provide the option to delay IOS to set the line protocol to down when a POS alarm triggers the
line protocol to go down.
It is recommended not to use Line triggers on interfaces that participate in APS, because Line
triggers interfere with APS operation. The “pos delay triggers line n” command does not allow
the line to go down on the brief LOS that comes from internally protected DWDM gear, from the
time an internal DWDM protection switch occurs. If the defect clears during the holdoff period, it
is like the defect never occurred.
all IP destination prefixes will be either a loopback or network interface address and normally
there should be no customer IP addresses in the global routing table.
The core P routers only have an understanding of labels that are associated with IP destinations in
the OSPF internal routing table. They have no knowledge of labels related to routes in customer
VPNs as these are created and distributed by the MP-BGP process between PEs only. Therefore,
in the P network, a labeled IP packet is switched to the next-hop based only on the outer label (the
one allocated by LDP from the global routing table) until it finally reaches its destination. In an
MPLS-VPN network, this final destination will always be the PE that originated the VPN route.
The IOS CLI accepts “tag” and “mpls” command interchangeably for most cases. For example,
“show tag tdp neighbor” and “show mpls ldp neighbor” produce identical output. In some
cases, there is only an “mpls” command such as “mpls label protocol ldp”. Wateen Telecom is
advised to use only the newer “mpls” form of the commands.
Prior to activating Label Switching on a router Cisco Express Forwarding must be enabled (on
7600 series router this is default switching mode), plus the mpls ip command must be issued on
each IP-enabled interface of P and PEs that are connected. mpls ip command should not be
enabled on PE-CE connections or on PE interfaces where hosts will reside in a VRF.
As with OSPF, MD5 based authentication should be enabled where LDP will be used to prevent
any DoS attacks, and to help avoid configuration errors.
To support label switching over an interface where the MTU size is default (1500bytes), the MTU
size for MPLS frames must be increased with the following command on every interface in the
core on P and PE routers.
The command shown in Figure 44 can be configured on all gigabit Ethernet interfaces in the
Wateen Telecom’s network. The default MTU on STM-1/STM-4/STM-16 interfaces is 4470
bytes and does not need to be changed.
Upto six labels have been allowed in MPLS to cater for other services on the network such as
traffic engineering and Carrier Supporting Carrier. Each service may require an increase in the
label stack from 2 to something greater. Given that there may be requirements to roll out traffic
engineering and other services, it may be desireable to set mpls mtu to a higher value to
accommodate maximum number of labels.
LDP Autoconfiguration
LDP autoconfiguration enables one to globally enable LDP on every interface associated with an
IGP instance. The IGP supported for LDP autoconfiguration are OSPF and ISIS. Support for
OSPF has been there in IOS 12.0S train since 12.0(30)S release.
Under current software releases, this feature can only be used for devices which are targeted to
run 12.0S release i.e. 12816 routers. This feature is currently not available for 7600 series
platform. Its support is slated for a subsequent release (i.e. a release scheduled to ship after
12.2(33)SRA).
Once OSPF configuration is complete, LDP autoconfiguration can be invoked under “router
OSPF x” process with following configuration:
mpls ip
mpls label protocol ldp
router ospf process-id
network ip-address wildcard-mask area area-id
mpls ldp autoconfig [area area-id]
Figure 45 LDP Autoconfiguation on P router
Once mpls ldp autoconfig command is configured under a routing process, all the interfaces that
belong to an OSPF area are enabled for LDP.
If it is desired to remove LDP from an interface, command: “no mpls ldp igp autoconfig” can be
used on the interface to disable LDP from the interface. It is suggested to use LDP
autoconfiguration on 12816 series P routers for convenience and consistency.
MPLS LDP MD5 Global configuration allows one to enable LDP MD5 globally instead of on a
per-peer basis. Using this feature one can set up password requirements for a set of LDP
neighbors to help prevent unauthorized peers from establishing LDP sessions and to block
spoofed TCP messages. Currently this feature is only available for our P devices on the latest
12.0(32)SY release. The feature is slated for an upcoming release for 7600 series platform. While
this feature is not available on our current target release for GSR platform, this may be of interest
once Wateen Telecom has decided to upgrade to this version of the code. For further details,
please see:
http://www.cisco.com/en/US/products/ps6566/products_feature_guide09186a00805f24da.html
This feature is currently available on 12.0S train starting 12.0(30) and hence available on
12000 series platform. This feature has recently been incoporated in IOS train available on
7600 series routers. This feature is available in 12.2(33)SRA release on 7600 series platform.
MPLS LDP-IGP Synchronization:
http://www.cisco.com/univercd/cc/td/doc/product/software/ios120/120newft/120limit/120s/12
0s30/fsldpsyn.htm
This feature is also available on 12.0S train starting 12.0(30)S. However this feature is not
available on 7600 series platform and is scheduled for a subsequent release (i.e. a release after
12.2(33)SRA).
Please note that as indicated above, these features can be used in future depending on various
factors including additional redundancy and growth of the network.
With filtering LDP labels for PE loopbacks only, following additional commands will need to be
added:
The “mpls ip” command merely enables LDP on an interface, and it does not control whether a
packet is switched using the MPLS Label at all – this is determined by the Ethertype on an
incoming frame and via various forwarding tables.
update X update X
Step1: Customer routes are injected into the VRF table at the Site1 PE using redistribute
connected
Step 2: At the Site1 PE, the routes in the customer VRF are then exported into MP-BGP,
updates contains RD for unicity of IPv4 + one or more RT for VRF identification.
Step 3: These exported routes are then sent across the MPLS backbone between the BGP peers.
This process would be carried out for any other BGP peers that had members in the
same VPN. Note that this step shows a logical connection between the two BGP peers
(actually MP-BGP update is sent to a Route-Reflector then RR reflect to other MP-
BGP peers, RR are explained further)
Step 4: The routes are imported into the correct VRF at the other end in respective PE which
also have same VRF defined, (RT configured in VRF define in which VRF the routes
should be imported).
Step 5: These routes are then accessible for users from VPN at respective site.
RR
Without RR RR
With RR
Figure 49 BGP Route Reflectors
As shown above all PE routers will peer with two VPNv4 route reflectors. The interim route
reflectors are two PE devices (7613) in Lahore and Karachi. These two routers will peer with each
other as well as peer with all PE devices, treating them as vpnv4 route reflector clients.
It should also be noted that two vpnv4 route reflector clients are also PE devices and hence will
carry full internet BGP routing table. Therefore vpnv4 route reflectors will also peer with two
ipv4 route reflectors. This is not a recommended solution, instead an interim measure. This
necessity can be eliminated in case of dedicated route reflectors.
Please refer to Interet gateway section for ipv4 route reflector description.
!
router bgp 65501
neighbour <ip address / peer group> password <password>
!
Figure 51 BGP Authentication
If the possibilities for a prefix were received in this order without deterministic MED then the
node would install path 2 over path 1 due to a lower IGP metric, it would then prefer path 3 to
path 2 due to the fact that it is an external route.
However, with deterministic MED this removes any order dependencies of MED based decisions,
it ensures that the best MED comparison is made across all routes from the same AS, in this case
path 1 would be chosen due to it having the lowest MED.
BGP Convergence
Several BGP timers can be used to tune the iBGP convergence in an IP MPLS / VPN Network.
• Advertisement-interval: To set the interval between the sending of BGP routing updates. The
default interval is 5 seconds for iBGP peers.
• Keepalive: Frequency, in seconds, with which the Cisco IOS software sends keepalive
messages to its peer. The default is 60 seconds.
• Holdtime: Interval, in seconds, after not receiving a keepalive message that the software
declares a peer dead. The default is 180 seconds.
• Scan-time: it defines the frequency at which the router validates routes in BGP table. It
validates the locally originated network and next-hop within the BGP table. Valid values used
for selecting the desired scanning interval are from 5 to 60 seconds. The default scan time is
60 seconds. (Per address-family use “bgp scan-time {5-60}”)
• Scan-time import: this timer is relevant for VPNv4 address-family. It defines the frequency at
which routes received from another PE through MP-BGP are imported into a VRF. This timer
is not relevant for next-hop deletion that is processed based on the “normal” BGP scan-time.
Default value is 15 seconds. (Per address-family “bgp scan-time import {5-60}”)
The following table summarizes the default values for the iBGP timers.
Table 8 BGP Default Timers
• iBGP convergence follows IGP convergence, This means that when a PE router or its
uplinks fail, BGP next-hop reachability will be lost after the IGP converges. The BGP
scanning process will detect the loss of the next-hop for all relevant BGP and MP-BPG
routes and stop selecting it in the BGP RIB.
• It is recommended initially to leave the timers at their default.
• In current release on 7600, BFD is not available for BGP. BFD will be really relevant for
eBGP sessions
As discussed in the IGP routing section, fast IGP convergence and related values that should be
tuned have already been provided under the event of an internal link or node failure.
This means that when a PE router or it’s uplinks fail, BGP next-hop reachability will be lost after
the IGP converges. The BGP scanning process will detect the loss of the next-hop for all relevant
BGP and MP-BPG routes and stop selecting it in the BGP RIB.
Table below shows potentially tuned BGP tuned times, these should be modified with care and
only after ensuring stability of the network infrastructure
BGP monitors the next hop of installed routes to verify next-hop reachability and to select, install,
and validate the BGP best path. By default, the BGP scanner is used to poll the RIB for this
information every 60 seconds. During the 60 second time period between scan cycles, Interior
Gateway Protocol (IGP) instability or other network failures can cause black holes and routing
loops to temporarily form.
BGP Support for Next-Hop Address Tracking
BGP monitors the next hop of installed routes to verify next-hop reachability and to select, install,
and validate the BGP best path. By default, the BGP scanner is used to poll the RIB for this
information every 60 seconds. During the 60 second time period between scan cycles, Interior
Gateway Protocol (IGP) instability or other network failures can cause black holes and routing
loops to temporarily form.
The BGP Support for Next-Hop Address Tracking feature is enabled by default when a supporting
Cisco IOS software image is installed. BGP next-hop address tracking is event driven. BGP
prefixes are automatically tracked as peering sessions are established. Next-hop changes are
rapidly reported to the BGP routing process as they are updated in the RIB. This optimization
improves overall BGP convergence by reducing the response time to next-hop changes for routes
installed in the RIB. When a bestpath calculation is run in between BGP scanner cycles, only
next-hop changes are tracked and processed.
BGP Next hop address tracking feature is available on 12816 series router staring 12.0(30)S
(actual release of the feature is in 12.0(29)S). However this feature only showed up in latest
release of 7600 i.e. 12.2(33)SRA. It is suggested that default scan-time value be used once 7600
series routers have been updgraded to 12.2(33)SRA.
ip tcp path-mtu-discovery
Every TCP session has a limit in terms of how much data it can transport in a single packet. This
limit is defined as the Maximum Segment Size (MSS) and is 536 bytes by default. This means
TCP will take all of the data in a transmit queue and break it up into 536 byte chunks before
passing packets down to the IP layer. Using a MSS of 536 bytes ensures that the packet will not
be fragmented before it gets to its destination because most links have a MTU of at least 1500
bytes. The following command will allow you to check the MSS of all BGP peers:
The problem is that using such a small MSS value creates a large amount of TCP/IP overhead,
especially when TCP has a lot of data to transport like it does with BGP. The solution is to
dynamically determine how large the MSS value can be without creating packets that will need to
be fragmented. This is accomplished by enabling "ip tcp path-mtu-discovery" (PMTU). PMTU
allows TCP to determine the smallest MTU size among all links between the ends of a TCP
session. TCP will then use this MTU value minus room for the IP and TCP headers, as the MSS
for the session. If a TCP session only traverses Ethernet segments then the MSS will be 1460
bytes. If it only traverses POS segments then the MSS will be 4430 bytes. The increase in MSS
from 536 to 1460 or 4430 bytes reduces TCP/IP overhead, which helps BGP converge faster.
Once we have enabled this knob and reset all of our BGP sessions via a reboot or "clear ip bgp *"
we can see the expected increase in MSS:
Similarly there will be two route reflectors for ipv4 traffic to carry all ipv4 prefixes learnt via
EBGP. The route reflectors will be second set of two core P devices 12816 series routers located
in Lahore and Faisalabad.
Recommendation of RR design for Wateen Telecom’s network:
• The RRs will peer with each other forming a full mesh.
• All PEs will peer with both RRs for vpnv4 and with second set of RR for ipv4 traffic.
• The AS number used between BGP peers will be 65501.
• The two Pes chosen as vpnv4 RRs will only reflect VPNv4 routes, there are separate RRs for
IPv4 routes. It is always recommended to have dedicated route refelectors, separate route
reflector for ipv4 and vpnv4 are recommended.
• Normally RRs are not required to enable MPLS, as they are not a transit point for any traffic,
therefore MPLS need not be configured on them. All routers will peer, using the BGP
neighbor statement, to the loopback address of the adjacency. The loopback address is used to
make sure the IP address of an interface stays up and is independent of an interface that might
be unstable.
• Passwords with MD5 authentication will be used on all BGP peers.
• Members of an external peer group have different autonomous system (AS) numbers.
Following is a configuration example of route reflector.
!
address-family ipv4
neighbor refip activate
neighbor refvpn activate
neighbor <RR1-ipv4 address> peer-group refip
neighbor <RR2-ipv4 address> peer-group refip
no auto-summary
no synchronization
exit-address-family
!
address-family vpnv4
neighbor refvpn activate
neighbor refip activate
neighbor refvpn send-community extended
neighbor <RR1-vpnv4 address> peer-group refvpn
neighbor <RR2-vpnv4 address> peer-group refvpn
exit-address-family
!
address-family ipv4 vrf voip
redistribute connected
no auto-summary
no synchronization
exit-address-family
!
The following parameters are suggested to be enabled on all BGP speaking routers.
• Use the description field to add the hostname of the neighbour in the configuration
The following BGP features should be disabled (if not disabled by default):
• Auto-summarization
• IGP Synchronization
The SSO feature takes advantage of available RP redundancy with in the dual-processor
environments (7600, 12000 and 10000 series routers etc). The protocol establishes one of the RPs
as the active processor while the other RP is designated as the standby processor, synchronizing
critical state information between the two devices occurs automatically and ensures that upon an
active failure the standby is initialised and ready to take over operations.
Following an initial synchronization between the two processors, SSO dynamically maintains RP
state information between them.
A switchover from the active to the standby processor occurs when the active RP fails, is removed
from the networking device, or is manually taken down for maintenance. Specifically following
conditions will cause a switchover:
• Hardware failure on the active supervisor engine
• Clock synchronization failure between supervisor engines
• Manual switchover
Without having a SSO enabled solution when a switchover of a RP occurs the state of the layer 2
control is lost.
With SSO when a switchover occurs the standby RP maintains the session state for a number of
protocols: Frame Relay, ATM, PPP, MLPPP, HDLC. With these sessions maintained there is no
protocol bounce, no interface flap so downstream traffic can continue to be forwarded as if there
is no interruption
In contrast to SSO, other operating mode that are supported on redundant supervisor engine in
7600 series chassis are RPR and RPR+. RPR is slowest form of redundancy supported, in this
mode, a switchover can take upto 2 or more minutes. In RPR mode, redundant supervisor engine
come out of reset mode but are not operational.
In RPR+ mode, supervisor engine switch over time can be 30 seconds or more. In this mode the
redundant supervisor engine is fully initialized and configured.
RPR+ Restrictions:
These guidelines and restrictions apply to RPR+:
• Network services are disrupted until the redundant supervisor engine takes over and the
router recovers.
• The Forwarding Information Base (FIB) tables are cleared on a switchover. As a result,
routed traffic is interrupted until route tables reconverge.
• Static IP routes are maintained across a switchover because they are configured from entries
in the configuration file.
• Information about dynamic states maintained on the active supervisor engine is not
synchronized to the redundant supervisor engine and is lost on switchover.
SSO provides a faster switchover relative to HSA, RPR, and RPR+ by fully initializing and fully
configuring the standby RP, and by synchronizing state information, which can reduce the time
required for routing protocols to converge due to the uptime gained by not re-initialising linecards
and interfaces
For 7600 series platform, please note that during switchover, system control and routing protocol
execution is transferred from the active supervisor engine to the redundant supervisor engine. The
switch requires between 0 and 3 seconds to switchover from the active to the redundant supervisor
engine.
Redundancy Restrictions:
Configuration changes made through SNMP are not synchronized to the redundant supervisor
engine. A work around is that after configuring the router through SNMP, the running-config file
should be copied to the startup-config file on the active supervisor engine to trigger
synchronization of the startup-config file on the redundant supervisor engine and with RPR+,
reload the redundant supervisor engine and MSFC.
Supervisor engine switchover takes place after the failed supervisor engine completes a core
dump. A core dump can take up to 15 minutes. A core dump configuration should only be used
when troubleshooting a specific problem. Under normal operations, core dump shouldn’t be
taken.
* With Sup720 and Release 12.2(18)SXF and later releases, if a fabric synchronization error
occurs, the default behavior of the router is to switchover to the redundant supervisor engine. In
some cases, a switchover to the redundant supervisor engine may be deemed more disruptive than
powering down the module that caused the fabric synchronization error. The “no fabric error-
recovery fabric-switchover “command should be used to disable the switchover and power down
the module with the fabric synchronization error. The command should be used in case a problem
is detected and confirmation is made that problem is due to fabric synchornization error.
Hardware Restrictions:
• Cisco IOS running on the supervisor engine and the MSFC supports redundant configurations
where the supervisor engines and MSFC routers are identical. If they are not identical, one
will boot first and become active and hold the other supervisor engine and MSFC in a reset
condition.
• Each supervisor engine must have the resources to run the router on its own, which means all
supervisor engine resources are duplicated, including all Flash devices.
• Separate console connections are required to each supervisor engine.
• Both supervisor engines must have the same system image
Redundancy Configuration:
!For SSO configuration
redundancy
mode sso
!
!For RPR+ configuration
redundancy
mode rpr-plus
It is recommended to use SSO mode should be deployed on all P, PE and Data Centre devices,
this will help maintain uptime in the eventuality of a RP switchover.
Non-Stop Forwarding in IOS allows for the forwarding of data packets to continue along known
routes while the routing protocol information is being restored following a failover. With NSF,
peer networking devices do not experience routing flaps. During failover, data traffic is forwarded
through line cards while the standby Route Processor (RP) assumes control from the failed RP.
When an OSPF NSF-capable router performs an RP failover, it must perform two tasks to
resynchronize its link-state database with its OSPF neighbours.
First, it must relearn the available OSPF neighbours on the network without causing a reset of the
neighbour relationship.
Second, it must reacquire the contents of the link-state database for the network.
After an RP failover, the NSF-capable router sends an OSPF NSF signal to neighbouring NSF-
aware devices. This signal is in the form of a link-local LSA generated by the failed-over router.
Neighbour networking devices recognize this signal as a cue that the neighbour relationship with
this router should not be reset. As the NSF-capable router receives signals from other routers on
the network, it can begin to rebuild its neighbour list.
After neighbour relationships are re-established, the NSF-capable router begins to resynchronize
its database with all of its NSF-aware neighbours. At this point, the routing information is
exchanged between the OSPF neighbours. After this exchange is completed, the NSF-capable
device uses the routing information to remove stale routes, update the RIB, and update the FIB
with the new forwarding information. OSPF on the router as well as the OSPF neighbours are now
fully converged.
Devices which have a single processor (e.g. 10720, 7200, 3800 series router etc) will not
experience RP switchover in the event of a fault; however, these devices can be NSF aware,
meaning that they run NSF software and can maintain session information with a peer device (a
Cisco 7600, 10000, or 12000 series router) following a switchover of the peer device.
BGP, OSPF, and IS-IS all have been enhanced with NSF-capability and awareness, which means
that routers running these protocols can detect a switchover and take the necessary actions to
continue forwarding network traffic and to recover route information from the peer devices. NSF
is a routing “knob”. Therefore it is configured under the routing portion of a routers configuration
router ospf 1
log-adjacency-changes
nsf
network 10.0.0.0 0.0.0.255 area 0
!
!
Figure 56 OSPF NSF configuration
It is recommended that Wateen Telecom deploy NSF as most of the infrastructure is singly
attached and some protocols may benefit from NSF in case of RP switchover. Please note that in
case of 7600 BFD is not sso aware and hence will cause routing protocol adjacencies to be reset
during RP switchover as documented below.
OSPF NSF requires that all neighbor networking devices be NSF-aware. If an NSF-capable router
discovers that it has non-NSF -aware neighbors on a particular network segment, it will disable
NSF capabilities for that segment. Other network segments composed entirely of NSF-capable or
NSF-aware routers will continue to provide NSF capabilities
7600 series platform has following restriction when deploying NFS & SSO:
• For NSF operation, SSO configuration is required on the device
The section details the MPLS VPN design to be used by Wateen Telecom. Some overview of
MPLS VPN will be given while detailing the sample configuration to be used by Wateen Telecom
for its initial deployment phase.
Definition of a VPN
A VPN is a set of systems, which are allowed to communicate with each other. Such a system is
defined by a set of administrative policies. The path between two systems in a VPN, and the
characteristics of that path may also be determined (wholly or partially) by policy. Whether a
system in a particular VPN is allowed to communicate with systems that are not in the same VPN
is also a matter of policy.
In MPLS, a VPN generally consists of a set of sites, but it is also possible to apply different
policies to different systems that are located at the same site. A given set of systems may be in
one or more VPNs. A VPN can consist of sites or systems which are all from the same enterprise
"intranet,” or from different enterprises "extranet;” it may consist of sites or systems which all
attach to the same MPLS backbone, or to different service provider backbones.
The mechanisms for stating the policies are very general, and do not require that two sites in the
same VPN be associated with a common Route Distinguisher or a common Route-Target
extended community attribute.
to customer’s end including access switch. It is understood that these design details may be
covered by Cisco partner Motorola in a separate document.
VRF Name
The VRF name is simply a unique name used to identify the VRF and the routes it contains. It is
suggested the name always be lower case to avoid any potential compatibility issues with some
NMS platforms that may only work with lower case names.
Based on the requirements put forth by the customer and information provided thus far, following
vpns will be created at the rollout phase. It is understood that nms vpn will be created at each PE,
therefore this vpn can also be considered as a test vpn to aid with troubleshooting and ongoing
maintenance and support of the IP/MPLS network.
Based on information provided by the customer so far, voip vpn may be split int several vpns. For
further details please refer to access section of this document.
Route-Distinguisher
Each VRF must have a unique RD (Route Distinguisher) which will have the following format:
<AS>:<Unique Number>
Where:
<AS> is the 16-bit autonomous system number allocated to the network. This will be set to
the value 65501 and will also be used for all MP-BGP configurations
<Unique Number> A unique 32 bit number within the AS that is allocated by Wateen Telecom.
(Decimal maximum of 4294967295)
The situation where multiple sites of the same customer are connecting to the same PE router,
each sub-interface should use the same VRF definition as they would normally be part of the same
routing policy. This would give the customer sites peer access to each other via the PE.
Route-Targets
Route targets define which routes will be imported into a particular VRF and how the routes
exported into the VPN-IPv4 address space, which MP-BGP operates on, will be tagged.
The import and export values for route-targets can match the RD value for the VRF although as
mentioned earlier they are unrelated values. The RD uniquely identifies customer IPv4 routes and
the route-targets define import export policies for routes into and out of the VRF. Using the same
route-target and RD values simplifies the configuration and management of it.
The default route-target for Wateen Telecom VRFs will be equal to the RD. Additional route-
targets may be required given the route import & export policies needed.
Once the name, route descriptor & route-targets are assigned for a VRF it can be defined from the
command line in global configuration mode:
ip vrf nms
rd 65501:11
route-target import 65501:11
route-target export 65501:11
Route-Maps
There may be occasion where just using route-targets may not provide sufficient granularity for
importing or exporting routes. Finer route control can be achieved by using route-maps which can
import or export routes based on many different characteristics. This can be accomplished by
using import map <route-map name> or export map <route-map name>
To identify a route-map for a particular VRF table, a recommended approach would be to precede
it with IN_ for import route-maps and OUT_ for export route maps. For example,
IN_Preferred_Gateway.
An example of vrf route-map is given below:
ip vrf svc
rd 65501:903
import map in-svc-imp
!
route-map in-svc-imp permit 10
match ip address 101
access-list 101 permit <ip|tcp|udp> network wildcard mask wildcard
Interface vlan10
description NMS VLAN 10
ip vrf forwarding nms
ip address 172.24.100.1 255.255.255.0
For Wateen Telecom each access vlan will be mapped to either a subinterface or SVI (switched
virtual interface) which will represent a routed interface for VLAN. This will in turn be assigned
to a vrf. Based on the information provided so far, each customer site will be mapped to a specific
vlan, this in turn would mean a subinterface or an svi can then be mapped to a specific vpn.
In the case, where ip telephony servers are connected to a vpn, it may make sense to configure
these as an svi, rather than configuring subinterfaces. Unless a specific requiremen exists, it is
generally suggested to use svi interface rather than suninterfaces.
When an interface is assigned a VRF, any existing IP address assignments on that interface are
automatically deleted and must be re-entered.
Please note that an interface can only belong to a single VRF.
• Intranet – Full Mesh VPN – How to provide access between different sites of the same
customer/service
It is understood that initial phase of deployment Wateen Telecom will be deploying intranet – full
mesh vpn form of services, this pertains to nms as well as voip vpn. Other forms of services are
mentioned as forward looking options.
The following sections will discuss these services.
By this configuration command, traceroute on host devices attached to PE will show other PE
devices to be just one hop away. In reality an IP packet may transit more than one core node,
though the host will not see this. It is also possible to use modified version of above mentioned
command where “forwarded” key word is added. With this option, while hosts see everything
one hop away, local troubleshooting is still possible on the PE router and all intermediate hops are
shown for traffic generated by the PE router.
All the sites of a VPN can communicate directly with each other in this model. All sites export
and import routes using the same RT.
Figure below shows a VPN with three sites. All the sites can communicate with each other. All
the sites use the same RT (65501:1) to export the routes and the same RT (65501:1) to import the
routes.
VRF RED
RD 65501:1 MPLS Network
RT export 65501:1
RT import 65501:1
PE1
RED
RED PE2 CE2
CE1 PE3
10.0.0.0/24 11.0.0.0/24
CE3
12.0.0.0/24
RED
ip vrf test
rd 65501:1
route-target export 65501:1
route-target import 65501:1
!
router bgp 65501
bgp router-id <ip-address>
no bgp default ipv4-unicast
<standard iBGP neighbour configuration>
!
address-family ipv4 vrf nms
redistribute <connected / static / rip>
no auto-summary
no synchronization
exit-address-family
Same configuration can be repeated on all PEs and any to any connectivity will be achieved for all
devices with in a single vpn . Please note in this case only a single best path will be forwarded by
the route reflector. Please see PE-CE routing protocol section for further details on routing
protocol between PE-CE link.
VRF RED
Rd 65501:1
RT exp 65501:1 - VPN RED
VP
VPN RT imp 65501:1 - VPN RED
RE
RED
CE1
10.0.0.0/24
10.0.0.0/24
MPLS Network CE2 VPN
VPN
RE
RED
RE
VP PE3
GREE CE1
GREEN
CE2 VPN
12.0.0.0/24
VRF BO TH
GREEN
GREE
Rd 65501:3
RT exp 65501:1 13.0.0.0/24
RT exp 65501:2
RT imp 65501:1
RT imp 65501:2 VRF GREEN
CE3 Rd 65501:2
Site in both RT exp 65501:2 - VPN GREEN
GREEN and RT imp 65501:2 - VPN GREEN
RED VPNs
14.0.0.0/24
Obviously, the IP addresses that are being used in the sites that belong to overlapping VPNs need
to be unique in both VPNs.
The way to implement an overlapping VPNs infrastructure is to:
• Have each VPN use its own Import / Export RT that the sites that participate in that VPN will
export and import.
• Have the sites that participate in more than one VPN import routes with route targets from any
VPN in which they participate and export routes with route targets for all the VPNs in which
they participate.
• Use 1 VRF per VPN and per overlapping VPN instance (per PE).
• Use 1 RD per VPN and per overlapping VPN instance.
• Use 1 Import / Export RT per VPN. The sites that belong to multiple VPNs will use multiple
Import / Export RTs.
! VPN-PE1
!
ip vrf RED
rd 65501:1
route-target export 65501:1
route-target import 65501:1
ip vrf GREEN
rd 65501:2
route-target export 65501:2
route-target import 65501:2
!
router bgp 65501
< BGP configuration goes here >
VPN-PE2
!
ip vrf RED
rd 65501:1
route-target export 65501:1
route-target import 65501:1
ip vrf GREEN
rd 65501:2
route-target export 65501:2
route-target import 65501:2
!
router bgp 65501
< BGP configuration goes here >
VPN-PE3
!
ip vrf BOTH
rd 65501:3
route-target export 65501:2
route-target export 65501:1
route-target import 65501:2
route-target import 65501:1
Obviously, this model assumes that client (spoke) IP addresses will not overlap, as this is a single
VPN where this should not be the case.
The design rules for a Central Services Hub and Spoke VPN scenario are:
VRF SPOKE
Rd 65501:2
RT Exp 65501:2 MPLS Network
RT Imp 65501:1
VPN-PE1
Regional
CE 1
VPN-PE2 CE 2 Site2
10.0.0.0/24
10.0.0.0/24 VPN-PE3
Regional 11.0.0.0/24
VRF HUB VRF SPOKE
Site1
Rd 65501:1 Rd 65501:2
RT exp 65501:1 RT exp 65501:2
RT imp 65501:2 RT imp 65501:1
CE 3
12.0.0.0/24
Central
Site
VPN-PE1
!
ip vrf SPOKE1
rd 65501:5
route-target export 65501:5
route-target import 65501:6
!
router bgp 65501
< BGP configuration goes here >
VPN-PE2
!
ip vrf SPOKE2
rd 65501:8
route-target export 65501:8
route-target import 65501:6
!
router bgp 65501
< BGP configuration goes here >
VPN-PE3
!
ip vrf HUB
rd 65501:6
route-target export 65501:6
route-target import 65501:5
route-target import 65501:8
!
router bgp 65501
< BGP configuration goes here >
This sample configuration will not allow spoke1 to talk to spoke 2 even if they are on the same
PE, this is because the routes are imported into a different VRF, both VRF’s will receive copies of
the Hub routing table as they are importing the Hub routes (65501:6)
This VPN model is used for VPNs that have a Central Site and Regional Sites and the Central Site
controls the connectivity between the Regional Sites. For e.g. the Central Site might have a
Firewall and all the traffic between the Regional Sites must pass through the Firewall.
Figure below shows a VPN with one central site and two regional sites. Communication between
the two regional sites is through the central site. The central site is connected to MPLS-VPN
Network with two different VRFs in order to implement this solution. VRF HUB-IMPORT is
used to import all the routes from the Spoke sites and the second VRF HUB-EXPORT is used to
export all the routes (including the routes from the Central Site) to the Spoke sites.
VRF SPOKE1
RD 65536:1
RT export 65536:1
MPLS Network Regional
RT import 65536:2 VPN-PE2 CE2 Site2
VPN-PE1
11.0.0.0/24
CE1 VPN-PE3 VRF SPOKE2
10.0.0.0/24
RD 65536:2
RT export 65536:1
Regional RT import 65536:2
Site1 VRF HUB-IMPORT VRF HUB-EXPORT
RD 65536:3
RD 65536:4
RT import 65536:1 RT export 65536:2
CE3 CE4
Firewall
12.0.0.0/24
Central Site
Figure 68 - Hub and Spoke Model with Connectivity between Spoke Sites
All the Spoke sites will be assigned unique Route Distinguishers. This scheme will help to avoid
the problem of importing the PE-CE links in some of the spoke sites.
Sample configuration for the PE’s in this scenario is shown below:
This is a Sample configuration for hub-and-spoke with two VRF on the hub site.
VPN-PE1
ip vrf SPOKE
rd 65536:5
export map SPOKE
route-target export 65536:5
route-target import 65536:6
VPN-PE2
ip vrf SPOKE
rd 65536:8
export map SPOKE
route-target export 65536:5
route-target import 65536:6
VPN-PE3
ip vrf HUB-EXPORT
rd 65536:7
export map HUB-EXPORT
route-target export 65536:6
!
ip vrf HUB-IMPORT
rd 65536:6
export map HUB-IMPORT
route-target import 65536:5
!
router bgp 65501
< BGP configuration goes here >
Please note that Hub-PE site must have two separate interfaces connecting to the hub site.
common management VRF can reach all customer VRFs; but they cannot reach each other. The
management workstations will originate from this VRF and the VRF will have visibility of each
VRF to be managed. Conversely, each customer VRF must contain the address of the
management workstation to allow two way communications between the Management
workstation and the CE router.
The figure below illustrates the importing and exporting of routes from the management VRF.
ip vrf vpn_customer2
rd 65501:10
route-target export 65501:10
route-target import 65501:10
route-target import 65501:4 # Import routes from the Management VRF
ip vrf vpn_customer2
rd 65501:10
route-target export 65501:10
route-target import 65501:10
route-target import 65501:4 # Import routes from the Management VRF
This configuration will import all routes from the management VRF into the customer VRF,
though with no control over exactly which routes are imported.
Similarly on the PE connecting to the Management VRF the configuration would be:
ip vrf management
rd 65501:4
route-target export 65501:4
route-target import 65501:4
route-target import 65501:10 !Import routes from the Customer VRF1
route-target import 65501:20 !Import routes from the Customer VRF2
ip vrf management
rd 65501:4
route-target export 65501:4
route-target import 65501:4
route-target import 65501:10 !Import routes from the Customer VRF1
route-target import 65501:20 !Import routes from the Customer VRF2
Again this configuration would mean all management hosts can access all hosts in the customer
VPN above.
ip vrf vpn_customer2
rd 65501:10600
route-target export 65501:10600
route-target import65501:10600
export map OUT_Extranet_Cust2 ! Export the Extranet subnet
route-target import 65501:107501 ! Customer 2 Extranet In
!
ip prefix-list Customer2_Prefix seq 10 permit <prefix/mask>
!
route-map OUT_ Extranet_Cust2 permit 10
match ip address prefix-list Customer2_Prefix
set extcommunity rt 65501:106501
ip vrf management
rd 65501:10
route-target export 65501:10
route-target import 65501:10
export map OUT_Extranet_Mgmt ! Export the Extranet subnet
route-target import 65501:106501 ! Customer 2 Extranet In
!
ip prefix-list Mgmt_Prefix seq 10 permit <prefix/mask>
!
route-map OUT_Extranet_Mgmt permit 10
match ip address prefix-list Mgmt_Prefix
set extcommunity rt 65501:107501
!
Routing Protocols
A routing protocol may be used between CE and the PE if the CE is a router and dynamic routing
information exchange between PE and CE is required.
A VPN Routing and Forwarding table (VRF) is associated with each CE interface on the PE and
contains the routing information associated with that site. A PE/CE routing protocol is necessary
so that the PE table can be populated with the customer’s routes.
The following routing protocols can operate between the CE and PE for an MPLS network;
• Connected
• Static
• RIPv2
• eBGP
• OSPF
• EIGRP
The above listed dynamic routing protocols have been modified to understand VRF tables by the
use of a feature called address families. Address families define the VRF contexts that the routing
protocol will operate in.
Note that the routing protocol that operates between the PE/CE is independent of any IGP that
may be operating in the VPN network. These routes may be redistributed into the PE/CE routing
protocol to populate the VRF. Therefore an edge device could be running EIGRP in the private
LAN, and redistributing to RIPv2 running between the PE/CE to populate the VRF.
It is important to understand that no special MPLS configurations are needed at the Customer
Edge devices. Only standard IOS routing commands are required. The Customer Edge is unaware
of the MPLS cloud it is connecting to and only requires the configuration of the appropriate
PE/CE routing protocol and redistribution commands into the IGP, if necessary.
Based on Wateen Telecom’s current requirements, it is understood that in most cases for small
customers only static routing will suffice. While for medium sized customers RIP and for large
enterprises BGP may be used as PE-CE routing protocol.
Therefor routing PE/CE routing protocols that may be used in Wateen Telecom’s environment can
be Connected, Static, RIPv2, and eBGP. As discussed while it is possible to offer all routing
protocols as potential PE-CE routing protocols, standardizing on specific protocols allows
customer to gain expertise and build an appropriate operations support model for these routing
protocol. Hence only connected, static, rip and eBGP will be discussed in the subsequent sections.
It is also suggested that in order to keep the changes in configuration of a VPN-PE to the
minimum, static routes may be used on the VPN-PE to propagate the VPN Routes where dynamic
routing information provides no additional value (e.g. single point of exit) or if a site has less than
3 subnets to advertise.
“Connected” Routing
“Connected” routing is not really a routing protocol as such, but it merely implies that the PE
router will inject the subnets of active interfaces directly connected to the VPN into the VRF
routing table.
The PE router can be set to redistribute the connected interface subnet addresses into the VRF test
for example; the routes should then be redistributed into MP-BGP to other PEs where vrf test is
defined.
!
router bgp 65501
address-family ipv4 vrf test
redistribute connected
!
Static Routing
Static routing is normally used when a VPN site is small and the IP addressing of devices is
unlikely to change. The CE router would have a default route pointing in the direction of the
MPLS cloud, whilst the PE would require a similar static route inserted into the appropriate VRF
table associated with the CE remote interface.
For example, on a PE device following static route is added for vrf test :
Figure 77 Configuring static routes redistribution into vrf context under BGP
Where 192.168.1.0/24 is a prefix on CE device attached to a PE. By adding this static route and
redistributing static routes in vrf configuration, other sites in test vpn can reach this network.
It is also possible to have vrf routes made visible into global routing table. This process for
sharing routes between a VPN routing table and the global routing table is known as “route-
leaking”.
configuration example for route-leaking is given below:
router ospf 1
redistribute static subnets
!
router ospf 2 vrf management
redistribute static subnets
!
ip route <network > <mask> interface NH ip address !interface for vrf, NH
(next hop address) is
needed on multipoint
interface
!
ip route vrf test <network> <subnet mask> next-hop address global
It is highly recommended that route-leaking should not be used until and unless absolutely
required.
RIP v2
RIP v2 can be used as the routing protocol on the CEs that do not support BGP or where the the
number of prefix are small.
Passwords can be used to authenticate RIP v2 sessions between VPN-PE and CE. The password
will be unique for each VPN.
Sample configuration is given below:
router rip
!
address-family ipv4 vrf cust1
network 172.16.0.0
no auto-summary
version 2
exit-address-family
For small sites (sites that do not require the complete routing table of the VPN) and stub sites
(sites that have only one network behind the CE router), RIP updates from the VPN-PE to the CE
can be suppressed. (passive interface on the VPN-PE sub-interface) Default route pointing to the
VPN-PE should be added to the routing table on the CE.
eBGP
Private AS Numbers (64512 to 65535.) can be used for the BGP session towards the CEs where
the CE does not have a globally assigned AS number.
The same AS Number can be used for all the sites of a single VPN to conserve the number of AS
Numbers this also allows for VPNs that have more than 1024 sites. AS-override will be used on
the VPN-PEs in order to reuse the same AS Number for all the sites.
In order to prevent loops, a BGP extended community Site-Of-Origin (soo) should be used to
identify each site. soo has the same format as the RD or the RT (X:Y) and uses 64 bits. Each site
must be assigned a unique soo.
The following design rules will apply to the eBGP session between the VPN-PE and CE:
1. Private AS numbers will be used for the BGP session on the CE Router
2. as-override will be used to conserve the private AS numbers.
3. Each Customer site belonging to a Central ServiceVPN will be assigned a unique SOO
attribute.
4. soo will be used to detect and prevent loops for multi-homed sites.
5. All the eBGP timers will be left to their default value in this phase of the project.
6. Passwords will be used to authenticate eBGP session between VPN-PE and CE. The
password will be unique for each VPN.
7. Peer-Groups will be used on the VPN-PE routes for eBGP sessions with CE Routers that
belong to the same VPN. This will help to reduce the size of the configuration and also
help in trouble-shooting.
The following IOS commands shows the configuration on a PE router that uses BGP as the
Routing Protocol with the CE. This configuration also uses as-override and soo. The soo in this
example is set to 65536:4
On the 7600 PEs, separate SVI interfaces (mapped to different customer/services) will be
configured for each vpn. The vpn can be L2vpn or L3vpn. Alternatively SVI can belong to global
routing table (internet access).
Assuming only several VLANs for MPLS VPN and global routing, multiple interfaces will need
to be configured.
VLAN Allocation
Different VLAN ranges can be reserved for different services on 7600 PE. Preliminary
assignment as provided by the customer is given below.
Customer has been advised to allocate a range of vlans set aside for internal use, coupled with
system configuration to assure internal vlans are allocated in descending order. It is also advised
that customer consider overall system scalability guidelines including a limit of 2K svis and
recommended maximum number of 1K vrfs along others.
Customer has devised an allocation scheme for NMS (management) VLAN and voip vlans. This
allocation is separately documented by the customer in a document called “VLAN-VPN Map
140806.pdf” and should be referenced independently.
The vlan allocation scheme uses non-overlapping vlans for regional PoPs.
Customer has also requested following vpns to be configured everywhere for voip network and
management vpn. Suggested vpn names are given below.
It is understood that SBC (session border controllers) are dual homed systems which will
participate in both sbc and ims vpns.
Interaction between these vpns is described by the customer is given below:
• IMS+AS VPN should be able to communicate to Class 4 VPN
• Management VPN should should be able to communicate to Class 4 VPN
In the absence of any further details, it may make sense to collapse three vpns into a single vpn. It
is expected that futher details will be provided by the customer. Connectivity requirements for all
vpns are any to any connectivity.
Following IP addressing scheme is understood for WiMAX AP and subscriber Module within
nms vrf context:
Islamabad 10.136.0.0-10.143.255.255
Faislabad 10.144.0.0-10.151.255.255
Karachi 10.152.0.0-10.159.255.255
It is understood that these subnets will be assigned in nms vrf to all access infrastructure
elements.
Similary allocations for other VPNs are provided by the customer on regional basis:
It is understood that customer will use same ip addresses in global routing table as well as in sbc
access vpn. This requirement is stated by the customer and per the customer driven by the
requirement of customer edge device.
IP addressing scheme for these vpns is documented by Motorola in a separate document titled: “IP
Summary 140806.pdf” and should be referenced for all vpn assignments, a subset of which is
shown above.
For design specific to Wateen Telecom, it has been advised that hosts (voip servers, either single
attached or dual attached via nic teaming) will connect directly to 7613 PE devices i.e. there is no
access switch providing distinction of access and distribution layer. There will be a dedicated link
between two 7613 PEs. This link will provide layer-2 connectivity between two switches for
specific access vlans on which voip servers will connect.
In this scenario, the HSRP configuration will track the priority depending on uplink availability,
should both uplink fail on active HSRP 7613 PE router, the virtual mac-address ownership will
automatically move to the backup 7613 PE router. Default priority of an HSRP router is 100, and
loss of link defaults to decrement of priority by 10.
For Wateen Telecom active router will have a priority of 120, since the PE is dual attached to the
single 12816 core router, loss of one link will not switch over virtual mac address ownership to
the backup router. In case when both uplinks are lost, HSRP priority will be decremented to 95 so
the backup router will takeover the virtual mac-address ownership and become the active router.
Once the uplink on active HSRP router is restored, the virtual mac-address ownership will be
restored back to the active router.
HSRP example configuration templates for both active and backup PE routers are shown below:
!
interface ethernet2/0
description Access Segment for voip server
bandwidth 1000000
ip vrf forwarding voip
ip address <interface ip address> <subnet-mask>
standby 2 ip <virtual ip-address>
standby 2 priority 120
standby 2 timers 1 4
standby 2 preempt
standby 2 track ethernet0/0 <uplink 1, priority decrement default of 10>
standby 2 track ethernet4/0 15 <uplink2, priority decrement of 15>
Figure 80 HSRP configuration on Active PE router
!
interface Ethernet2/0
description Access segment for voip servers
bandwidth 1000000
ip vrf forwarding voip
ip address 172.24.5.2 255.255.255.0
standby 2 ip 172.24.5.254
standby 2 timers 1 4
standby 2 preempt
Figure 81 HSRP configuration on backup PE router
A second scenario can be considered where yet another dedicated link is configured between two
7613-PE devices (via GE-SPA ports), this will be a layer-3 link will be included in IGP updates
and will provide reachability to a PE in case when both uplinks have been lost. Should this
scenario be chosen, we can potentially do away with HSRP link tracking requirements, as we will
always have an uplink path either directly to 12816 or via the adjacent 7613-PE device. The
HSRP configuration will be stay the same, except no uplinks will be tracked. The added benefit of
this approach is that only time HSRP timers kick into failure condition detection is when a
complete node is lost. This approach is preferred, however it adds complexity to the design and
uses additional ip addresses, precisious WAN ports etc.
The HSRP group assignment is typically done with odd vlans active on one PE e.g. PE1 while
even VLAN active on PE2. This is standard practice to distribute traffic and VLANs across
distribution routers (in our case these routers are also Provider Edge devices) from access
switches.
HSRP pre-empt delay can also be tuned best on the overall boot time of the host to avoid
possibility of having hsrp become active before the 7600 is ready ( Layer-3 with IGP
convergence) . This however needs to be tuned with some real life data and can be set using
following command: “standby 1 preempt delay minimum <sec>”
Overall access architecture (in terms of access to the customer) is lead by Cisco partner, Motorola.
It is therefore not covered in this document and is considered out of scope for the purpose of this
low level design. However based on discussion with Wateen Telecom and Motorola, following is
understood to be a high level representation of access network.
Following are generic recommendations based on above mentioned scenario. It is suggested that
customer consider these recommendation within the context of access architecture and only apply
relevant features/configurations where deemed appropriate. It has been advised that Cisco
Systems should consider access as a simple Ethernet link, specifically a .1Q trunk terminating on
a PE device via Gigabit Ethernet links.
It is required that we assure no spanning tree loops exist. This implies no layer2 loop when
connecting PE devices to access infrastructure.
Enable UDLD
UDLD in aggressive mode should be configured when connecting to access switch infrastructure
to deal with unidirectional link failures. Following syntax can be used to enable UDLD on an
interface:
interface GigabitEthernet1/0/28
udld port aggressive
Figure 85 Enabling Udld
UDLD should be enabled on 7600 PE ports connecting to access switches for unidirectional link
failure.
There are a couple of ways to change what VLANs are allowed on a trunk:
PE(config-if)#switchport trunk allowed vlan ?
WORD VLAN IDs of the allowed VLANs when this port is in trunking mode
add add VLANs to the current list
all all VLANs
except all VLANs except the following
none no VLANs
remove remove VLANs from the current list
When adding new VLANs to an existing trunk it is very important to use the “ADD” keyword. If
that is not done, all currently allowed VLANs will be removed from the allow list.
interface GigabitEthernet1/0/1
switchport trunk encapsulation dot1q
switchport trunk allowed vlan 1-299,350,401-4094
switchport mode trunk
no cdp enable
spanning-tree bpduguard enable
!
• At least one VLAN must be available for native VLAN tagging (vlan dot1q tag native
option). If all the available VLANs are used and then trying to enable the vlan dot1q tag
native option, the option will not be enabled.
• By enabling this command on 7600 (SP core switch), tag native VLAN egress traffic and
drop untagged native VLAN ingress traffic
• If this option is enabled on one switch and disabled on another switch, all traffic is
dropped; all access switches must have this option configured the same on each switch.
For simplicity and to avoid any interoperability/misconfiguration issues (at the access layer), it is
recommended to use first option, i.e. dedicating an unused vlan as native vlan in each trunk port.
The section details the Ethernet over MPLS design to be used by Wateen Telecom. The basic
premise is that a customer will be provided layer-2 connectivity across Wateen Telecom’s
IP/MPLS backbone core to other sites of the same customer. This customer is not interested in
Layer-3 vpn services (i.e. doesn’t peer with serive provider’s PE device) and instead is only
interested in building a layer-2 pipe to its multiple destinations. This form of connectivity is point
to point. i.e. an Ethernet link is provided between any two sites of the customer using a tunneling
mechanism by Wateen Telecom’s IP/MPLS core.
AToM overview
Virtual Private Wire Services (VPWS) allow the carriage of common layer 2 protocols across an
MPLS network in a point-to-point topology. This includes layer 2 protocols such as Ethernet,
Frame Relay (FR) etc. This service is commonly referred to as a Pseudo Wire service as it
provides a virtual direct neighbouring between CE devices. VPWS is enabled in Cisco PE routers
by the feature Any Transport over MPLS (AToM).
Any Transport over MPLS (AToM) provides a common framework to encapsulate and transport
supported Layer 2 traffic types over an MPLS network core.
AToM supports the following like-to-like transport types:
• ATM AAL5 over MPLS
• ATM Cell Relay over MPLS
• Ethernet over MPLS (VLAN and port modes)
• Frame Relay over MPLS
• PPP over MPLS
• HDLC over MPLS
The following process shows a Layer 2 packet traveling from Customer Edge 1 (CE1) across the
MPLS network to CE 2:
1- L2 transport route entered on ingress PE
2- PE1 starts LDP session with PE2 if one does not already exist
3- PE1 allocates a VC label for the new interface & binds to configured VCID
4- PE1 sends a label mapping message containing VC FEC TLV & VC label TLV
5- PE2 receives VC FEC TLV & VC label TLV that matches local VCID
6- PE2 repeats steps 1-5 so that bi-directional label/VCID mappings are established
cannot use overlapping VLAN ranges, unless customer-provided 802.1Q headers are
themselves tunneled into SP-only significant 802.1Q headers.
• EoMPLS VCs as defined in the Martini draft are point-to-point transmissions only.
• Traffic sent between the imposition/disposition routers (between LERs) over an EoMPLS VC
takes the same path across the IP/MPLS backbone. The Label Switched Path (LSP) may
change due to routing changes inside the provider network.
• Adding or removing a point-to-point Layer 2 VC requires configuration of the two VC
endpoints (at the two LERs), as each VC is unidirectional. Provisioning a VC involves
defining an endpoint of a Layer 2 VC at each of the VLAN interfaces at the PE router on the
interface that connects to the CE.
• The two LERs at the ingress/egress points of the IP/MPLS backbone (the N-PE routers) are
the only routers with knowledge of the Layer 2 transport VCs. All other LSRs have no table
entries for the Layer 2 transport VCs. This implies that only the PEs requires software with
EoMPLS functionality, making it very easy to scale an EoMPLS-based network.
Vlan Use
0 Reserved by Standard
1 Default Vlan (Native)
2 – 1001 Usable Vlans
1002 – 1005 Cisco defaults for FDDI
and Token Ring
1006 - 4094 Usable Vlans
4095 Reserved by Standard
It should not be assumed that if a device supports the full range of Vlan ID’s that it also supports
their use concurrently. Most access ethernet switch devices support a smaller number of
concurrent active Vlan ID’s. The exact number is device dependent. In case of 7600, following
should be observed:
o It is possible to use Vlan 1 , but it cannot be deleted. It is highly recommended that vlan
1 should not be used for data traffic (customer or internal).
o Extended system ID should be enabled to use extended range of VLANs ( 1006-4094)
using the command: “spanning-tree extend system-id” in global configuration mode.
o While it is understood that there are no switches running Catalyst OS in the access side,
please note that switches running the Catalyst operating system do not support
configuration of VLANs 1006-1024.
o 7600 uses internal vlans for L3 interfaces etc; these should be accounted for when
determining maximum vlans available.
Following components are needed when configuring a pseudowire:
VC ID
This is a unique identifier that is shared between the two PE routers. The vcid is a 32-bit
identifier.
The combination of the peer-router-id and the VC ID must be a unique combination on the router.
Two circuits cannot use the same combination of peer-router-id and VC ID.
!
interface GigabitEthernet0/3
xconnect 192.168.2.4 10 pw-class port-to-vlan Å 10 is used as vcid here
Figure 89 L2vpn configuration
VC Label
AToM uses one MPLS label to identify the AToM VCs (VC label). Therefore, the minimum
MPLS label stack is 1 for directly connected AToM PEs, which are PE routers that do not have a
P router between them
carry all VC Label (Tunnel) information between the pair for all subsequent VCs. Please see LDP
section for configuration details.
MTU Issues
MTU values at each end of an EVC must match for the EVC to transition to UP state.
For this service type, point-to-point circuits between customer routers (CEs) are established
through an Ethernet over MPLS (EoMPLS) core. The customer only needs one physical
connection to the Service Provider (SP), but is able to direct traffic to different destination routers
via the appropriate VLAN-ID (a sub interface will be used on the 7600 for each vlan).
In an EoMPLS-based Layer 2 VPN implementation, a customer’s point-to-point traffic based
upon a specific VLAN-ID is mapped to an MPLS label switched path where it extends across the
MPLS backbone. If the customer access port is an 802.1Q trunk, in the PE there will be as many
EoMPLS VCs as there are 802.1Q VLAN IDs. In other words, one physical Ethernet customer
UNI can have different destinations for each point-to-point connection identified by a VLAN-ID.
In each PE a mapping is made between each VLAN and Virtual Circuit (VC) which is tunneled
through a Label Switched Path (LSP). As specified in the IETF Martini draft, EoMPLS
technology provides point-to-point connections only. As long as the system is provisioned so that
there are only two endpoints to the circuit, traffic will be properly carried across the network.
For this Ethernet Relay service, the customer CPE equipment is ideally a router. This offers the
benefit of having only one MAC address on each end of the point-to-point circuit; i.e., only one
sub interface or svi has to be provisioned per neighboring router.
It should also be noted that ERS gives the possibility of creating arbitrary topologies as per
customer request, e.g. full mush or partial mesh.
The ERS service is less suitable if the customer has no CE router or has the requirement of
connecting directly the SP’s CPE a server or host Layer 2 network with no routing capability.
This mode of operation does not require any OSM or SIP line cards, and can be used with a
regular switched interface as the core facing interface.
Even though the VLAN IDs are configured as subinterfaces, VLAN ID numbers must be unique
per the whole device. It is impossible to use the same VLAN ID on separate trunks, and it is
impossible to create an SVI interface with the same VLAN number.
In order to allow local VLAN switching, a different approach has to be taken, and the use of SVI
interfaces (interface Vlan) should be introduced.
In this approach the trunk port is configured a regular switchport, and dot1q trunking is enabled
on it.
Every vlan has to be created in the vlan database and an “interface vlan” would be created. All the
relevant configuration would be done on the “interface vlans”:
• For L3 VPN, the “ip vrf forwarding” command would be configured on the SVI.
• For L2 VPN, the “xconnect” command would be configured on the SVI. No ip address is
configured in such a case.
• For locally layer 2 switched vlans there is no need for configuring an SVI interface, but a vlan
has to be created in the vlan database by using the “vlan <vlan-id>” command.
vlan 2000
name customer1
interface Vlan 2000
no ip address
xconnect <remote PE> <VC ID> encapsulation mpls
!
In order for the above configuration to work the core facing uplink link MUST be over a WAN
connection of a SIP line card. The CE facing interface can be a regular switch line card.
When configuring the xconnect command on a SVI interface, the following error message may
appear:
%C6K_MPLS_COMMON-3-L3_CONFIG_NOT_RECOMMENDED: LAN interfaces have
MPLS configured.
Do not configure xconnect on interface vlans.
This is a warning message which is used to make sure the uplink is a SIP enabled interface. It
would appear if there is any interface with “mpls ip” enabled which is not a SIP/OSM interface.
Similarly layer 3 vpns would be configured using an SVI interface as indicated in access section.
This method of configuration is recommended for Wateen Telecom in case that local layer 2
switching between two branches of a customer e.g. site A and site B is required. This will only
work if uplink (i.e. core facing interface) is on a SIP linecard. This also means that in case of both
uplink failures at core PoPs, interconnect between two 7613s should be on a SIP line card (instead
of LAN based card (WS-67xx series)) otherwise this configuration will no longer work.
It is recommended to use this method from day one if local switching is planned, as migrating
between the two different methods would require a service interruption. It is also recommended to
keep internconnect between two 7613 chassis to be on SIP-line cards if this method is chosen.
The technique that allows a customer’s VLAN space to have some independence from the SPs
VLAN space is called Q-in-Q. This feature is set-up by configuring interfaces on the customers’
CEs as VLAN trunks. The corresponding interfaces on the SP PE edge are deliberately configured
as non-trunking with the provider VLAN IDs (native VLANs). As a result, the data traffic gets
VLAN tagged on both the CE and PE, and is therefore double-tagged within the SP’s switching
infrastructure. For this solution the SP network is not capable of looking into the Q-in-Q frame
and working with the double-encapsulated frame. Therefore all the frames with the same outer
VLAN tag are switched everywhere in the SP’s network. Frames that seem to be going to the
wrong VLAN are dropped when they reach the egress PE and the outer tag is popped. At this
point, unless there is a match with the inner Customer VLAN tag at that port, traffic is dropped.
The customers’ control traffic (namely CDP, VTP, and Spanning Tree PDUs) is tunneled through
the EoMPLS cloud by enabling the Protocol Tunneling feature on Q-in-Q ports.
If configured in conjunction with EoMPLS in the core, Q-in-Q may increase the VC’s space. It
can thus be seen to be a special case of the EMS service, however currently no true EWS service
is possible due to the inability of the QinQ network to emulate the transparency of a wire with
respect to protocols such as PagP, etc.
With port mode the full interface is forwarded on the remote PE through the MPLS backbone.
The PE would not be able to send any traffic by itself over the interface, and all traffic would flow
only over the PW.
7600 restrictions
• Ensure that the maximum transmission unit (MTU) of all intermediate links between
endpoints is sufficient to carry the largest Layer 2 packet received.
• EoMPLS supports VLAN packets that conform to the IEEE 802.1Q standard. The 802.1Q
specification establishes a standard method for inserting VLAN membership information into
Ethernet frames.
• If QoS is disabled globally, both the 802.1p and IP precedence bits are preserved. When the
QoS is enabled on a Layer 2 port, either 802.1q P bits or IP precedence bits can be preserved
with the trusted configuration. However, by default the unpreserved bits are overwritten by
the value of preserved bits. For instance, if you preserve the P bits, the IP precedence bits are
overwritten with the value of the P bits. PFC3BXL or PFC3B mode provides a new command
that allows you to trust the P bits while preserving the IP precedence bits. To preserve the IP
precedence bits, use the no mls qos rewrite ip dscp command.
• EoMPLS is not supported with private VLANs.
VPLS / H-VPLS
VPLS
It is understood that with suggested release of code:12.2(18)SXF, Wateen Telecom cannot deploy
VPLS on their network. This is due to the fact that VPLS support for SIP-400 is not available in
12.2(18)SXF5 release. An upgrade to 12.2(33)SRA (internally codenamed cascades) is required
for this functionality to be supported. Wateen Telecom has agreed not to deploy VPLS at this
stage. Generic information and sample configurations are included for reference and potentially
forward looking requirement.
VPLS enables Ethernet multipoint services over a packet-switched network infrastructure. VPN
users get an emulated LAN segment that offers a Layer 2 broadcast domain. End users perceive
the service as a virtual private Ethernet switch that forwards frames to their respective destination
within the VPN. The next figure shows the logical view of a VPLS connecting three sites. Each
customer edge (CE) device requires a single connection to the network to get full connectivity to
the remaining sites. A multipoint technology allows a user to reach multiple destinations through a
single physical or logical connection, which requires the network to make a forwarding decision
based on the destination of the packet. Within the context of VPLS, this means that the network
makes a forwarding decision based on the destination MAC address of the Ethernet frame. From
the end customer's perspective, a multipoint service is attractive because fewer connections are
required to get full connectivity between multiple points. An equivalent level of connectivity
based on a point-to-point technology requires a much larger number of connections or the use of
suboptimal packet forwarding.
Figure 94 VPLS
In its simplest form, a VPLS consists of a collection of sites connected to a number of provider
edge (PE) devices implementing the emulated LAN service. A virtual switching instance (VSI) is
used at each PE to implement the forwarding decisions of each VPLS. The PE devices make the
forwarding decisions between sites and encapsulate the Ethernet frames across a packet-switched
network using an Ethernet virtual circuit (VC) or pseudo-wire. PEs use a full mesh of Ethernet
VCs to forward the Ethernet frames between PEs. VPLS relies on the same encapsulation defined
for point-to-point Ethernet over MPLS. The frame preamble and frame check sequence (FCS) are
removed, and the remaining payload is encapsulated with a control word, a VC label, and an
Interior Gateway Protocol (IGP) or transport label. VPLS has been initially specified and
implemented over an MPLS transport.
PEs would automatically populate the VSI with the forwarding information required to switch
frames within the VPLS. PEs acquire this information using the standard MAC address learning
and aging functions used in Ethernet switching. The VSI forwarding information is updated with
the MAC addresses learned from physical ports and from the virtual circuits. These functions
imply that all broadcast, multicast, and destination unknown MAC addresses are flooded over all
ports and VCs associated with a VSI. PEs use split-horizon forwarding on the VCs to form a loop-
free topology. In this way, the full mesh of VCs provides direct connectivity between the PEs in a
VPLS, and there is no need to use more resource-intensive protocols to generate a loop-free
topology (for example, Spanning Tree Protocol, or STP).
Configuration on the PE
Creation of the virtual switch instances (VSIs) and associated VCs.
For PE 1
!
l2 vfi <PE-VPLS-A> manual
vpn id <ID number>
neighbor <PE2 loopback IP> encapsulation mpls
neighbor <PE3 loopback IP> encapsulation mpls
!
interface loopback <number>
ip address <IP address> <mask>
For PE 2
!
l2 vfi <PE-VPLS-A> manual
vpn id <ID number>
neighbor <PE1 loopback IP> encapsulation mpls
neighbor <PE3 loopback IP> encapsulation mpls
!
interface loopback <number>
ip address <IP address> <mask>
!
For PE 3
!
l2 vfi <PE-VPLS-A> manual
vpn id <ID number>
neighbor <PE1 loopback IP> encapsulation mpls
neighbor <PE2 loopback IP> encapsulation mpls
!
interface loopback <number>
ip address <IP address> <mask>
Configuration of the CE interface side (there can be multiple Layer 2 interfaces in a VLAN).
On all the CE
Configuration of the attachment circuit (VLAN) associated with the VSI and enablement of the
layer 2 VLAN instance.
On all the PE
Interface vlan <vlan number>
no ip address
xconnect vfi <PE-VPLS-A>
VPLS Restrictions
• On the Cisco 7600 series router, the Virtual Forwarding Instance (VFI) is supported only with
the interface vlan command.
7600 restriction
• The core facing interface must be installed on an OSM module
• The Customer facing interfaces are all Ethernet/ Fast Ethernet/ Gigabit Ethernet interfaces
based on Layer 2 Catalyst LAN ports
This section details Wateen Telecom’s Dial-up network infrastructure. The Dial infrastructure will be
used to provide the following services:
Remote access to MPLS VPNs
As a backup to MPLS VPN primary link
Backup internet connection
Before discussing Wateen Telecom’s Dial design specifics it is believed that the reader should be familiar
with certain technologies and protocols that are integral to the overall operation of dial services. The next
section will discuss each of these protocols and technologies in sufficient detail in order to facilitate
satisfactory understanding of Wateen Telecom’s Dial Services architecture.
Technology Primers
For the purpose of this section the terms NAS/LAC are used interchangeably with LAC, VHG/PE with
LNS and CAR with RADIUS.
PPP
PPP is a symmetrical peer-to-peer protocol used for transporting Layer 2 and Layer 3 traffic over point-
to-point links. There are three main components:
Encapsulation
Link Control Layer (LCP)
Network Control Layer (NCP)
First off, the datagram is encapsulated in PPP. Secondly, there is the Link Control Layer (LCP) part,
which allows for configuration options to be negotiated allowing Link Establishment and Network
Control Protocols (NCP), which are negotiated for each of the Layer 3 protocols that will run on the link.
During the life of a PPP session, there are four distinct phases the link goes though:
Link Establish
Authentication
Network Layer
Link Termination
As part of the Link Establish phase, PPP uses an LCP function that must be completed and declared open
prior to entering the Authentication phase (this is an optional phase dependant on end-user configuration)
and negotiating the opening of the Network Layer. LCP is also used to terminate the PPP link.
The Authentication phase is implementation-specific and not a mandatory requirement to move from
LCP to NCP. If negotiated and agreed during the LCP phase, then the remote peer must identify itself and
pass the agreed authentication method prior to PPP moving to Network Layer.
Network Control Protocol (NCP) negotiation ensures both peers agree on the characteristics of the Layer
3 protocol. In the case of IP, the Control Protocol is called IPCP. In addition to the negotiation between
peers, there is also an element of assignment. This is common with Windows-type remote access clients
who have no allocated IP address and rely on the Service Provider allocating on connection.
The Link termination phase can be entered anytime during the lifecycle of the call. LCP is used to deliver
the Termination request.
L2TP extends the Point to Point nature of PPP by providing an encapsulation method for sending
tunneled PPP frames, thereby allowing the PPP endpoints to be tunneled over a packet switched network.
This is most commonly deployed in Remote Access type scenarios using the internet to offer intranet type
services; a concept of a Virtual Private Network (VPN).
The two primary physical elements of L2TP are the L2TP Access Concentrator (LAC) and the L2TP
Network Server (LNS).
Control Message
Data Message
L2TP passes Control and Data messages over separate Control and Data channels. The in-band Control
channel is used to pass sequenced control connection management, call management, error reporting and
session control messages. Initiation of the Control Connection is not specific to either the LAC or the
LNS rather the tunnel originator and receiver that has relevance in the Control Connection Establishment.
A shared secret challenge authentication method is used between the tunnel endpoints
Data Messages are used to encapsulate the PPP frames that are sent into the L2TP tunnel
L2TP uses the registered UDP port 1701; the whole L2TP packet is encapsulated within the UDP
datagram. As per normal UDP operation, the tunnel initiator selects an available UDP port and sends port
number 1701 to the UDP destination. In the reply, the destination port number is the same as the source
port number used in the incoming UDP header; the source port is set based on free port found. Once the
source and destination ports are established, they must remain the same for the duration of the tunnel. In
Cisco IOS the source and destination port numbers are always set to UDP port number 1701.
AAA
AAA describes the access control and accounting requirements to allow control of who is allowed to
access the network and what services they are allowed to use and then how to report on that particular
user’s activity for the lifetime of that connection. Various methods are employed to perform these
functions and network security services. Commonly these are set up and administered on a centralized
database. To send the information to the centralized database and forward the response to the respective
network device a protocol is required.
Authentication
Provides a method of identifying users - login and password. Using challenge and response to allow and
refuse the connection. Authentication is the way a user is identified prior to being allowed access to the
network and network services.
Authorization
Provides a method for remote access control, including one-time authorization or authorization for each
service, per-user account list and profile, user group support amongst many others. AAA authorization
works by assembling a set of attributes that describe what the user is authorized to perform.
Accounting
Provides a method for collecting and sending security server information used for billing, auditing, and
reporting, such as user identities, start and stop times, executed commands (such as PPP), number of
packets, and number of bytes. Accounting enables you to track the services users are accessing as well as
the amount of network resources they are consuming.
RADIUS
The RADIUS protocol is an access server authentication and accounting protocol. Communication
between a client and the RADIUS server is based on the User Datagram Protocol (UDP). Generally, the
RADIUS protocol is considered a connectionless service. Issues related to server availability,
retransmission, and timeouts are handled by the RADIUS-enabled devices rather than the transmission
protocol. RADIUS is a client/server protocol. The client passes user information to the configured
RADIUS servers and acts on the response that is returned. RADIUS servers receive user connection
requests, authenticate the user, and then return the configuration information necessary for the client to
deliver services to the user.
Introduction
As mentioned before Wateen Telecom’s dial infrastructure will be used to provide the following services:
Remote access to MPLS VPNs
As a backup to MPLS VPN primary link
Backup internet connection
Bill of Quantity
The following figure details the equipment that has been ordered for the dial network and its distribution:
As part of the Wateen Telecom network buildout process Cisco Access Servers AS5400XM and
AS5350XM’s will be deployed throughout the Main, City and Satellite PoPs to provide dial up internet
access as well as dial access into MPLS VPNs. A Cisco 7206 router will be deployed in each Main PoP to
offer Virtual Homegateway/Provider Edge (VHG/PE) support. Cisco Access Registrar (CAR) has been
added in the Data Center at Lahore to authenticate, authorize and account the dialin users and assign them
to the relevant VRF at the Home Gateway
The following table summarizes the hostname, type, location, and deployment role of various dial
components as applicable to Wateen’s network. The hostnames mentioned in the following table will be
used throughout the documentation, whenever a reference has to be made to dial components.
Deployment Device
Device Hostname Deployed As
Location Model
Khi-7206-VHG1 Karachi 7206VXR VHG (Virtual Home Gateway) / PE (LNS)
Khi-AS5400-AS1 Karachi AS-5400XM NAS / LAC (Network Access Server / L2TP Access Concentrator)
Hyd-AS5350-AS1 Hyderabad AS-5350XM NAS / LAC (Network Access Server / L2TP Access Concentrator)
Qta-AS5350-AS1 Quetta AS-5350XM NAS / LAC (Network Access Server / L2TP Access Concentrator)
Skr-AS5350-AS1 Sukkhur AS-5350XM NAS / LAC (Network Access Server / L2TP Access Concentrator)
Ryk-AS5350-AS1 R.Y.Khan AS-5350XM NAS / LAC (Network Access Server / L2TP Access Concentrator)
Please refer to the Layer 2 Tunneling Protocol sub-section for details on LAC and LNS functionality.
Additionally, the above table does not show Cisco Access Registrar (CAR). A pair of CAR servers will
be deployed at Wateen’s Lahore Data Center for providing AAA services.
As part of the initial connection sequence when a user connects they are prompted for an authentication
password. When the authentication response is received this triggers the NAS/LAC to send the CAR an
Access-Request. The Access-Request message contains various attributes including the Domain Name in
the User-Name attribute, NAS/LAC-IP-Address and Acct-Session-Id. Each NAS/LAC uniquely identifies
itself to the CAR by the source address of the IP packets that it generates will match a corresponding
entry in the CAR clients file.
If the Access-Request is invalid, the CAR will respond to the NAS/LAC with an Access-Reject, this will
cause the NAS/LAC to terminate PPP and disconnect the user. Otherwise, the CAR returns Access-
Accept to the NAS/LAC. The Access-Accept for an MPLS user contains various Tunnel attributes that
allows the NAS/LAC to build an L2TP tunnel to the appropriate VHG/PE router. For an Internet user the
Access-Accept will contain information that will allow the user to terminate on the NAS/LAC, be
allocated an IP address and connect to the internet.
As part of the L2TP protocol the NAS/LAC signals the negotiated LCP state to the VHG/PE thus
allowing the VHG/PE to bring up PPP LCP and continue the PPP Authentication phase and if successful
the NCP phase (see technology briefs below for detailed protocol explanation).
Once the L2TP session is established the NAS/LAC sends an Accounting-Start to the CAR accounting
server. Amongst others the following attributes are included - User-Name, Acct-Session-Id and Tunnel-
Server-Endpoint.
When the VHG/PE accepts the tunnel/session it sends an Access-Request to the CAR then completes the
remote user's authentication. The VHG/PE creates a unique virtual access interface for each dialin user,
completes PPP NCP, and terminates the PPP session into an IP connection.
The remote user's session becomes part of the appropriate VRF. An Accounting-Start packet is now sent
and subsequent interim accounting packets are sent at a configured interval.
Packets now flow from and to the remote user. When the user disconnects the PPP session will terminate
and the L2TP session will disconnect then both the VHG/PE and the NAS/LAC will send Accounting-
Stops.
For the internet users the NAS/LAC sends Accounting-Start, Interim and Stop packets throughout the
lifetime of the call.
1. The remote user initiates a PPP connection to the NAS/LAC using either analog POTS or ISDN.
2. NAS/LAC accepts the connection and a PPP link is established.
3. The NAS/LAC partially authenticates the user with CHAP or PAP. The Domain name is used to
determine whether the user is an MPLS customer. The NAS/LAC queries the CAR for tunnel information.
The CAR returns the address of a VHG/PE.
4. If an L2TP tunnel does not exist, the NAS/LAC initiates a tunnel to the VHG/PE. The NAS/LAC and the
VHG/PE authenticate each other before any sessions are attempted within a tunnel.
5. Once the tunnel exists, a session within the tunnel is created for the remote user and the PPP connection is
extended to terminate on the VHG/PE.
6. The NAS/LAC propagates all available PPP information (the LCP negotiated options and the partially
authenticated CHAP/PAP information) to the VHG/PE.
7. The VHG/PE associates the remote user with a specific customer MPLS VPN. The VPN's VRF (routing
table and other information associated with a specific VPN) has been preinstantiated on the VHG/PE.
8. The VHG/PE completes remote user's authentication with the CAR.
9. The VHG/PE obtains an IP address for remote user.
10. The remote user becomes part of the customer VPN. Packets flow from and to the remote user.
The dial backup connection sequence is identical to the dial in model however for the dial backup
solution the CPE has two connections into the network and loses the primary interface.
1. The CPE loses the primary interface and initiates backup PPP connection to the NAS/LAC using
either analog POTS or ISDN.
2. NAS/LAC accepts the connection and a PPP link is established.
3. The NAS/LAC partially authenticates the user with CHAP or PAP. The Domain name is used to
determine whether the user is an MPLS customer or Internet customer; the absence of a domain in the
username represents internet users. The NAS/LAC queries the CAR for tunnel information. If the user
is an Internet customer authentication continues on the NAS/LAC. If the user is an MPLS customer
the CAR returns the address of a VHG/PE.
4. If an L2TP tunnel does not exist, the NAS/LAC initiates a tunnel to the VHG/PE. The NAS/LAC and
the VHG/PE authenticate each other before any sessions are attempted within a tunnel.
5. Once the tunnel exists, a session within the tunnel is created for the remote user and the PPP
connection is extended to terminate on the VHG/PE.
6. The NAS/LAC propagates all available PPP information (the LCP negotiated options and the partially
authenticated CHAP/PAP information) to the VHG/PE.
7. The VHG/PE associates the remote user with a specific customer MPLS VPN. The VPN's VRF
(routing table and other information associated with a specific VPN) has been preinstantiated on the
VHG/PE.
8. The VHG/PE completes remote user's authentication with the CAR.
9. The VHG/PE obtains an IP address for remote user via an IP Pool or delivered from CAR as Framed-
IP-Address.
10. The remote user becomes part of the customer VPN. Packets flow from and to the remote user.
11. When the primary link is restored the primary route is also restored and the backup is terminated by
the remote user
12. The backup route is deleted by the VHG/PE
The complete call flow is detailed below, showing the initial PPP connection through to the final ISP call
acceptance.
In this section we will discuss the dial component physical connectivity details as they pertain to each
type of Wateen PoP; Main PoP, City PoP, and Satellite PoP.
One AS5400XM will be deployed in the Lahore, Karachi and Islamabad PoPs. All remaining PoPs will
be deployed with a single AS5350XM.
A 7206/NPE-G1 will be deployed as VGH/PE in the Lahore, Karachi, Islamabad and Faisalabad PoPs.
All City and Satellite PoPs will be equipped with one AS5350XM access server. Within a City or
Satellite PoP the AS5350XM access server will connect to the local PE via a Gigabit Ethernet UTP
interface.
The following diagram shows what the Dial infrastructure of a City and a Satellite PoP will look like. For
the sake of simplicity, only Gujranwala has been chosen as the reference City PoP while Gujrat has been
chosen as the reference Satellite PoP. All City and Satellite PoPs will follow the same NAS/LNS
connectivity model as depicted in the following figure:
Local Remote
Local Device Local Interface Local Interface IP Remote Remote Remote Interface
Interface Interface Comments
Name Type Address Device Name Interface Type IP Address
ID ID
Loopback 0 Virtual - Public 58.27.190.55 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Gi 0 / 1 GE - Fiber - MM 58.27.194.194 / 30 Khi-7613-PE1 Gi 1/22 GE - Fiber - MM 58.27.194.193 / 30 Provides L3 interconnect to local PE1
Khi-7206-VHG1
Gi 0 / 2 GE - Fiber - MM 58.27.194.198 / 30 Khi-7613-PE2 Gi 1/22 GE - Fiber - MM 58.27.194.197 / 30 Provides L3 interconnect to local PE2
Gi 0 / 3 GE - Fiber - MM - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.45 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.1 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Khi-AS5400-AS1 Loopback 2 Virtual - Private 172.24.0.1 / 23 - N/A - - N/A - - N/A - - N/A - Used to originate summary route for configured dial pool
Gi 0/0 GE - Copper 58.27.194.130 / 30 Khi-7613-PE1 Gi 12/46 GE - Copper 58.27.194.129 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper 58.27.194.134 / 30 Khi-7613-PE2 Gi 12/46 GE - Copper 58.27.194.133 / 30 Provides L3 interconnect to local PE2
Loopback 0 Virtual - Public 58.27.190.47 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.3 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Hyd-AS5350-AS1 Loopback 2 Virtual - Private 172.24.2.1 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.194.138 / 30 Hyd-7609-PE1 Gi 8/46 GE - Copper 58.27.194.137 / 30 Provides L3 interconnect to Local PE
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.49 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.5 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Qta-AS5350-AS1 Loopback 2 Virtual - Private 172.24.2.129 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.194.142 / 30 Qta-7609-PE1 Gi 8/46 GE - Copper 58.27.194.141 / 30 Provides L3 interconnect to local PE
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Skr-AS5350-AS1 Loopback 0 Virtual - Public 58.27.190.51 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.7 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Loopback 2 Virtual - Private 172.24.3.1 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.194.146 / 30 Skr-7609-PE1 Gi 8/46 GE - Copper 58.27.194.145 / 30 Provides L3 interconnect to local PE
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.53 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.9 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Ryk-AS5350-AS1 Loopback 2 Virtual - Private 172.24.3.129 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.194.150 / 30 Ryk-7609-PE1 Gi 8/46 GE - Copper 58.27.194.149 / 30 Provides L3 interconnect to local PE
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.29 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for device management
Gi 0 / 1 GE - Fiber - MM 58.27.192.194 / 30 Lhe-7613-PE1 Gi 1/22 GE - Fiber - MM 58.27.192.193 / 30 Provides L3 interconnect to local PE1
Lhe-7206-VHG1
Gi 0 / 2 GE - Fiber - MM 58.27.192.198 / 30 Lhe-7613-PE2 Gi 1/22 GE - Fiber - MM 58.27.192.197 / 30 Provides L3 interconnect to local PE2
Gi 0 / 3 GE - Fiber - MM - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.21 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.16 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Lhe-AS5400-AS1 Loopback 2 Virtual - Private 172.24.32.1 / 23 - N/A - - N/A - - N/A - - N/A - Used to originate summary route for configured dial pool
Gi 0/0 GE - Copper 58.27.192.130 / 30 Lhe-7613-PE1 Gi 12/46 GE - Copper 58.27.192.129 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper 58.27.192.134 / 30 Lhe-7613-PE2 Gi 12/46 GE - Copper 58.27.192.133 / 30 Provides L3 interconnect to local PE2
Loopback 0 Virtual - Public 58.27.190.23 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.18 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Skt-AS5350-AS1 Loopback 2 Virtual - Private 172.24.34.1 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.192.138 / 30 Skt-7609-PE1 Gi 8/46 GE - Copper 58.27.192.137 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.25 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.20 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Guj-AS5350-AS1 Loopback 2 Virtual - Private 172.24.34.129 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.192.142 / 30 Guj-7609-PE1 Gi 8/46 GE - Copper 58.27.192.141 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Gjt-AS5350-AS1 Loopback 0 Virtual - Public 58.27.190.27 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.22 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Loopback 2 Virtual - Private 172.24.35.1 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.192.146 / 30 Gjt-7609-PE1 Gi 8/46 GE - Copper 58.27.192.145 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.81 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for device management
Gi 0 / 1 GE - Fiber - MM 58.27.196.194 / 30 Isb-7613-PE1 Gi 1/22 GE - Fiber - MM 58.27.196.193 / 30 Provides L3 interconnect to local PE1
Isb-7206-VHG1
Gi 0 / 2 GE - Fiber - MM 58.27.196.198 / 30 Isb-7613-PE2 Gi 1/22 GE - Fiber - MM 58.27.196.197 / 30 Provides L3 interconnect to local PE2
Gi 0 / 3 GE - Fiber - MM - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.71 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.32 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Isb-AS5400-AS1 Loopback 2 Virtual - Private 172.24.64.1 / 23 - N/A - - N/A - - N/A - - N/A - Used to originate summary route for configured dial pool
Gi 0/0 GE - Copper 58.27.196.130 / 30 Isb-7613-PE1 Gi 11/46 GE - Copper 58.27.196.129 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper 58.27.196.134 / 30 Isb-7613-PE2 Gi 11/46 GE - Copper 58.27.196.133 / 30 Provides L3 interconnect to local PE2
Loopback 0 Virtual - Public 58.27.190.73 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.34 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Pes-AS5350-AS1 Loopback 2 Virtual - Private 172.24.66.1 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.196.138 / 30 Pes-7609-PE1 Gi 8/46 GE - Copper 58.27.196.137 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.75 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.36 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Dik-AS5350-AS1 Loopback 2 Virtual - Private 172.24.66.129 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.196.142 / 30 Dik-7609-PE1 Gi 8/46 GE - Copper 58.27.196.141 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.77 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.38 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Jeh-AS5350-AS1 Loopback 2 Virtual - Private 172.24.67.1 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.196.146 / 30 Jeh-7609-PE1 Gi 8/46 GE - Copper 58.27.196.145 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.79 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.40 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Abt-AS5350-AS1 Loopback 2 Virtual - Private 172.24.67.129 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.196.150 / 30 Abt-7609-PE1 Gi 8/46 GE - Copper 58.27.196.149 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.111 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for device management
Gi 0 / 1 GE - Fiber - MM 58.27.198.194 / 30 Fsd-7613-PE1 Gi 1/22 GE - Fiber - MM 58.27.198.193 / 30 Provides L3 interconnect to local PE1
Fsd-7206-VHG1
Gi 0 / 2 GE - Fiber - MM 58.27.198.198 / 30 Fsd-7613-PE2 Gi 1/22 GE - Fiber - MM 58.27.198.197 / 30 Provides L3 interconnect to local PE2
Gi 0 / 3 GE - Fiber - MM - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.99 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.48 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Fsd-AS5350-AS1 Loopback 2 Virtual - Private 172.24.96.1 / 25 - N/A - - N/A - - N/A - - N/A - Used to originate summary route for configured dial pool
Gi 0/0 GE - Copper 58.27.198.130 / 30 Fsd-7613-PE1 Gi 11/46 GE - Copper 58.27.198.129 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper 58.27.198.134 / 30 Fsd-7613-PE2 Gi 11/46 GE - Copper 58.27.198.133 / 30 Provides L3 interconnect to local PE2
Loopback 0 Virtual - Public 58.27.190.101 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.50 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Sgd-AS5350-AS1 Loopback 2 Virtual - Private 172.24.96.129 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.198.138 / 30 Sgd-7609-PE1 Gi 8/46 GE - Copper 58.27.198.137 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.103 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.52 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Mul-AS5350-AS1 Loopback 2 Virtual - Private 172.24.97.1 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.198.142 / 30 Mul-7609-PE1 Gi 8/46 GE - Copper 58.27.198.141 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.105 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.54 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Swl-AS5350-AS1 Loopback 2 Virtual - Private 172.24.97.129 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.198.146 / 30 Swl-7609-PE1 Gi 8/46 GE - Copper 58.27.198.145 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.107 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.56 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Bwp-AS5350-AS1 Loopback 2 Virtual - Private 172.24.98.1 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.198.150 / 30 Bwp-7609-PE1 Gi 8/46 GE - Copper 58.27.198.149 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
Loopback 0 Virtual - Public 58.27.190.109 / 32 - N/A - - N/A - - N/A - - N/A - Loopback interface for management functions
Loopback 1 Virtual - Private 192.168.64.58 / 32 - N/A - - N/A - - N/A - - N/A - Provides address to unnumbered interfaces
Okr-AS5350-AS1 Loopback 2 Virtual - Private 172.24.98.129 / 25 - N/A - - N/A - - N/A - - N/A - Will be used in future if dynamic routing is enabled
Gi 0/0 GE - Copper 58.27.198.154 / 30 Okr-7609-PE1 Gi 8/46 GE - Copper 58.27.198.153 / 30 Provides L3 interconnect to local PE1
Gi 0/1 GE - Copper - Not Used - - N/A - - N/A - - N/A - - N/A - This interface will not be used
IP Addressing Details
IP Addressing for both NAS/LAC and VHG/PE devices is described separately in the following sub-
sections:
NAS IP Addressing
NAS devices will require IP Addresses for:
Physical and virtual interfaces
Address pools for internet users
For physical and logical interface IP Addressing please refer to Table 16.
A separate IP Address pool has been allocated to each NAS device for internet dialup users. In order to
conserve public IP Address space RFC 1918 addresses have been used for this purpose. The size of the
pool assigned to each NAS depends on the maximum number of simultaneous users the NAS can
support; this in turn depends on the number of ordered PRIs.
The private pool 172.24.0.0 / 16 has been reserved for Dialup users. This pool has been further subnetted
to allocate each NAS the required number of addresses while still maintaining regional hierarchy. The
following table shows how this prefix has been subnetted.
Using the above ip schema as the base the following table has been formulated to provide details such as
maximum number of concurrent users, number of PRIs, assigned pool, etc. for each NAS:
Max. # of
NAS/LAC No. of PRIs Concurrent Dialup Suggested Pool
Location Model Assigned Pool
Hostname (Per BoQ) Users (As per # of Size
PRIs)
VHG/PE Addressing
Each VHG/PE device will require IP Addresses for:
Physical and Logical interfaces
Address pools per customer VPN.
For physical and logical interface (management loopback only) addressing details please refer to Table
16.
Address pools per MPLS VPN will have to be defined by Wateen in consultation with the customers. One
option could be to use one unique pool per VHG/PE and reuse it across all customer VPNs. Wateen will
have to ensure that this does not overlap with customer’s internal addressing. For example, Wateen can
configure VHG/PE devices as:
Karachi VHG/PE Æ Use 172.24.4.0 / 23 for MPLS VPNs
Lahore VHG/PE Æ Use 172.24.36.0 / 23 for MPLS VPNs
Islamabad VHG/PE Æ Use 172.24.68.0 / 23 for MPLS VPNs
Faisalabad VHG/PE Æ Use 172.24.100.0 / 23 for MPLS VPNs
These pools can be reused per VRF using the ‘overlapping address pools’ feature described earlier. Please
note that this is just a recommendation. Wateen Telecom can adopt this recommendation in case they
have not formulated a policy of their own on the subject matter.
It shall be noted that for each customer VPN on loopback interface will need to be created per VHG/PE.
The purpose of this interface is to preinstantiate the customer routes in the VPN so that a user dialing in
does not experience unnecessary delay (due to BGP convergence) in having connectivity established. It is
recommended that a single /32 address be used on all such loopbacks (all customer VRFs) to conserve
address space.
We have used Hyderabad City PoP as an example; however, the same routing architecture applies to all
Wateen City/Satellite PoPs.
utilize the dual entry/exit points both during normal conditions and during periods of a single link failure.
Consequently each NAS device installed in a Main PoP will be configured to run OSPF. The Core OSPF
Area 0 will be extended to these NAS devices by configuring them to include their Gigabit Ethernet
interfaces (connecting to local PE) in the OSPF Area 0. Additionally, the Loopback 2 interface of these
NAS devices will also be included in the OSPF Area 0 in order to send a summary of the configured dial-
up towards the rest of the network.
Since a default route is not being propagated in the core of the network (core PEs have visibility to full
internet routes) we need a mechanism to add a default route to the dual-homed NAS devices even though
they are configured to run OSPF with the core. Keeping simplicity in mind it is suggested to configure
two static default routes per NAS device installed in a Main PoP. One default will point to local PE1
device while the other default will point to the local PE2 device. ECMP (Equal Cost Multipath) load
sharing will be performed by the NAS device using the underlying CEF per destination load sharing
scheme.
The following figure is a graphical representation of the routing architecture of NAS devices located at
Wateen Main PoPs:
The digram uses Lahore Main PoP as an example; however, the routing architecture of all Main PoPs will
follow a similar design from NAS connectivity perspective.
VHG/PE devices located in Main PoPs will also be dual-homed; one Gigabit Ethernet link will connect
each such device to each of the local PE routers. Since the purpose of the VHG/PE routers is to provide
remote access to MPLS VPNs, they will require Multiprotocol BGP (MP-BGP) configuration for vpnv4
addresses. The VHG/PE routing architecture can therefore be summarized by the following points:
OSPF Area 0 to extend to the VHG/PE Gigabit Ethernet interfaces connecting to the local PE
routers.
The figure uses Lahore VHG/PE deployment as an example only. All VHG/PE deployments will follow
the same routing architecture as depicted in Figure 105.
VHG/PE Resiliency
VHG/PE redundancy is achieved by passing tunnel attributes from CAR to the NAS/LAC. By specifying
(on a per domain basis) multiple tunnel endpoints and tunnel preference values it is possible to achieve
failover and load balancing. The recommended failover and load balancing scheme will be for the
NAS/LAC’s to be returned tunnel attributes that allow the primary VGH/PE (that is the geographically
closest) to be the primary tunnel endpoint. The remaining three VHG/PE’s will then have equal priority in
the next priority group. An example being the NAS/LAC in Lahore will be returned the VGH/PE in
Lahore as its primary tunnel endpoint. Should this VHG/PE not be available the NAS/LAC will failover
to the next priority group that will have the endpoints of the VHG/PE’s in Karachi, Islamabad and
Faisalabad. The subsequent connections from this particular NAS/LAC will then be shared amongst the
three VHG’s above, until the primary endpoint becomes available again. The following table provides a
map for VHG/PE resiliency showing which VHG/PE will be primary and which VHG/PEs will be
backup for a particular NAS.
NAS/LAC
Location Model Primary VHG/PE Backup VHG/PE Group
Hostname
Khi-AS5400-AS1 Karachi AS-5400XM
Hyd-AS5350-AS1 Hyderabad AS-5350XM Lhe-7206-VHG1
Qta-AS5350-AS1 Quetta AS-5350XM Khi-7206-VHG1 Isb-7206-VHG1
Lyp-7206-VHG1
Skr-AS5350-AS1 Sukkhur AS-5350XM
Ryk-AS5350-AS1 R.Y.Khan AS-5350XM
There is no special configuration required on the NAS or the VHG/PE to implement this resiliency
scheme. The tunnel attributes will be returned dynamically from the CAR and the NAS will always
initiate a tunnel to the primary VHG/PE. If the primary VHG/PE is unavailable then the NAS will attempt
to establish a tunnel with one of the three backup VHG/PE routers.
For the VHG/PE resiliency functionality to be effective it is reqiured that VPNs requiring resiliency are
pre-configured on all four VHG/PE routers. To ensure that backup VHG/PE devices have adequate
IPaddress resources to accommodate the clients normally associating with the primary VHG/PE, it is
recommended to over-provision the dial pools per customer.
Configuration Templates
In this section an attempt has been made to provide configuration snippets that will be used to provision
various components of Wateen Telecom’s Dial infrastructure. Please note that these templates are strictly
for the dial configuration parameters; it is assumed that other configuration, specific to device security
and ip connectivity, will co-exist with these templates.
NAS/LAC
http://www.cisco.com/en/US/tech/tk801/tk36/technologies_tech_note09186a008009428b.shtml
http://www.cisco.com/univercd/cc/td/doc/product/access/acs_serv/as5400/sw_conf/atnextpt.htm
http://www.cisco.com/warp/public/471/recc_modemcaps.html
!
interface Serial1:15
no ip address
encapsulation ppp
no logging event link-status
dialer rotary-group 1
no snmp trap link-status
isdn switch-type primary-net5
isdn incoming-voice modem
Repeat the above for the number of configured PRI interfaces
!
interface Group-Async1
ip unnumbered loopback 1
encapsulation ppp
no logging event link-status
dialer in-band
dialer rotary-group 1
async mode dedicated
no snmp trap link-status
peer default ip address pool internet_users
ppp authentication chap pap callin
group-range 1 XX
The group-range should cover all modems in the access server
!
interface Dialer1
ip unnumbered Loopback0
encapsulation ppp
no logging event link-status
no snmp trap link-status
peer default ip address pool internet_users
ppp authentication chap pap callin
!
line 1 XX
The line number should cover all modems in the access server
no flush-at-activation
modem InOut
modem autoconfigure type Wateen_MODEMCAP
no modem log rs232
no exec
transport input all
transport output none
PPP
Generic config for the IP local-pool, this is defined on the NAS/LAC to allocate Internet users IP
addresses as they connect. The pool name can be configured on the physical interface or downloaded
from RADIUS.
It is recommended that the pool range is summarized as an aggregate route to be advertised by OSPF.
L2TP
The following is the generic required L2TP configuration; the per-user tunneling information will be
retrieved from the CAR.
vpdn enable
vpdn source-ip A.B.C.D (Management Loopback Address)
vpdn search-order domain
Tunnel-Type = :2:L2TP,
Tunnel-Preference = :2:2,
Tunnel-Client-Auth-ID = :2:"NAS/LAC",
Tunnel-Server-Endpoint = :2:"E.F.G.H",
Tunnel-Server-Endpoint = :2:"I.J.K.L",
Tunnel-Server-Endpoint = :2:"M.N.O.P ",
Tunnel-Password = :2:"xxxx"
Where:
Customer_1 is the domain name of the MPLS VPN user sent to LNS by LAC; however, CAR has no
notion of domain names so for CAR Customer_1 is just another username in the database.
AAA
The following is the generic required AAA and RADIUS configuration.
aaa new-model
aaa authentication ppp default group radius
aaa authorization network default group radius
aaa accounting delay-start
aaa accounting update periodic xx
The required interval for sending interim accounting packets
aaa accounting network default start-stop group radius
RADIUS
VHG/PE
As previously discussed the VHG/PE router sends an Access-Request to the CAR. CAR then returns an
Access-Accept that completes the remote user's authentication and delivers, as part of network
authorization, an unnumbered loopback interface (that places the interface into the required VRF) and ip
pool or framed-ip-address. The VHG/PE then creates a unique virtual access interface using the ip
address of the unnumbered loopback interface, completes PPP NCP, allocates the remote user an IP
address from the local pool or the delivered framed-ip-address and terminates the PPP session into the
delivered VRF.
Customer VRF’s
The following is the generic required VRF configuration to define the customer VRF’s and Loopback
associated with VRF’s
ip vrf Customer_1
rd 10:10
route-target export 10:10
route-target import 10:10
!
ip vrf Customer_2
rd 20:20
route-target export 20:20
route-target import 20:20
!
interface Loopback10
description Customer_1
ip vrf forwarding Customer_1
ip address A.B.C.D 255.255.255.255
!
interface Loopback20
description Customer_2
ip vrf forwarding Customer_2
ip address A.B.C.D 255.255.255.255
Virtual Template
Configure the Virtual-Template that will be used as the generic clone block for creating a virtual access
interface for each dialin user and assigning them to the respective VRF.
interface Virtual-Template1
no ip address
ip verify unicast reverse-path
no snmp trap link-status
BGP
Generic configuration for BGP and address family.
address-family vpnv4
neighbor <ip address of bgp peer> activate
neighbor <ip address of bgp peer> next-hop-self
neighbor <ip address of bgp peer> send-community extended
exit-address-family
!
address-family ipv4 vrf Customer_1
redistribute connected
no auto-summary
no synchronization
exit-address-family
!
address-family ipv4 vrf Customer_2
redistribute connected
no auto-summary
no synchronization
exit-address-family
PPP
Below is a generic config for the overlapping IP address pools, this is defined on the VHG/PE to
allocate MPLS users IP addresses as they connect. The pool names are downloaded from RADIUS.
With the above configuration the pools the below is the sample output showing the overlapping
addresses.
It is recommended that the pool range is summarized as an aggregate route to be advertised by MP-
BGP.
In order to accommodate the VHG/PE Failover and Load Sharing scheme as discussed in the
NAS/LAC configuration template the below is an example showing the ip local pool configuraion for
two MPLS customers, Customer_1 & Customer_2 (assuming a /24 requirement). Both have Khi as
the primary VHG/PE with Lhe, Isb and Fsd in a failover group all with equal priority.
Khi-7206-VHG1
ip local pool pool1_Customer_1 10.0.0.2 10.0.0.254 group Customer_1
ip local pool pool1_Customer_2 11.0.0.2 11.0.0.254 group Customer_2
Lhe-7206-VHG1
ip local pool pool1_Customer_1 10.0.1.2 10.0.1.254 group Customer_1
ip local pool pool1_Customer_2 11.0.1.2 11.0.1.254 group Customer_2
Isb-7206-VHG1
ip local pool pool1_Customer_1 10.0.2.2 10.0.2.254 group Customer_1
ip local pool pool1_Customer_2 11.0.2.2 11.0.2.254 group Customer_2
Fsd-7206-VHG1
ip local pool pool1_Customer_1 10.0.3.2 10.0.3.254 group Customer_1
ip local pool pool1_Customer_2 11.0.3.2 11.0.3.254 group Customer_2
In this example one must consider, if Khi-7206-VHG1 fails and then Lhe-7206-VHG1 & Isb-7206-
VHG1 fail then Fsd-7206-VHG1 must be able to terminate all users for Customer_1 & Customer_2.
This configuration will allow Customer_1 or Customer_2 clients to dial into any NAS across Pakistan
and be connected to their private VPN. The actual IP Address allocated to a client will depend on the
VHG that terminates the L2TP tunnel; it will be the VHG closest to the NAS the user lands on
L2TP
The following is the generic required L2TP configuration; the per-user tunneling information will be
retried from the CAR.
vpdn enable
vpdn source-ip A.B.C.D
vpdn authen-before-forward
!
vpdn-group 1
AAA
The following is the generic required AAA and RADIUS configuration.
aaa new-model
aaa authentication ppp default group radius
aaa authorization network default group radius
aaa accounting delay-start
aaa accounting update periodic 60
aaa accounting network default start-stop group radius
RADIUS
ip radius source-interface Loopback <#> (Management Loopback)
radius-server host A.B.C.1 auth-port <#> acct-port <#> (For CAR_1)
radius-server host A.B.C.2 auth-port <#> acct-port <#> (For CAR_2)
radius-server source-ports extended
radius-server timeout 3
radius-server deadtime 10
radius-server key <key>
CAR Attribute
The attributes below are an example of those required on the CAR to assign the user to the correct VRF
and allocate the respective IP Pool or Framed IP address.
Cisco:Avpair = "ip:ip-unnumbered=Loopback10"
Cisco:Avpair = "ip:addr-pool=Customer_1" or Framed-IP-Address = A.B.C.D
IOS versions
Below are the IOS versions that will be deployed on the NAS and VHG devices.
The section details the internet connectivity, with upstream service provider(s). It details current plans of
Wateen Telecom for multihomed internet connectivity. The section also coveres internet as a customer
service offering and related infrastructure details.
It is understood that Wateen Telecom will like to peer with multiple internet service providers (and be
dual homed). Wateen Telecom has plans for two external peering points. Currently it is envisaged that
possible peering point will be located in Lahore. However it is possible that Wateen may expand the
number of external peering points beyond two, it is also possible that peering point location may be
extended to other cities or possibly countries. In addition to previous diagram depicting Lahore PoP,
following represents internet connectivity for internet border router (these devices are labeled internet
gatew router and the two terms can be used interchangeably).
The internet gateway routers are singly attached to respective PE routers due to SCE (service control
engine requirements). SCE requires symmetrical flow of traffic and hence cross connects are not used.
The Gigabit Ethernet links are on LAN based cards (67xx) on both gateways as well as PE devices.
From services perspective Wateen Telecom will like to provide transit internet connectivity to their
customers, this would mean it will be necessary to carry full internet routing tables in many parts of the
network.
iBGP will be used to propagate BGP learnt internet routes within Wateen Telecom’s network. iBGP has a
limitation that an iBGP peer never advertises an iBGP learnt route to another iBGP neighbour. As
previously discussed, this means creating a full iBGP mesh between all iBGP speakers. While this may
be possible for a very small number of nodes, this does not scale for any sizeable network. Therefore
IPv4 route reflectors are required to avoid a full iBGP mesh (please refer to see MP-BGP4 section for
details).
Possible options for relaxing this requirement are either using confederations or route reflectors.
Confederations require splitting the network into several autonomous systems (AS), route reflectors are
designed to be hierarchical and are a much better solution. Route Reflector approach is most widely
deployed approach.
While it is possible to provide BGP Next-Hop reachability via MPLS in the core, and hence eliminate the
need to run BGP in the core of the network (i.e. maintain a hollow core). As an alternative we can run
BGP in the core, assuming that core network is rather stable and there are no foreseeable plans to grow
the core.
For a hollow core, all the network elements within the core are not required to run BGP and carry full
internet routes. It becomes mandatory in this case to modify the next-hop at the border gateway router,
when peering with iBGP speakers.
For Wateen Telecom, both options are possible and administrative overhead of running BGP everwhere is
rather small. (P devices in this case will need to carry all interner routes). It is suggested that initially
BGP be used everywhere, and Wateen Telecom can eventually move to hollow core depending on growth
of the network and customer requirements.
As noted in the diagram above, remote (satellite) PE devices(7609) routers can be configured such that
full internet routes are not required and hence IPv4 configuration of BGP may be omitted for these
devices.
Granted that these PoP devices have scaled down version of Supervisor engine in 7609 series router
(Sup32 instead of Sup720-3BXL), it is suggested to make optimal use of resources by not carrying full
internet routing table in these devices. Instead it is suggested that
• A default route be statically configured pointing to upstream devices’s loopback address
• While it is possible to originate default route elsewhere in the network and distribute via IGP, in
satellite(remote) PoPs, there is little value to this approach as only single exit to upstream device
exists. This also assures that any unknown prefixes are only carried upto upstream PEs before being
discarded rather than being carried thoughout the core.
• Any customers in these PoPs be provided internet routing table by allowing them to peer directly
with an upstream PE (by using multihop knob). Disadvantage of this approach is that every time a
BGP session flaps, complete internet routing table will be sent across STM-1 link. However it is
expected that few customers will subscribe to this service in satellite PoPs. Wateen Telecom will
always have the option of transitioning these PoPs to follow same configuration as other city PoP
PEs.
• As indicated earlier, VPNv4 peering session will still be configured on satellite PoP PEs.
October 2, 2006 Wateen Telecom 188
Company Confidential. A printed copy of this document is considered uncontrolled.
Internet Services
• These PoP PEs will be configured identically to other PEs for IGP, LDP and any internet related
services.
For all other PE devices, please see MP-BGP4 (Multi-protocol BGP) section earlier in this design
document.
Following section replicates some information which is common for ipv4 iBGP session as well as eBGP
configuration (be it for upstream provider or downstream node) from MP-BGP4 section.
BGP Security
BGP by default does not use encryption for the sessions established, it is however recommended to
ensure that BGP is configured with neighbour passwords.
While the sessions may be internal and the requirement exists for both-ends of the session to be
configured to ensure BGP session establishment it is still recommended to use MD5 authentication on the
sessions for maximum security
BGP Passwords are disabled by default and are configured on a per-neighbour basis
It is also suggested that all external BGP sessions be secured with a unique password mutually agreed
with upstream provider or with a customer.
When the BGP session is established, it is told (on both routers) to implement a TTL of 255, the control
plane of the peering router then checks that the packet coming in has a TTL of 254 (decremented by one
hop).
In order to configure this the following configuration should be applied to eBGP neighbours
!
router bgp 65501
neighbor <IP address> ttl-security hops <hop count>
!
This command allows the user to configure the max numbers we receive from a peer or peer-group, when
the number of routes learnt exceeds the maximum number configured the default action is to terminate
the peering, the peering is then only re-established by manual intervention with a “clear ip bgp
neighbour” statement. We can use the keyword warning-only to generate a warning message to the
console and the network management platforms to indicate there has been a violation of the peering
policy.
The threshold value is set to indicate an initial warning that we are receiving a quantity of routes
approaching the pre-defined limit, this default is 75% this can be modified by inserting a integer value in
the <threshold> value with defining the maximum-prefix.
Wateen Telecom will use a maximum-prefix value of 200000 when peering with upstream BGP
neighbors for eBGP connectivity.
BGP MED
When there is no Multi-Exit Discriminator (MED) attribute set to a route, BGP IOS assumes a value of 0
for this route, in some circumstances other routing vendors have assumed absence of MED to mean a
value of 4294967295 (2^32-1), this difference caused possible potential to cause routing loops between
vendors with differing BGP implementations.
A belated IETF draft concluded that absence of a MED should cause it to have an infinite value (as
shown above), this would make a route without the MED attribute the least preferred, this is the inverse
generic behaviour in Cisco IOS where lack of MED indicates the best route available.
An additional command was introduced to ensure compatibility with other vendors using the IETF
implementation of MED.
!
router bgp 65501
bgp bestpath med missing-as-worst
!
This will cause IOS to act in the same way as described by the IETF, however this should not be
provisioned without due consideration.
If the possibilities for a prefix were received in this order without deterministic MED then the node would
install path 2 over path 1 due to a lower IGP metric, it would then prefer path 3 to path 2 due to the fact
that it is an external route.
However, with deterministic MED this removes any temporal dependencies of MED based decisions, it
ensures that the best MED comparison is made across all routes from the same AS, in this case path 1
would be chosen due to it having the lowest MED.
BGP Communities
BGP Communities are community tags that can be applied to a BGP route and carried throughout the
network of iBGP peerings.
These communities can identify a route or be used to apply policies applicable to only routes with certain
community tags.
It is proposed that Wateen Telecom use communities to classify the network routes:
• Lahore Subnets
• Karachi Subnets
• Islamabad Subnets
• Faisalabad Subnets
• National Subnets
• International Internet Subnets
Each edge node with eBGP peerings to customers connecting to the Wateen Telecom network would use
their own prefix so that it could be identified as coming from a certain PoP.
Table 20 Proposed BGP Communities
Lahore 65501:42001
Karachi 65501:42002
Islamabad 65501:42003
Faisalabad 65501:42004
National Prefixes 65501:42005
International 65501:42000
Prefixes
These BGP communities can then easily be checked with ACL’s to match Subnets from a major core
PoP, national or international traffic etc. A route can optionally be tagged with multiple communities if
necessary such as a regional and a national community, however care should then be taken when
configuring policies that are applied based on communities.
This community should be set outbound when peering from the Edge / Border router to the route-
reflectors,
October 2, 2006 Wateen Telecom 192
Company Confidential. A printed copy of this document is considered uncontrolled.
Internet Services
!
router bgp 65501
neighbour <RR IP Address> route-map setcommunity out
!
route-map setcommunity permit 10
Set community 65501:xxxx additive
!
Figure 115 BGP Community Tagging
Note, the keyword additive adds the community to any existing communities that exist on the route
already, without the additive keyword any existing communities would be overwritten by the single
community being set in the route-map.
To utilise the new style formatting where AA represents the AS number and NN is a 2 bytes community
value we need to configure the router.
!
ip bgp-community new-format
!
This will show the communities (as seen in a show running-config) in the new format, however data can
still be entered in any of the three supported formats.
BGP Fast-external-fallover
BGP fast-external-fallover is an option that may be disabled on eBGP peering routers (or on routers with
known unstable interfaces) where the Wateen Telecom network peers with outside service providers or
large clients with eBGP peerings. Typically with eBGP peers (as opposed to iBGP peers) the peering is
achieved using link addresses, rather than loopback addresses.
With fast-external-fallover, if the directly connected link for BGP peering fails then the session is
immediately reset. This then means that when the link recovers all the BGP routes have to be re-
advertised again causing CPU spikes and unnecessary potentially large BGP updates.
Fast-external-fallover is by default, enabled on Cisco routers, to disable this the syntax is:
!
router bgp 65501
no bgp fast-external-fallover
!
Usage of this knob (i.e. disabling fast-external-fallover) is dependent on the link stability and facilities
condition, in a stable environment the feature may be deployed to deal with a short hit on the link.
BGP Dampening
BGP route dampening is a feature that is designed to minimise the propagation of flapping routes
throughout the Internet.
If a link between differing ISPs flaps then the routes that that may be advertised will be withdrawn, this
has a rippling effect as these withdraws are propagated through the Internet, then during this propagation
the link returns and re-advertises the routes. When this process occurs several times in a short period of
time we have a number of routes that are de-advertised and then re-advertised, this propagated through
the Internet can cause routing instabilities and cause higher than normal CPU utilisation.
BGP route dampening uses a number of configurable options to enable dampening to a peer.
Table 21 BGP Route Dampening Parameters
BGP route dampening operates in the following fashion (as defined by RIPE229),
1. For each flap, a penalty (1000 points) is incrementally added to the routes advertised.
2. As soon as the penalty exceeds the suppress value the advertisement of the route will be
suppressed.
3. The penalty is then exponentially decayed using the pre-configured half-life value.
4. Once the value has fallen below the reuse value then the route is re-advertised.
The penalty is decayed at a granularity of 5 seconds and the entries are un-suppressed with a granularity
of 10 seconds.
Note that external routes that have been learnt through iBGP are not dampened (these are routes within
our own network and hence do not affect external peers and the Internet in general).
!
router bgp 65501
bgp dampening 15 750 2000 60
!
Figure 118 BGP Dampening Configuration
• Bgp dampening 30 2000 3000 60 – These are valid BGP dampening values, the maximum
suppress time is 60 minutes with a reuse-value of 2000, the maximum penalty is 8000
(2000*2)^(60/30) == 4000*2 == 8000. The example suppress limit of 3000 means that penalties
can easily exceed this limit. If the maximum penalty is incurred then the penalty will decay from
8000 to 4000 after one half-life, which is 30 minutes, and to 2000 after two half-lifes, which is
one hour.
• Bgp dampening 15 750 3000 45 – These are again valid BGP dampening values, the maximum
suppress time is 45 minutes, with a reuse-value of 750, the maximum penalty is 6000
(750*2)^(45/15) == 1500*2*2 == 6000, the suppress limit of 3000 is easily reachable and
penalties can exceed this value. If the maximum penalty is incurred then the penalty will decay
from 6000 to 3000 after one half-life (15 minutes)
• Bgp dampening 15 500 2500 30 – This is an example of an illegal set of dampening values that
will cause no routes to be dampened. The maximum suppress time is 30 minutes, so working the
equation out to reach the re-use limit of 500 we will need a route with a max penalty of 2000 to
be decayed through two half lives (2000 -> 1000 ->500) in order to be reused. However
considering we do not start suppressing routes until a penalty of 2500 then these are invalid
combinations.
This shows the importance of making the suppress limit LESS than the maximum possible value that the
route can achieve, otherwise no suppression will take place. It is important to note that the prefixes will
only be suppressed when a neighbour advertises them and it has a flap-history and a penalty value which
will cause them to be suppressed. A prefix will attract the 1000 penalty when the WITHDRAW is heard
from the neighbour eBGP peer. However because the prefix has been withdrawn it is not present in the
BGP table apart from the history entry. The penalty will start to decay, when the prefix is reannounced if
the penalty is still above the prefix limit the prefix will be suppressed, if the penalty has fallen below the
reuse limit it will not be suppressed, as an example:
• Bgp dampening 30 751 3000 60 – will generate a maximum penalty of 3004, which is only 4
higher than the suppress limit, the fastest the prefix would re-appear in the BGP table is one
minute later – when the BGP process runs, during that intervening time it would have attracted a
delay enough to have the penalty value below 3000 and not cause any dampening, if we had a
more realistic re-use limit of around 800 then the decay would be still high enough to cause the
prefix to be suppressed.
BGP Scaling
Peer Groups:
Update generation for BGP peers without the use of Peer Groups requires that the entire BGP table is
walked for every peer, the selected prefixes to be selected are then filtered through any outbound policies.
Then the updates are generated and sent to one peer. This process is repeated for each peer making this a
CPU intensive operation.
With peering groups the BGP table is walked only once, this is then sent to the peer-group leader after
applying all outbound filtering policies. These updates are then replicated for each peer-group member;
this ensures the peers are synchronised and receive the same information.
Peer-Group members are defined as being synchronised if all updates sent to the leader have also been
sent to the particular member. If the peer-groups are in a constant synchronised state then BGP can
replicate the update messages, this replication process is significantly faster then formatting an update
because this does not require a walk of the BGP table and the application of performing outbound routing
policies.
Peer groups may become out of synchronisation if the router in question is receiving slow TCP
throughput; differing CPU speeds or if the peer has a high CPU utilisation performing other tasks.
It is worth noting that the use of peer groups in a route-reflector design offers between 35% and 50% in
BGP scalability, allowing us to push more routes to more peers with less CPU utilisation in a shorter
period of time.
ip tcp path-mtu-discovery
Every TCP session has a limit in terms of how much data it can transport in a single packet. This limit is
defined as the Maximum Segment Size (MSS) and is 536 bytes by default. This means TCP will take all
of the data in a transmit queue and break it up into 536 byte chunks before passing packets down to the IP
layer. Using a MSS of 536 bytes ensures that the packet will not be fragmented before it gets to its
destination because most links have a MTU of at least 1500 bytes. The following command will allow
you to check the MSS of all BGP peers:
The problem is that using such a small MSS value creates a large amount of TCP/IP overhead, especially
when TCP has a lot of data to transport like it does with BGP. The solution is to dynamically determine
how large the MSS value can be without creating packets that will need to be fragmented. This is
accomplished by enabling "ip tcp path-mtu-discovery" (PMTU). PMTU allows TCP to determine the
smallest MTU size among all links between the ends of a TCP session. TCP will then use this MTU
value minus room for the IP and TCP headers, as the MSS for the session. If a TCP session only
traverses Ethernet segments then the MSS will be 1460 bytes. If it only traverses POS segments then the
MSS will be 4430 bytes. The increase in MSS from 536 to 1460 or 4430 bytes reduces TCP/IP overhead,
which helps BGP converge faster. Once we have enabled this knob and reset all of our BGP sessions via
a reboot or "clear ip bgp *" we can see the expected increase in MSS:
Router#show ip bgp neighbors | include max data
Datagrams (max data segment is 1460 bytes):
Datagrams (max data segment is 1460 bytes):
Datagrams (max data segment is 1460 bytes):
Datagrams (max data segment is 1460 bytes):
It is recommended to turn on TCP MTU discovery on all 12xxx and 7x00 network routers.
SPD Headroom
Traditionally the default SPD headroom has been 100 packets. In later releases this has been changed to
1000. Starting 12.0(22)S it is changed to 1000 and hence need not be changed. This should be confirmed
by typing sho spd or sho ip spd in privileged exec mode.
October 2, 2006 Wateen Telecom 196
Company Confidential. A printed copy of this document is considered uncontrolled.
Internet Services
!
interface pos 6/0
hold-queue 1500 in
!
Figure 119 Hold Queue Configuration
The Hold-queue refers to the buffer size on the GRP/Sup etc that is used to hold packets that have been
punted from the linecards, typically in the GSR’s distributed environment this is a relatively small
number of packets (IGP/BGP updates etc).
It is recommended to increase the interface hold-queue on the interfaces of all Routers.
BGP Timers
Several BGP timers can be used to tune the iBGP convergence in an IP MPLS / VPN Network.
• Advertisement-interval: To set the interval between the sending of BGP routing updates. The default
interval is 5 seconds for iBGP peers.
• Keepalive: Frequency, in seconds, with which the Cisco IOS software sends keepalive messages to its
peer. The default is 60 seconds.
• Holdtime: Interval, in seconds, after not receiving a keepalive message that the software declares a
peer dead. The default is 180 seconds.
• Scan-time: it defines the frequency at which the router validates routes in BGP table. It validates the
locally originated network and next-hop within the BGP table. Valid values used for selecting the
desired scanning interval are from 5 to 60 seconds. The default scan time is 60 seconds.
• Scan-time import: this timer is relevant for VPNv4 address-family. It defines the frequency at which
routes received from another PE through MP-BGP are imported into a VRF. This timer is not
relevant for next-hop deletion that is processed based on the “normal” BGP scan-time. Default value
is 15 seconds.
The following table summarizes the default values for the iBGP timers.
• iBGP convergence follow IGP convergence, This means that when a PE router or its uplinks fail,
BGP next-hop reachability will be lost after the IGP converges. The BGP scanning process will
detect the loss of the next-hop for all relevant BGP and MP-BPG routes and stop selecting it in
the BGP RIB.
• It is recommended initially to leave the timers at their default.
• In current release on 7600, BFD is not available for BGP BFD will be really relevant for eBGP
sessions and is not deemed significant for Wateen Telecom at this stage.
As discussed in the IGP routing section, the IGP convergence has already been tuned to provide fast
convergence under the event of an internal link or node failure.
This means that when a PE router or it’s uplinks fail, BGP next-hop reachability will be lost after the IGP
converges. The BGP scanning process will detect the loss of the next-hop for all relevant BGP and MP-
BPG routes and stop selecting it in the BGP RIB.
It is recommended initially to leave the timers at their default, unless there is a specific convergence
requirement for BGP.
Table below shows potentially tuned BGP tuned times, these should be modified with care and only after
ensuring stability of the network infrastructure
Table 23 - Potential Modified values for iBGP timers
BGP Configuration
• The BGP AS# for Wateen Telecom will be 65501 (interim AS number as indicated by the customer)
• iBGP Neighbour addresses will always be to the public loopback addresses
NAT Service
For end user internet access (i.e.customers not doing BGP peering), it is expected that most of the
internet users will be using rfc1918 private ip address space. These users will be NATed (doing Port
Address Translaton since N to 1 mapping) to an interface address using following CLI:
ip nat inside source list 1 interface pos0/0 overload
int pos0/0
ip nat outside
int vlan10
ip nat inside
It is suggested to use a separate loopback address for NATing all the internet traffic, hence keeping a
clear distinction with infrastructure traffic vs user internet traffic. However in the interest of conserving ip
address space it is possible to NAT the users to one of the egress interfaces of PE device.
Please note that with above configuration, all CPE devices which are getting NATed by the PE device,
will no longer be reachable (as they are being NATed) without access to PE device first. Should this be a
concern, a static NAT mapping can be done per device, such that various devices can be reached directly.
This may be required for a NMS application managing CPE devices which are only in the global routing
table. i.e. a Customer’s CPE which has only subscribed to internet service.
Following example would allow telnet access to NATed hosts at different ports of public ip address:
For Wateen Telecom, another alternative to this approach is to bypass NAT for certain source and
destination prefiexes using extended ACLs (access-lists) while performing NAT. This approach is
simpler and recommended, as long as both prefixes are reachable within Wateen autonomous system via
IGP or BGP.
Ingress Access-Lists
For customer facing ports, it is suggested to deploy access-lists filtering certain traffic which can
commonly be exploited. e.g. permitting all traffic except icmp traffic etc.
On the Cisco 7600 router uRPF processing is done in hardware, resulting in no performance impact by
enabling this feature.
If uRPF is configured in combination with ACLs, the ACLs are checked before and after the reverse path
validation as explained in the following:
Step 1 - Check input ACLs configured on the inbound interface.
Step 2 - uRPF validates the packet has a return path through the inbound interface by checking the CEF
table (CEF or dCEF).
Step 3 - CEF table (FIB) lookup is carried out for packet forwarding.
Step 4 - Output ACLs are checked on outbound interface.
Step 5 - Packet is forwarded.
Unicast RPF’s advantage when used for IP source address validation is that it dynamically adapts to
changes in the routing tables, including static routes. Additionally it requires less operational maintenance
than traditional approaches which use IP access or extended-access lists
October 2, 2006 Wateen Telecom 200
Company Confidential. A printed copy of this document is considered uncontrolled.
Internet Services
The rx keyword enables uRPF in strict mode, the any keyword enables uRPF in loose mode. It is
important to add the “allow-self-ping” at the end of this command. If allow-self-ping is not present, the
uRPF will filter out “self-ping” packets from the router, which can be a useful troubleshooting tool.
In the Wateen Telecom network uRPF should be used on all edge interfaces.
There will be two types of external access connections, single home and dual homed links.
• Single homed connection. In this scenario, a “simple” connection exists from SP core to the customer
edge i.e. traffic will always be synchronous to and from this customer. This is due to the fact that the
customer only has one uplink, and no “dual-homing” is present. This is expected to be majority of
Wateen Telecom’s customer access scenario.
• Multi homed connections: In this scenario, because of redundant links we can not ensure symmetric
routing. All Internet connections are classified as connections where traffic flow may be asymmetric,
as well as all customer connections where the customer has multiple connections to one or multiple
service providers .
For single homed connections, the following command should be applied on the customer facing
interface:
ip verify unicast source reachable-via rx allow-self-ping
For multi homed connections, the following command should be applied on the customer facing
interface:
ip verify unicast source reachable-via any allow-self-ping
Inbound traffic:
In the inbound direction (traffic coming from the Internet towards the Wateen Telecom network) an ACL
should be deployed that stops traffic with an source IP address that is a RFC1918 or RFC 3330 address,
or that have a Wateen Telecom owned subnet as the source address. For more details on Cisco
recommendation on Infrastructure ACL, please refer to public white paper at
http://www.cisco.com/en/US/tech/tk648/tk361/technologies_white_paper09186a00801a1a55.shtml
Outbound traffic:
For outbound traffic towards the Internet ACL’s should be configured to stop known “non-valid” source
addresses, like RFC1918 and RFC 3330 address.
It is recommended that totally stubby area be numbered as area 1. Following configuration will be
required on both PEs and 6513 DC switches for area 1 configuration.
router ospf 1
network <network address> <wild-card mask> area 1
area 1 stub no-summary
!
The “no-summary” keyword is only required at ABR i.e. Lhe-76130-PE1&2. For further details on Data
Center architecture and related services please refer to relevant LLD document.
interface vlan10
ip address <interface ip address>
ip helper-address <dhcp (CNR) server-1>
ip helper-address <dhcp (CNR) server-2>
!
By allowing summarized prefixes into IGP, individual subnets are no longer required to be advertised in
area 0. This approach would allow IGP to scale while providing optimal reachability.
Same objective can be accomplished via advertising the summary into BGP. Care must be taken such that
these RFC1918 addresses are not advertised externally. In order to achieve this objective, these summary
prefixes can be tagged with a no-export community string. IF BGP was used to carry these prefixes,
following configuration will be required.
ip route 194.168.0.0 255.255.0.0 Null0
router bgp 65501
<standard BGP configuration for a PE>
address-family ipv4
!configure community to be sent to BGP Peers
neighbor refip send-community both
neighbor refip route-map setcom out
neighbor 10.100.0.1 activate
neighbor 10.100.0.4 activate
no auto-summary
no synchronization
network 194.168.0.0 mask 255.255.0.0
exit-address-family
route-map setcom permit 10
match ip address 112
set community no-export
access-list 112 remark rfc1918 adv via BGP with community set
access-list 112 permit ip 194.168.0.0 0.0.255.255 255.255.0.0 0.0.255.255
Once the Customer Edge (CE) device has obtained an ip address via DHCP, any traffic destined for
internet will be NATed by respective 7600 PE device.
In this case it is anticipated that each prefix for a customer facing network will span both PE devices. i.e.
same vlan/prefix (e.g. vlan 10 will terminate on PE-1 and PE-2, prefix assigned to vlan10 e.g.192.168.1.0
will appear on both PEs). It is understood that layer-2 connectivity will be provided by access switch (to
avoid any spanning tree issues). In this case, both PEs will advertise static summaries into area 0. This is
in line with other PEs (which will be advertising static summaries). Since it is possible that one of the PEs
may loose an access link and continue to advertise a summary, a separate routing process (e.g. RIP) will
be configured between both PEs over a dedicated link (or a dedicated vlan). This link will advertise
specific prefixes/subnets to its neighbor so as to avoid blackholing any traffic.
An alternative to this approach is to define a separate area between two PEs, with a dedicated link
between the two PEs. This way both PEs will become ABR (area border routers) and OSPF can advertise
summaries from each PE for non backbone area into area 0. ABR will summarise their respective
prefixes using the router ospf command: “area area-id range address mask”. By using this
summarization technique the discard route (summary pointing to null0) is automatically generated by
default by OSPF process. Either approach will work for ipv4 traffic. Based on the information available
to date for dual attached PE scenario, later approach is suggested.
Introduction
In order to fulfill Wateen Telecom requirements of having 5 distinct classes of service, each with their
specific service characteristics, QoS mechanisms will be deployed in the core of the network. Wateen
Telecom has requested for following classes:
• Real Time Traffic (LLQ, voice)
• Video Traffic (potential future requirement)
• Management Traffic (snmp, management, signaling)
• Business Data (Premium Traffic)
• Best Effort (Internet Traffic)
It is to be noted that indeed, the CE to PE link is the most bandwidth constrained and as a result,
congestion will occur most likely either on the CE (egress to the PE ) or on the PE (egress to the CE).
It is understood that WiMAX, WiMAX aggregaton and Microwave infrastructure can provide equivalent
five classes of traffic with appropriate prioritization to keep end to end Qos commitment and SLA to the
end customer.
Characteristics of each class are given below:
• Real-Time (VoIP) traffic. A minimum bandwidth together with minimum loss, delay and jitter needs
to be guaranteed. It is advised that the VPN customer implements some kind of Call Admission
Control method for their Voice traffic. This is required because, in the event that a VPN customer
exceeds the amount of Voice traffic that has been allocated in their Voice Class of Service, Voice
traffic could be dropped in case of congestion, irrespective of to which voice call this traffic belongs.
This would then potentially lead to degradation in Voice quality for all voice calls.
• Video Traffic (potential future requirement) : Streaming Video is considered a class with minimum
bandwith / service gaurntee. Streaming video applications have more lenient QoS requirements
because they are delay-insensitive and are largely jitter-insensitive (because of applicationbuffering).
• Business/Premium traffic. This is mission critical traffic for which the VPN customer requires
preferential treatment. A minimum bandwidth needs to be guaranteed to see SLA.
• Manage\ment Class: This is a class carrying all inband management traffic and is protected with
minum bandwidith gaurntee.
• Best Effort traffic. This traffic requires no special of preferential treatment. This would typically be
E-mail and Internet type traffic. A (low) minimum bandwidth needs to be guaranteed.
Note: typically QOS is described as IP Precedence, however in MPLS this maps to MPLS EXP bits
as this is how the QOS is delivered in an MPLS environment, IP Precedence is or can be maintained
throughout the network even though the SP can modify the MPLS EXP bits to suit their internal
QOS policies.
QOS Disclaimer
Since no clear requirements were given by Wateen Telecom in terms of expected traffic patterns, or the
strategy with regards to what type of QoS packaging that should be offered, what qos guarntees should be
made for each class, amount of expected link utilization etc. The QoS section in this document will
contain the following sections
• Generic QoS recommendation
• QoS architectures for the Cisco Platforms deployed in Wateen IP/MPLS core
• Generic configuration templates for the Cisco devices for QoS
These assumptions will be applied on all CORE Links, CORE-DIST, DIST-DIST and DIST-ACCESS
links.
This section is intended as an introduction to the Differentiated Services (DiffServ) reference model.
DiffServ is a new model by which traffic is treated by intermediate systems with relative priorities based
on the type of services (ToS) or Differentiated Services Code Point (DSCP) field. Defined in RFC’s 2474
and 2475, the DiffServ standard supersedes the original specification for defining packet priority
described in RFC 791.
The new DiffServ standard proposes a new way of interpreting a field that has always been part of an IP
packet. In the DiffServ standard, the ToS field will be renamed to Differentiated Services Code Point
(DSCP) and will have new meaning. The DiffServ standard proposes to increase the number of definable
priority levels by re-allocating bits of an IP packet for priority marking.
The ToS field describes one entire byte (eight bits) of an IP packet. Precedence refers to the three most
significant bits of the ToS field---that is, [XXX]XXXXX. There may be some confusion because people
occasionally use the term ToS to refer to the next three bits---XXX[XXX]XX. To be consistent with the
original RFC 791 specification, this document uses the term ToS to refer to the entire eight bits.
The three most significant bits of the ToS field---the precedence bits---define the IP packet priority under
RFC 791. The precedence portion of the ToS field---[XXX]XXXXX---defines the priority or importance
of the field. The following diagram and tables illustrate the old method used to interpret this eight-bit
portion of the IP packet.
0-2 3 4 5 6 7
Precedence Delay Throughput Reliability Reserved = 0 Reserved = 0
This one-byte ToS field has been almost completely unused since it was proposed almost 20 years ago.
Only in the last few years have Cisco and other router companies begun utilising the Precedence bits for
making forwarding decisions.
The DiffServ standard follows a similar scheme to RFC 791, but utilises more bits for setting priority.
The new standard maintains backward compatibility with RFC 791 implementations, but allows more
efficient use of bits 3, 4, and 5. (Bits 6 and 7 will still be reserved for future development.) With the
additional 3 bits, there are now a total of 63 classes instead of the previous 7 classes.
RFC 2475 defines Per Hop Behaviour (PHB) as the externally observable forwarding behaviour applied
at a DiffServ-compliant node to a DiffServ Behaviour Aggregate (BA).
With the ability of the system to mark packets according to DSCP setting, collections of packets with the
same DSCP setting and sent in a particular direction can be grouped into a BA. Packets from multiple
sources or applications can belong to the same BA.
In other words, a PHB refers to the packet scheduling, queuing, policing, or shaping behaviour of a node
on any given packet belonging to a BA, as configured by a service level agreement (SLA) or a policy
map.
Default PHB
The default PHB essentially specifies that a packet marked with a DSCP value of 000000 (recommended)
receives the traditional best-effort service from a DS-compliant node (that is, a network node that
complies with all of the core DiffServ requirements). Also, if a packet arrives at a DS-compliant node,
and the DSCP value is not mapped to any other PHB, the packet will get mapped to the default PHB.
For more information about default PHB, refer to RFC 2474, Definition of the Differentiated Services
Field in IPv4 and IPv6 Headers.
Class-Selector PHB:
To preserve backward-compatibility with any IP Precedence scheme currently in use on the network,
DiffServ has defined a DSCP value in the form xxx000, where x is either 0 or 1. These DSCP values are
called Class-Selector Code Points. (The DSCP value for a packet with default PHB 000000 is also called
the Class-Selector Code Point.)
The PHB associated with a Class-Selector Code Point is a Class-Selector PHB. These Class-Selector
PHBs retain most of the forwarding behaviour as nodes that implement IP Precedence-based
classification and forwarding.
For example, packets with a DSCP value of 110000 (the equivalent of the IP Precedence-based value of
110) have preferential forwarding treatment (for scheduling, queuing, and so on), as compared to packets
with a DSCP value of 100000 (the equivalent of the IP Precedence-based value of 100). These Class-
Selector PHBs ensure that DS-compliant nodes can coexist with IP Precedence-based nodes.
The DiffServ standard utilizes the same precedence bits (the most significant bits: 0, 1, and 2) for priority
setting, but further clarifies their functions/definitions, plus offers finer priority granularity through use of
the next three bits in the ToS field. DiffServ reorganizes (and renames) the precedence levels (still
defined by the three most significant bits of the ToS field) into the following categories:
Precedence 5 Class 5
Precedence 4 Class 4
Precedence 3 Class 3
Precedence 2 Class 2
Precedence 1 Class 1
For more information about class-selector PHB, refer to RFC 2474, Definition of the Differentiated
Services Field in IPv4 and IPv6 Headers.
For example, network traffic can be divided into the following classes:
• Gold: Traffic in this category is allocated 35 percent of the available bandwidth.
• Silver: Traffic in this category is allocated 40 percent of the available bandwidth.
• Bronze: Traffic in this category is allocated 25 percent of the available bandwidth.
Further, the AFxy PHB defines four AF classes: AF1, AF2, AF3, and AF4. Each class is assigned a
specific amount of buffer space and interface bandwidth, according to the SLA with the service provider
or policy map.
Within each AF class, you can specify three drop precedence (dP) values: 1, 2, and 3. Assured
Forwarding PHB can be expressed as shown in the following example:
AFxy
In this example, x represents the AF class number (1, 2, or 3) and y represents the dP value (1, 2, or 3)
within the AFx class.
In instances of network traffic congestion, if packets in a particular AF class (for example, AF1) need to
be dropped, packets in the AF1 class will be dropped according to the following guideline:
dP(AFxy) <= dP(AFxy) <= dP(AFxy)
where dP (AFxy) is the probability that packets of the AFxy class will be dropped. In other words, y
denotes the dP within an Afx class.
In the following example, packets in the AF13 class will be dropped before packets in the AF12 class,
which in turn will be dropped before packets in the AF11 class:
dP(AF13) <= dP (AF12) <= dP(AF11)
The dP method penalises traffic flows within a particular BA that exceed the assigned bandwidth. Packets
on these offending flows could be re-marked by a policer to a higher drop precedence.
An AFx class can be denoted by the DSCP value, xyzab0, where xyz can be 001, 010, 011, or 100, and ab
represents the dP value.
Bits 3 and 4 of DiffServ field allow further priority granularity through the specification of a packet drop
probability for any of the defined classes. Collectively, Classes 1-4 are referred to as Assured Forwarding
(AF). The following table illustrates the DSCP coding for specifying the priority level (class) plus the
drop percentage. (Bits 0, 1, and 2 define the class; bits 3 and 4 specify the drop percentage; bit 5 is always
0.)
Using this system, a device would first prioritize traffic by class, then differentiate and prioritize same-
class traffic by considering the drop percentage. It is important to note that this standard has not specified
a precise definition of "low," "medium," and "high" drop percentages. Additionally, not all devices will
recognize the DiffServ bit 3 and 4 settings. Remember also that even when the settings are recognized,
they do not necessarily trigger the same forwarding action to be taken by each type of device on the
network---each device will implement its own response in relation to the packet priorities it detects. The
DiffServ standard is meant to allow a finer granularity of priority setting for the applications and devices
that can make use of it, but it does not specify interpretation (that is, action to be taken).
EF PHB is ideally suited for applications such as VoIP that require low bandwidth, guaranteed
bandwidth, low delay, and low jitter.
For more information about EF PHB, refer to RFC 2598, An Expedited Forwarding PHB.
It is Cisco’s experience that a Quality of Service design and deployment is never a straightforward
process – after an initial deployment, a performance assessment phase and subsequent tuning of the QoS
deployment is a necessity. Therefore, we strongly recommend a tuning phase while beta customers are
connected.
The following table and figure gives an overview of the various QoS mechanisms that can be used on the
CE (for traffic flowing from CE to PE) and on the PE (for traffic flowing from PE to CE).
Best Effort DSCP 0 0 CE: ACL CE: class- CE: LLQ WRED
PE: precedence based BEST_EFFORT
PE: MDRR – queue 0
Routing DSCP 48 6 CE: ACL BGP CE: LLQ RP
Protocol PE: precedence precedence Min (8 Kbps, 1% of
6 access)
PE: MDRR – queue 3
Table 28 - CE to PE QoS mechanisms overview
Bottleneck Bottleneck
The following table and figure gives an overview of the various QoS mechanisms that can be used on the
PE and P routers, for traffic flowing between PE and P routers, and P to P routers.
Bottleneck Bottleneck
PE-Ingress P PE-Egress
classification classification classification classification
queuing queuing queuing queuing
congestion avoidance congestion avoidance congestion avoidance congestion avoidance
QoS Tools
There are several types of tools that are used to provide and deploy a QoS enabled network, be it local to
a network element or across the network:
• Classification and Marking
• Congestion Avoidance
• Congestion Management
• Traffic Conditioning
• Signaling
• Link Efficiency Mechanisms
The following diagram provides a visual representation of various QoS functions that are performed in a
network.
.
Classification
On the CE, packets will be classified through the use of extended access lists (ACLs). These ACLs can
match packets on IP Source or Destination address, protocol type, and UDP/TCP port numbers.
Dependent on the actual signalling method used (packet sizes), speed of the access links and the number
of concurrent voice call set-ups that need to be supported, 2 possible design options can be taken with
regards to the queuing method used.
• Queue the voice signalling packets in the same PQ as the actual voice bearer packets. This will result
in a simpler design but could delay the transmission of some of the voice bearer packets (dependent
on voice signalling packet size, access link speed and number of concurrent voice call set-ups). This
could than have an impact on the voice delay / jitter.
• Queue the voice signalling packets in another normal class queue. This should ideally be a separate
class queue from the ones that are used for regular data traffic to ensure delivery of the voice
signalling packets. This will result in a more complicated design where bandwidth needs to be
allocated for the voice signalling class. Also, voice signalling packets might be delayed through the
network resulting in a delay in the voice call set-up process. The advantage is that the actual voice
quality will not be impacted as no voice signalling packets will travel in the PQ.
Testing has indicated that, without cRTP (Compressed Real Time Protocol) enabled, the effect of
mapping VoIP signalling packets together with the VoIP bearer packets in the same priority queue is
negligible. The signalling packets have little effect on the latency nor do they cause any drops due to the
default bust size of 200ms that has been built into the priority queue. Therefore, the design
recommendation is to match the VoIP signalling packets with ACL <x> and queue them together with the
VoIP bearer packets in the priority queue.
It should however be understood that VoIP signalling implementations differ and that some might have a
negative effect on the performance of the priority queue. In that event, the VoIP signalling traffic needs to
be mapped in another class queue (Business, Management class for example).
The classified traffic will subsequently be mapped in their respective classes. A maximum of 1024
classes can be defined on a single router, 256 classes per policy map and 256 policy maps maximum not
to exceed the 1024 classes.
For example:
Note: the above samples would need to be modified to include both real ACL’s that reflect real business
traffic along with real class-maps that define the classes within the network. Following lists all possible
options in a class-map on a 7600 series router:
7606-8c(config)#class-map test
7606-8c(config-cmap)#match ?
access-group Access group
any Any packets
atm Match on ATM info
bgp-index BGP traffic index value
class-map Class map
cos IEEE 802.1Q/ISL class of service/user priority values
destination-address Destination address
discard-class Discard behavior identifier
dscp Match DSCP in IP(v4) and IPv6 packets
fr-de Match on Frame-relay DE bit
fr-dlci Match on fr-dlci
input input attachment circuits
input-interface Select an input interface to match
ip IP specific values
mpls Multi Protocol Label Switching specific values
not Negate this match result
packet Layer 3 Packet length
precedence Match Precedence in IP(v4) and IPv6 packets
protocol Protocol
qos-group Qos-group
source-address Source address
vlan VLANs to match
Marking
After classification, packets need to be marked with their appropriate IP precedence or DSCP value. The
following marking mechanism will be used here.
• Class Based Marking. This is the currently recommended method of marking or colouring traffic.
Class Based Marking is part of the Modular QoS Command Line Interface (or MQC) method of
configuring QoS parameters. Class Based Marking will be used for the Best Effort, Real-Time, and
Business classes of traffic. The BGP traffic will be marked with precedence 6 by the router.
The following is the required configuration for Class Based Marking of the Best Effort traffic. Best Effort
traffic will be coloured with a DSCP value of 0.
!
Policy-map customer_profile
class best_effort
set ip prec 0
As depicted in then following figure, when a Priority Queuing class is configured, the PQ has direct
access to the transmit (TX) ring. This is, of course, unless interleaving is configured, in which case
interleaving occurs prior to placing the PQ traffic onto the TX-ring.
For example, a G.729A codec at 50 pps would generate 25.6Kbps. 5% should be added to this to include
the VoIP signalling overhead (RTCP / H.225 / H.245). As discussed before, the VoIP signalling traffic
will be mapped together with the VoIP bearer traffic in the priority queue.
Call admission control is another important issue that needs to be considered. Call admission control is a
mechanism for ensuring that voice flows do not exceed the maximum provisioned bandwidth allocated
for voice conversations. After doing the calculations to provision the network with the required
bandwidth to support voice, data, and possibly video applications, it is important to ensure that voice does
not oversubscribe the portion of the bandwidth allocated to it. While most QoS mechanisms are used to
protect voice from data, call admission control is used to protect voice from voice. This is illustrated in
the following figure, which shows an environment where the network has been provisioned to support
two concurrent voice calls. If a third voice call is allowed to proceed, the quality of all three calls is
degraded. Call admission control should be external to the network.
After defining the service policy in a policy-map, it needs to be applied on an interface (service-policy).
By default, on the non-distributed router platforms, the sum of the minimum bandwidths needs to be
lower than 75 % of the configured access bandwidth. Since the actual configured sum of minimum
bandwidths will probably be larger (Real-Time + Business + Best Effort + 8 Kbps management + 8 Kbps
routing protocol), this default parameter setting can be changed (maximum-reserved-bandwidth) to 100
%. However, it is also a very good design practice not to push the design boundaries to the edge without
allowing for any margin of error or unexpected traffic patterns. Therefore, it is still recommended to keep
the sum of all minimum bandwidths below 100 %. Keeping the sum of all minimum bandwidths around
90 % will allow for unaccounted traffic such as layer 2 overhead, layer 2 keepalives, LMI (in the case of
Frame Relay), etc.
The following is the required configuration for LLQ class queuing. The order in which the classes are
configured under the policy-map is important as traffic will be queued in the first queue that matches the
defined ACL.
!
!
policy-map CUSTOMER_PROFILE
class REAL_TIME
set ip prec 5
priority <bps>
class BUSINESS
set ip prec 3
bandwidth <bps>
class BEST_EFFORT
set ip prec 0
bandwidth <bps>
!
interface Serial0/1
bandwidth 2048
encapsulation ppp
max-reserved-bandwidth 90
service-policy input CUSTOMER_PROFILE
When congestion occurs in the class queues, packets will be dropped. By default, tail dropping (First In
First Out – FIFO) will occur. Tail dropping is in general not a desired behaviour. TCP stacks react to
multiple packet drops by severely reducing their window sizes, thus causing severe interruption to higher
layer applications. In addition, multiple packet drops can lead to a phenomenon called “global
synchronisation” whereby TCP stacks become globally synchronised resulting in a wave-like traffic flow.
This is also not desirable.
Therefore, Weighted Random Early Detection (WRED) will be implemented on the Business and Best
Effort traffic classes. FIFO queuing will still be applied for the other classes.
WRED is a congestion avoidance and control mechanism whereby packets will be randomly dropped
when the average class queue depth reaches a certain minimum threshold (min-threshold). As congestion
increases, packets will be randomly dropped (and with a rising drop probability) until a second threshold
(max-threshold) where packets will be dropped with a drop probability equal to the mark-probability-
denominator. Above max-threshold, packets are tail-dropped.
Figure 142 depicts the WRED algorithm.
WRED
WREDValues
Values
min = 20
min = 20
max
max==5050
prod = 1
prod = 1
Packet Drop
Probability
WRED Values
WRED Values
min = 55
min = 55
max
max==7070
prod
prod==1010
Average
Queue Size
Min1 Max1 Min2 Max2
WRED will selectively instruct TCP stacks to back-off by dropping packets. Obviously, WRED has no
influence on UDP based applications (besides the fact that their packets will be dropped equally).
The average queue depth is calculated using the following formula:
• new_average = (old_average * (1-2-e) + (current_queue_depth * 2-e)
The “e” is the “exponential weighting constant”. The larger this constant, the slower the WRED
algorithm will react. The smaller this constant, the faster the WRED algorithm will react.
The exponential weighting constant can be set on a per-class basis. The min-threshold, max-threshold and
mark probability denominator can be set on a per precedence or per DSCP basis.
The mark probability denominator should always be set to 1 (100 % drop probability at max-threshold).
WRED will be applied on the Business and Best Effort traffic classes.
Again, it should be stressed that tuning QoS parameters is never a straightforward process and the results
are depending on a large number of factors, including the offered traffic load and profile, the ratio of load
to available capacity, the behavior of end-system TCP stacks in the event of packet drops, etc.
Therefore, it is strongly recommended to test these settings in a testbed environment using expected
customer traffic profiles and to tune them, if required. In addition, after an initial production beta
deployment, a performance assessment phase and subsequent tuning of the QoS deployment is
recommended.
Table 32 summarises the recommendation for setting the exponential weighting constant (e). B equals the
link bandwidth in MTU sized packets (MTU = 1500 bytes).
• The reference bandwidth should be the link speed for the Best Effort traffic class.
• The reference bandwidth should be the class bandwidth for the Business traffic class. This will result
in a better quality Business class.
32 3 3 3
64 6 6 3
128 11 11 3
256 22 22 4
512 43 43 5
1024 86 86 6
2048 171 171 7
34684 2891 289.1 8
45210 3767 376.7 9
100000 8334 833.4 10
155000 12917 1291.7 10
622000 51834 5183.4 12
2400000 200000 20000 14
Table 32 - WRED - exponential weighting constant
These tables summarize the recommendation for setting the min-threshold and max-threshold for the Best
Effort and Business traffic classes. The required values without having out-of-contract behavior are
“minTHin” and “maxTHin”.
• The reference bandwidth should be the link speed for the Best Effort traffic class.
• The reference bandwidth should be the class bandwidth for the Business traffic class.
Link
Speed minTH maxTH minTH maxTH
in Mbps B in in out out
34684 2891 87 290 29 87
45210 3767 110 367 37 110
100000 8334 251 834 84 251
155000 12917 388 1292 130 388
622000 51834 1556 5184 519 1556
2400000 200000 6000 20000 2000 6000
!
policy-map CUSTOMER_PROFILE
class REAL_TIME
set ip dscp ef
priority <bps>
class BUSINESS
set ip dscp af 41
bandwidth <bps>
random-detect dscp-based
random-detect exponential-weighting-constant <e>
random-detect dscp 34 <minTH> <maxTH> 1
class MANAGEMENT
set ip dscp 1
bandwidth 8
queue-limit 40
class RP
bandwidth 8
queue-limit 40
class BEST_EFFORT
set ip dscp 0
bandwidth <bps>
random-detect dscp-based
random-detect exponential-weighting-constant <e>
random-detect dscp 0 <minTH> <maxTH> 1
!
interface Serial0/1
bandwidth 2048
encapsulation ppp
max-reserved-bandwidth 100
service-policy output CUSTOMER_PROFILE
Deficit Round Robin is a scheduling algorithm that provides relative amounts of bandwidth to the
different classes. Relative means that there is no hard limit but the amount is limited through “weights“ in
relation to the other configured and active queues. For a simple example we assume we have three
classes, weighted 1, 1 and 2. If there is heavy congestion and traffic present in all three queues the
proportions are 25%, 25% and 50% of the bandwidth. However if there is no traffic in the first class and
still oversubscription caused by the remaining two, the percentages are 33% and 66%, as the active
weights are 1and 2. This shows that the algorithm always distributes the bandwidth in a very efficient
way while maintaining fairness between the classes.
The actual implementation on the GSR is different from the simple example above. The basic principle of
operation is that the scheduling algorithm cycles through the configured classes and sends a certain
amount of traffic. How much is determined in the following way: Each queue has a deficit counter in
bytes, which initially is set to zero. When the queue is processed the counter is increased by the
Maximum Transmission Unit (MTU)+512*(weight-1). Packets are dequeued from the respective queue
and for every packet its number of bytes is subtracted from the deficit counter. If the counter becomes
zero or less, the queue has used up its amount and the next queue is serviced. If there are no more packets
in the queue the scheduler also moves to the next queue. If the deficit counter is negative it keeps its value
for the next cycle, and the amount of MTU+512*(weight-1) is added to it. This leads to slightly less data
that can be sent, keeping the amount of traffic sent over multiple cycles within the defined range. If the
deficit counter is greater than zero (not all of it has been used up), it will be reset to zero. This means that
a class cannot accumulate quotas in advance.
Some of the GSR line cards support a slightly modified DRR algorithm that provide for a specific low-
latency queue. This queue can run in one of two modes: strict priority or alternate priority. In strict
priority mode the low-latency queue is serviced as soon as packets arrive in it and it is serviced until it is
empty. This gives very low latency at the expense of possible unfairness towards the other queues.
Alternate priority works differently in that it services the low-latency queue every other time between the
regular queues. The low-latency queue has a weight assigned to it and can only send a certain amount of
traffic, however because the time between its turns is much shorter it will provide low latency.
The weight is the only variable parameter in DRR. It influences two things: The relative amount of
bandwidth a traffic class can use and how much traffic is sent in one turn. Using larger weights means
that the overall cycle takes longer and possibly increases latency. Also it needs to be noted that the low
latency queue in alternate priority mode is serviced more than once per cycle and thus takes more
bandwidth than other queues with the same nominal weight. How much more is a function of how many
queues are defined, e.g. with three queues total the low latency queue is serviced twice as often as the
other queues and sends twice its weight per cycle. If eight queues are defined the low latency queue is
serviced seven times more often and the “effective” weight is seven times higher.
A specific low latency queue is a key component to guarantee low delay and jitter across the backbone
network. Backbone routers have large packet buffers to accommodate bursts at high speeds, those buffers
can add tens of milliseconds of delay. Routers that do not have low latency queuing mechanisms are not
able to provide the necessary quality of service for some real-time applications.
transmission through the fabric as well as for packets queued on outgoing interfaces. The following two
diagrams show where CAR, WRED and DRR features are used in the GSR architecture.
On the receiving or ingress line card queuing takes place on a per destination line card basis, for every
possible destination line card in the system up to 8 queues for 8 Classes of Service can be defined.
On the transmit side a set of 8 output queues is available on every port. Please note that availability of features
differs across the different types of line cards for the GSR. Also note the difference between DRR and MDRR,
the former uses strict priority for the low-latency queue while the latter has the additional option of using
alternate priority.
It is obvious that with the current network topology and installed capacity, the crossbar switch capacity
will not become a bottleneck, therefore tofab queuing is not covered.
Typically, there will be 3 separate MDRR queues on every egress POS linecard (FrFab) within the core
network. Core network links are between PE, P. Core link speeds are:
• OC-48 (STM-16) on the Engine 5 line card (12816 Core Router)
• OC-12 (STM-4) on the Engine 3 line card (12816 Core Router)
• Gigabit Ethernet (1 G) on the Engine 5 line card (12816 Core Router)
As mentioned earlier, the quantum will determine how many bytes can be served from the queue each
time the DRR scheduler visits that particular queue. The quantum for each queue is calculated as follows.
• Quantum (bytes) = MTU + (Weight – 1) * 512
Engine 3 and Engine 5 linecards allow use of MQC and hide the intricacies of MDRR from the user.
http://www.cisco.com/univercd/cc/td/doc/product/software/ios112/ios112p/gsr/wred_gs.htm#xtocid945524
As you read the descriptions of the features above, it should quickly become apparent that using some of
these features implies that other features cannot be used or must be configured in a more-specific manner.
Below are some common examples of feature interaction.
policy-map foo
class Prec_5
bandwidth 124400
shape average 104960000
policy-map foo
class Prec_5
bandwidth 124400
class Prec_2
bandwidth 186600
class Prec_1
bandwidth 186000
policy-map foo
class Prec_5
priority
class Prec_2
bandwidth 186600
class Prec_1
bandwidth 186000
policy-map foo
class Prec_5
priority
police 128000000 32000 64000 conform-action transmit exceed-action drop
class Prec_2
bandwidth 186600
class Prec_1
bandwidth 186000
Using WRED
This example shows how to configure multiple WRED profiles for various traffic classes. WRED can
only be applied when you have defined a queue by using either the shape or bandwidth keyword.
policy-map foo
class Prec_5
bandwidth percent 5
random-detect
random-detect precedence 5
class Prec_4
bandwidth percent 30
random-detect
random-detect precedence 4 1952 4000 1
class Prec_3
bandwidth percent 10
random-detect
random-detect precedence 3 1952 4000 1
class Prec_2
bandwidth percent 10
random-detect
random-detect precedence 2 1952 4000 1
class Prec_1
bandwidth percent 10
random-detect
random-detect precedence 1 1952 4000 1
class Prec_0
bandwidth percent 25
random-detect
random-detect precedence 0 2952 5000 1
class class-default
policy-map foo
class Prec_5
police 64000000 64000 128000 conform-action transmit exceed-action drop
priority
class Prec_4
shape average 200000000 262143
bandwidth percent 30
random-detect
random-detect precedence 4 1952 4000 1
class Prec_3
shape average 200000000 262143
bandwidth percent 10
random-detect
random-detect precedence 3 1952 4000 1
class Prec_2
shape average 200000000 262143
bandwidth percent 10
random-detect
random-detect precedence 2 1952 4000 1
class Prec_1
shape average 200000000 262143
bandwidth percent 10
random-detect
random-detect precedence 1 1952 4000 1
class Prec_0
shape average 200000000 262143
bandwidth percent 25
random-detect
random-detect precedence 0 2952 5000 1
class class-default
random-detect
random-detect precedence 0 2952 5000 1
random-detect precedence 1 2952 5000 1
random-detect precedence 2 2952 5000 1
random-detect precedence 3 2952 5000 1
random-detect precedence 4 2952 5000 1
random-detect precedence 5 2952 5000 1
random-detect precedence 6 2952 5000 1
random-detect precedence 7 2952 5000 1
!
interface pos 8/2
service-policy output foo
Input scheduling
Input scheduling prioritizes and schedules packets out of ingress packet queues based on several QoS
values including CoS and DSCP. However, most of Catalyst switches can deliver packets to the switching
fabric at line rate or a specified rate. This specific rate defines the maximum throughput of the switch. If
the input rate is not exceeded, input scheduling is not crucial in implementing QoS architecture.
The use of input scheduling is meant to deal with rare conditions of fabric oversubscription. Use of input
scheduling is therefore not deemed necessary in most scenarios. It will not be implemented in the Wateen
Telecom Network.
Internal DSCP
During processing of the packets, the priority of all traffic (including non-IP traffic) is represented
with an internal DSCP value in the Cisco 7600. During the classification process, QoS derives the
internal DSCP value depending on the trust state of the input interface, and the type of traffic:
• When PFC3B receives an IP packet (ip2ip, ip2mpls), it uses the input interface trust state and, if
configured, the policy-map trust command.
• For trust-cos input interface, internal DSCP is derived from received or port Layer 2 CoS values. The
following conversion table is used
7600#
7600#sh mls qos maps cos-dscp
Cos-dscp map:
cos: 0 1 2 3 4 5 6 7
------------------------------------
dscp: 0 8 16 24 32 40 48 56
7600#
• For trust-ipprec input interface, internal DSCP is derived from received IP precedence values.
The following conversion table is used:
7600#sh mls qos maps ip-prec-dscp
IpPrecedence-dscp map:
ipprec: 0 1 2 3 4 5 6 7
------------------------------------
dscp: 0 8 16 24 32 40 48 56
7600#
• For trust-dscp input interface, internal DSCP is derived from received DSCP values.
• For untrusted input interface, internal DSCP is derived from port CoS or policy map DSCP
values.
• When PFC3B receives an MPLS packet (mpls2mpls), it trusts the mpls EXP bits. The interface
trust state and the policy-map trust command have no effect.
• When PFC3B receives an MPLS packet (mpls2ip), trust depends on the type of label. However,
in all cases, the interface trust state and the policy-map trust command have no effect.
o Non-aggregate label: PFC3B trusts EXP in the topmost label.
o Aggregate label in VPN CAM: PFC3B trusts IP DSCP.
o Aggregate label not in VPN CAM: After recirculation, PFC3B trusts IP DSCP.
Packet in
Forwarding Operation Trust
Tycho
Routing ip2ip IP Interface trust; policymap trust
Imposition ip2mpls MPLS Topmost EXP
Swapping mpls2mpls MPLS Topmost EXP
Swapping +
mpls2mpls MPLS Topmost EXP
Imposition
mpls2mpls MPLS Topmost EXP
Disposition
Agg label in VPN CAM: IP DSCP
mpls2ip (no recirc) MPLS
Non-agg label: EXP
1st pass MPLS EXP
mpls2ip Agg label not in VPN CAM: IP DSCP
(recirc) nd
2 pass IP Non-agg label with output IP policy: IP
DSCP
o PFC3B propagates the popped EXP if the egress interface is known during the initial
lookup.
• Non-aggregate labels
• Aggregate labels in VPN CAM
o PFC3B cannot propagate the popped EXP if the egress interface is not known during
the initial lookup.
• For aggregate labels not in VPN CAM, propagation will not occur unless all
interfaces in the VRF have ‘mls propagate-cos’ configured.
• For explicit NULL labels, PFC3B cannot determine the egress interface,
hence, propagation does not occur.
Frames marked
Output Port
with CoS=5
placed into SP
Queue if one is
present SP Queue
Drop
Threshold 2
Drop
Threshold 1
Within each queue, drop thresholds are used to
indicate which CoS tagged packets can be dropped
once the queue has filled beyond a certain threshold
WRR
Weighted Round Robin (WRR) is another mechanism used in output scheduling used on the Catalyst
7600. WRR works between two or more queues. The queues that are making WRR are emptied in a
round robin fashion, and you can configure the weight for each queue.
Buffer Allocation
The transmit buffer is shared by the different queues of a port. Depending of the amount of traffic which
is expected in each of the traffic classes, and how bursty the traffic might be, the allocation of the buffer
assigned to each queue will be modified. The settings to be used in the Wateen Telecom network will be
described later in this document.
For each of the linecards, the different types of traffic (identified by the CoS value) will be mapped to the
different egress queues. Please note that LAN cards use WRR commands to enforce QoS policies. This
syntax is specific to LAN cards and respective capabilities of the card.
In the Wateen Telecom network following classe of service identifiers are suggested be used to determine
which flows are queued in which queue (or are distinguished for QoS treatment):
Wateen Telecom can adjust these according to their own requirements, once these requirements are
gathered and defined in terms of various classes of services desired.
As mentioned earlier there are several components and positions in network where QOS needs to be
defined, these are shown below:
Managed Services Scenario:
Unmanaged Services Scenario:
Unmanaged Services:
Managed Services:
• CPE: Policing, Classification, Marking, Queueing, congestion avoidance - WRED
• Edge: Queueing, WRED
• Core: Queueing, WRED.
Please note that lot of qos functionality in this case is off loaded to CPE devices.
This section talks a little more about each of the classes of traffic and the profiles for each of them for
features such as PHB (Per-Hop Behaviour, WRED configuration etc).
Link Bandwidth Allocation for Voice: Delay due to VoIP contention increases with VoIP traffic
volume as a percentage of link speed. It also increases with decrease in link bandwidth, for the same
percentage of link bandwidth usage (due to higher serialization delay). In real networks, VoIP delay will
depend several factors such as traffic mix, burstiness, link utilization, and an optimal Voice traffic
volume for a given link bandwidth is difficult to determine except via simulating the network in question,
or tests in the actual network. Many service providers try not to allocate more than 20-30% of link
bandwidth to Real-Time traffic class.
Per Hop Behavior (PHB): This traffic class must get Expedited Forwarding (EF) PHB via priority
queuing. This traffic will be served by low latency queueing.
WRED: WRED is NOT recommended for this class, since packet drop target is 0% (or nearly 0%), and
typically this class carries UDP traffic which does not react to WRED drops.
Policing:
The real-time class packets shouldn’t be dropped by the service provider, except when the traffic volume
exceeds the contracted bandwidth per SLA. This traffic class therefore uses ingress policing for SLA
enforcement at QoS trust boundary. Policing is configured to drop any out-of-contract traffic (rather than
marking to a lower priority class and transmitting to avoid the possibility of delayed delivery of real-time
traffic.
class SP-Real-Time
priority [percent <%>|<kbps>] <burst>
class SP-Real-Time
priority [percent <%>|<kbps>] <burst>
In the above configuration, the priority queue traffic is only policed when the interface is congested.
Policing doesn’t take place if there is no congestion. This mode is initiated by specifying the percent or
the class “rate” in the priority command.
class SP-Real-Time
priority
police <bps> bc <bytes> conform-action transmit exceed-action drop
The police command in the above configuration polices the priority queue permanently, even when there
is no congestion. This always-on policer is more favorable for service providers who wish to strictly
enforce the VoIP traffic volume contract. Congestion aware policer, on the other hand, can potentially
allow line rate realtime traffic if there is no other traffic (of other classes) to cause congestion.
It is recommended that Wateen Telecom uses option two (always-on-policer). In core of Wateen
Telecom, 12000 series engine 3 and engine 5 cards require second option i.e. an explicit policer must be
configured. In the absence of explicit policer on engine 3 and engine 5 lincards, all other traffic may be
starved as mentioned earlier. Using always on policer allows not only for consistency everywhere in the
network it also will allow Wateen Telecom to do bandwidth provisioning in their core appropriately.
Marking: Traffic of this class can be marked with different ip precedence values which can then be
aggregated in the core. It is recommended that traffic with ip precedence 3, 6 and 7 be aggregated in this
class.
PHB: This traffic class gets Assured Forwarding (AF) PHB via CBWFQ queuing
A minimum bandwidth guarantee is assured via a “bandwidth” command.
Latency for the class is minimized by assigning proportionately more minimum bandwidth guarantee to
this class that its average traffic volume (assumed to be the case for most platforms). However it is to be
noted that only minimum bandwidth guarantee is made for this class. No guarantees for delay or jitter can
be made.
WRED: WRED may be configured for this class to avoid TCP synchronization. WRED can also be used
in this class to ensure that packets with certain ip precedence /DSCP values get dropped during
congestion before those with other ip precedence/ DSCP values (by assigning them different WRED
thresholds). This would be useful, for example, if transactional-data (DSCP CS3) packets needs to be
dropped during congestion before mission-critical data (DSCP AF31) is dropped (both belong to this
traffic class). If, however, this class is optimized to provide very low drop rates, WRED may not be
configured.
Queue limit: If queue-limit is set for a class, verify (or set) the class queue size so that, it can
accommodate the expected traffic burst for the class. However, too large a queue-limit might increase the
queue size significantly to cause excessive latency due to delay when the queue fill up during congestion.
Since the SLA for this class includes some amounts of packet loss, the excessive latency can be reduced
by reducing the queue-limit, at the cost of increasing the possibility of packet drops during congestion.
If L = latency limit to be achieved in ms, and Cbw = Class bandwidth in bps, then
Queue-limit in bytes ~ L/1000 * Cbw /8
For example, if the queuing delay goal is 5ms on a router and the class minimum bandwidth guarantee is
10 Mbps, then the queue-limit in byte should be about 5/1000* 10*1024*1024/8 = 6553 bytes
Policing:
Ingress Policing is used to impose contractual traffic limits per SLA, and optionally, to mark traffic
exceeding the contractual limit with a DSCP/EXP of a lower priority traffic class such as best effort
traffic class. The policing burst size should be sufficient to accommodate expected traffic bursts in this
class.
Marking: Traffic of this class can be marked with different ip precedence values. It is suggested that
traffic with ip precedence 2 to be aggregated in this class.
PHB: This traffic class gets Assured Forwarding (AF) PHB via CBWFQ queuing
A minimum bandwidth guarantee is assured via a “bandwidth” command.
Latency for the class: Only minimum bandwidth guarantee is made for this class. No guarantees for delay
or jitter can be made.
WRED: WRED may be configured for this class to avoid TCP synchronization (for TCP based
applications).
Queue limit: If queue-limit is set for a class, verify (or set) the class queue size so that, it can
accommodate the expected traffic burst for the class. However, too large a queue-limit might increase the
queue size significantly to cause excessive latency due to delay when the queue fill up during congestion.
Since the SLA for this class includes some amounts of packet loss, the excessive latency can be reduced
by reducing the queue-limit, at the cost of increasing the possibility of packet drops during congestion.
If L = latency limit to be achieved in ms, and Cbw = Class bandwidth in bps, then
Queue-limit in bytes ~ L/1000 * Cbw /8
For example, if the queuing delay goal is 5ms on a router and the class minimum bandwidth guarantee is
10 Mbps, then the queue-limit in byte should be about 5/1000* 10*1024*1024/8 = 6553 bytes
Policing:
No ingress policing is expected be in place for this class as this is a class which carries SP infrastructure
traffic.
Marking: Traffic of this class can be marked with different ip precedence values. It is suggested that
traffic with ip precedence 4 to be placed in this class.
PHB: This traffic class gets Assured Forwarding (AF) PHB via CBWFQ queuing
When addressing the QoS needs of Streaming Video traffic, the following guidelines are recommended:
Streaming Video (whether unicast or multicast) should be marked to DSCP CS4 as designated by the
QoS Baseline.
Loss should be no more than 5 percent.
Latency should be no more than 4–5 seconds (depending on video application buffering capabilities).
There are no significant jitter requirements.
Guaranteed bandwidth (CBWFQ) requirements depend on the encoding format and rate of the video
stream. Streaming video applications have more lenient QoS requirements because they are delay-
insensitive (the video can take several seconds to cue-up) and are largely jitter-insensitive (because of
application buffering).
WRED: WRED may not be configured for this class, instead policing should be used based on video
application requirements and number of simultaneous flows .
Policing:
Ingress Policing is used to impose contractual traffic limits per SLA. No mark down is expected as the
class is strictly targeted for a single class of service.
This class usually has no latency, jitter, or loss commitments, and includes all traffic that are not included
in the other edge classes.
Marking: Traffic of this class is typically marked with ip precedence/ DSCP of 0 and 1 and EXP value is
set to 0 in the MPLS core.
PHB: This traffic class is assigned the class-default, and receives best effort treatment, bandwidth is
guaranteed if a bandwidth statement is assigned to this class.
WRED: WRED is configured for this class to avoid TCP synchronization. WRED can also be used to
drop certain type of best effort traffic prior to other types of best effort traffic (as mentioned in the
Business class)
Policing:
Similar to business class, except that exceeding traffic is always dropped (not re-marked to a lower
color).
There are a number of actions that can be configured at the CE. Wateen Telecom can have three ways to
which they want to classify their traffic:
• IP address range
• TCP/UDP Port Numbers
• TOS/COS Values
The most scalable method, and indeed the most commonly deployed and easiest to understand method is
to use TCP/UDP Port numbers (layer 4 information), this allows us to classify traffic into varying classes
based on the applications rather than using the IP addressing information.
The predominant drawback to using the IP addressing scheme is that it really requires the operator to
have a good knowledge of their customer network, the flows within etc, additionally as the customer
network grows and new prefixes are added the QOS policies will need to change, this will cause a
constant state of flux.
Basing QOS on the application layer means that regardless of the source/destination of the traffic the
correct QOS policy is defined based on the application (Voice, FTP, HTTP, SAP R/3, Oracle etc all use
well known ports or port ranges). This helps the SP to scale the QOS offerings and additionally makes the
per-customer policies much simpler.
At the CE, packets will be classified through the use of extended access lists (ACLs). These ACLs can
match packets on IP Source or Destination address, protocol type, and UDP/TCP port numbers.
ACLs for Real-Time and Business traffic should be agreed with the customer. VoIP typically uses UDP
ports in the range of 16384 to 32767.
!
access-list <x> permit udp any any range 16384 32767
!
An alternative for matching the VoIP traffic is to use “match ip rtp” in the class-map.
The ACL for Management traffic should match SNMP, TFTP, TELNET and any other required traffic to
and from the network management systems IP address range.
!
access-list x permit tcp any 192.168.1.0 0.0.0.255 eq telnet
access-list x permit udp any 192.168.1.0 0.0.0.255 eq snmp
access-list x permit udp any 192.168.1.0 0.0.0.255 eq tftp
!
The following is the ACL for the Best Effort traffic (100).
!
access-list x permit ip any any
!
Voice signalling traffic will need to be classified and marked appropriately. Depending on the customer
VoIP implementation, the different possibilities are:
Defining mission critical data is something that the customer has to supply to Wateen Telecom, each
customer will likely have a different set of requirements for their critical data, and therefore a different
QOS service policy will likely be necessary for each customer.
If Wateen Telecom trusts the traffic that is coming into the CE then there is no requirement for the CE to
have an input service policy because we are assuming that the ingress traffic is appropriately marked and
requires no further conditioning.
If Wateen Telecom wants to apply their own policies to ingress traffic at the CE i.e where the traffic is
not trusted from the customer then there should be an input service policy at CE, this basic service policy
will simply remark the traffic so that on egress it can be appropriately queued in the correct fashion.
Following is a sample input policing policy, it basically categorizes the ingress voice and video traffic as
well as the other classes to use a particular IP precedence.
policy-map INGRESS
class VOICE
set ip precedence 5
class VIDEO
set ip precedence 4
class BUSINESS-DATA
set ip precedence 3
class CRITICAL
set ip precedence 2
class INTERNET
set ip precedence 0
We can see on these access-lists that the varying traffic is classified and remarked at the ingress point of
the CE based on the TCP/UDP port ranges, this could be also based on IP address ranges, existing IP
Precedence / DSCP code point values and so on. This is really upto Wateen Telecom and the customer to
define.
The egress policy (from CE towards PE) is where we can configure the Priority Queue (LLQ) and the
constraints per traffic class such as bandwidth requirements, WRED etc.
This sample assumes that there is a 10mb link between CE-PE
policy-map EGRESS
class VOICE
priority 4000 # This uses 4Mb LLQ for Voice
police cir 4000 bc 1500 be 1500 conform-action transmit exceed-action drop
# # This polices the traffic to ensure
# there is no more than 40Mb
class VIDEO
bandwidth percent 20 # This uses 2Mb for Video
class CRITICAL
bandwidth percent 10 # This uses 20% of the 10mb available
random-detect # This uses WRED on the Data Class
class BUSINESS-DATA
bandwidth percent 10 # This uses 10% of the 10mb available
class INTERNET
bandwidth percent 10 # This uses 10% of the 10mb available
random-detect
class class-default
random-detect # run WRED on this class
Note: by default an interface (on the low end CE equipment) can only have 75% of it’s bandwidth used
for QOS, this is to cover overhead etc. This can be changed by using the command “max-reserved-
bandwidth” and expressing more than 75. This limitation does not exist on the 75/7600.
Note: in this example Voice and Video is split into two classes (VOICE and VIDEO) in two separate
classes. However if it is desired to place both video and voice in the same priority class and if the
customer access link (CE-PE) is less than 2mb, it is not recommended to combine Video with Voice in
the LLQ, this is because the serialization delays caused by large video packets will cause jitter and
latency in the voice traffic. Therefore in this scenario there should be a separate class for VOICE which
uses the priority mechanism and a separate class for VIDEO.
In that case, as shown in the above example, the video class is allocated a separate queue with a
bandwidth percent 20 to allow for 2Mb of 10Mb (but not in a LLQ).
Note: the size of the LLQ and the percentage of the bandwidth specified will vary depending on the
services that Wateen Telecom will sell and the requirements of the end users. Please also note that there is
only one LLQ on the system wide basis, regardless of number classes defined and configured with
“priority” command.
In the case of an unmanged CE service, classification and marking can be achieved at the Edge (PE)
layer. It is understood in intial deployment of this service Wateen Telecom will be using both managed
and unmanaged CE services model. PE however is not the most recommended position in the network to
configure these services as these will cause additional load on the device.
The configurations for classification and marking on the edge have to be re-applied here using the Cisco
MQC in absence of managed CE device.
Under managed CE scenario, at the Edge of the network we should be focused on queueing towards the
core and queuing towards the Access Layer.
As we saw earlier at the edge we can have more than three classes, however in the edge / core some of
these classes should be combined into a single queue to make QOS easier to provision in the core – most
SP’s have only 3 or 4 classes in their core. It is understood that Wateen Telecom has decided to keep all
five classes within the core as well.
For unmanaged CE, classification and marking will happen much the same way as if it was managed CE
but on the PE device, these are repeated here for convenience. These can be used on 7600.
policy-map INGRESS
class VOICE
set ip precedence 5
class VIDEO
set ip precedence 4
class BUSINESS-DATA
set ip precedence 3
class CRITICAL
set ip precedence 2
class INTERNET
set ip precedence 0
These in turn utilize the following Extended named access-lists as shown earlier:
!
ip access-list extended VOICE
permit udp any any range 16384 32767
!
ip access-list extended VIDEO
permit udp any any range 35001 36000
!
ip access-list extended CRITICAL
permit tcp any any eq bgp
!
ip access-list extended BUSINESS-DATA
permit tcp any any eq 3000
!
ip access-list extended INTERNET
permit ip any any
Please note that above acls are given as examples only. This policy can be applied to customer facing
(PE-CE) link ingress direction.
Interface serial 0/0
Service-policy input INGRESS
The 7600 has many varying different ways to provision Quality Of Service based on the ingress linecard
type, egress linecard type and so on, this document focuses on QOS for Gigabit Ethernet linecards and
WAN (SIP) based linecards.
The 7600 also has some limitations for QOS, WAN based linecards can only support a single match in a
class (this match could be an ACL to provide more complexity) , additionally the following match
commands are not supported on SIP-400:
* In Release 12.2(18)SXE and later releases, Supervisor Engine 720 supports the match protocol ipv6 command.
Additionally, the Cisco 7600 SIP-400 does not support the following QoS classification features with
AToM:
• Matching on data-link connection identifier (DLCI) is unsupported.
• Matching on virtual LAN (VLAN) is unsupported.
• Matching on class of service (CoS) is unsupported
• Matching on input interface is unsupported.
• Matching on packet length is unsupported.
• Matching on media access control (MAC) address is unsupported.
• Matching on protocol type, including Border Gateway Protocol (BGP), is unsupported.
* In Release 12.2(18)SXF and later relases: When .1Q encapsulation is configured on the 7600-SIP-400 with GE SPA, the CoS bits
in the VLAN tag can be used to differentiate the packet priority and a QoS service policy can be applied to the interface.
Note: 7600 same when using WAN (SIP) baesd linecards, the major exception being that the 7500 can
support the “class class-default” where as the 7600 (OSM only) cannot have a class-default and therefore
October 2, 2006 Wateen Telecom 253
Company Confidential. A printed copy of this document is considered uncontrolled.
Quality of Service
all possibilities have to be accounted for in the classes, the Flexwan is as per the 7500 and can have a
class default.
The following is a sample MQC policy-map that would be applied to a GigE uplink interface (egress)
towards the GSR core (PE-P) link or on a POS uplink STM-4 (oc-12) interface towards the GSR core
(PE-P).
class-map match-all VOICE
match mpls exp <topmost> 5
class-map match-all VIDEO
match mpls exp <topmost> 4
class-map match-all CRITICAL
match mpls exp <topmost> 2
class-map match-all BUSINESS-DATA
match mpls exp <topmost> 3 6
class-map match-all INTERNET
match mpls exp <topmost> 1 0
policy-map edge_oc12_policy
class VOICE
priority
police 248800000 conform-action transmit exceed-action drop
class VIDEO
bandwidth percent 20
random-detect
random-detect precedence 4 375 2423 1 !WRED threshold per Cisco Recomm.*
class BUSINESS-DATA
bandwidth percent 10
random-detect
random-detect precedence 3 375 2423 1
random-detect precedence 6 375 2423 1
class CRITICAL
bandwidth percent 10
random-detect
random-detect precedence 2 375 2423 1
class INTERNET
bandwidth percent 10
random-detect
random-detect precedence 1 375 2423 1
random-detect precedence 0 175 1500 1 # randomly populated to
#differentiate and drop prec 0 first
Exception to above policy is when LAN based cards are being used. This will be discussed below:
Following scenarios are possible with 7600:
Ingress line card: Gigabit Ethernet linecard, & Egress linecard: SIP-400
1. Ingress policy as described above can be applied to gigabit Ethernet linecards for
marking/remarking.
In order for Ethernet linecards to be able to classify, police and mark, qos must first be globally
enabled by using the following command:
mls qos
To verify qos is globally enabled, show mls qos command can be used.
To trust incoming traffic, the ports must be classified as trusted by using the following command
under the interface:
mls qos trust dscp
Qos queueing features are determined by the type of ethernet linecard on 7600 series router.
While PFC on Supervisor 720 performs classification, marking, mapping, and policing functions, all
queuing and congestion avoidance policies are administered by the linecards. This inevitably leads to per-
linecard hardware-specific capabilities and syntax when it comes to configuring queuing and dropping.
While ingress queuing capabilities exist (i.e.queueing towards switch fabric), receive queues are
extremely difficult to congest, even in controlled lab environments. Ingress congestion implies that the
combined ingress rates of traffic exceed the switch fabric channel speed, and thus would need to be
queued simply to gain access to the switching fabric. On Supervisor 720, this means that a combined
ingress rate of up to 40 Gbps per slot would be required to create such an event, Therefore it is
recommended to implement egress / transmit queuing.
There are currently six main transmit queuing/dropping options for Ethernet linecards:
• 2Q2T—Indicates two standard queues, each with two configurable tail-drop thresholds.
• 1P2Q1T—Indicates one strict-priority queue and two standard queues, each with one configurable
WRED-drop threshold (however, each standard queue also has one nonconfigurable tail-drop threshold).
• 1P2Q2T—Indicates one strict-priority queue and two standard queues, each with two configurable
WRED-drop thresholds.
• 1P3Q1T—Indicates one strict-priority queue and three standard queues, each with one configurable
WRED-drop threshold (however, each standard queue also has one nonconfigurable tail-drop threshold).
• 1P3Q8T—Indicates one strict-priority queue and three standard queues, each with eight configurable
WRED-drop thresholds (however, each standard queue also has one nonconfigurable tail-drop threshold).
• 1P7Q8T—Indicates one strict-priority queue and seven standard queues, each with eight configurable
WRED-drop thresholds (on 1p7q8t ports, each standard queue also has one nonconfigurable tail-drop
threshold).
Almost all Ethernet linecards support a strict-priority queue and when supported, the switch services
traffic in the strict-priority transmit queue before servicing the standard queues. When the switch is
servicing a standard queue, after transmitting a packet, it checks for traffic in the strict-priority queue. If
the switch detects traffic in the strict-priority queue, it suspends its service of the standard queue and
completes service of all traffic in the strict-priority queue before returning to the standard queue.
The Transmit Queuing/Dropping capabilities can be determined by using the “show queueing interface”
commands.
In summary we have two distinct qos capabilities on transmit queue side : 1p3q8t and 1p2q2t. Each of
queuing structure is discussed below:
1P2Q2T
Qos must first be globally enabled by using “mls qos” command in config mode. Please note that the
buffer allocation for the PQ (Q3) is not configurable, but is, by default (for the 1P2Q2T queuing structure
only) set to equal the size defined for Q2.
There are two points to observe: A) there are only three queues, (one strict priority and two queues .B)
Size of PQ is set same size as Q2.
This means that In order to closely mimick our qos allocation of:
40% voice
20% video
10% business-data
10% critical
And 10% of internet, some classes will need to be aggregated.
Therefore, the queue size ratios have been changed, specifically Q1 is set to 30% and Q2 is set to 35%
(which indirectly sets Q3 to match, at 35%). Voice will be served by strict priority queue with 35%. The
internet traffic (0,1) will be mapped to Q1. Video, Business Data and Critical class will be mapped to Q2.
Q2 will be configured such that it is serviced 75% of the time, compared to Q1 which will be served 25%.
These values are chosen as examples, Wateen Telecom can modify them based on their requirements.
1P3Q8T
Under this mode, there are three queues while one strict priority queue, this is queue number 4.
wrr-queue random-detect max-threshold 1 100 100 100 100 100 100 100 100
#Sets Max WRED Threshold
for Q1T1 to 100% and all
others to 100%
wrr-queue random-detect min-threshold 2 80 100 100 100 100 100 100 100
#Sets Min WRED Threshold
for Q2T1 to 80% and all
others to 100%
wrr-queue random-detect max-threshold 2 100 100 100 100 100 100 100 100
#Sets Max WRED Threshold
for Q2T1 to 100% and all
others to 100%
wrr-queue random-detect min-threshold 3 50 60 70 80 90 100 100 100
#Sets Min WRED Threshold
for Q3T1 to 50%, Q3T2 to
60%, Q3T3 to 70% Q3T4 to
80%, Q3T5 to 90% and all
others to 100%
wrr-queue random-detect max-threshold 3 60 70 80 90 100 100 100 100
# Sets Max WRED Threshold
for Q3T1 to 60%, Q3T2 to
70%, Q3T3 to 80% Q3T4 to
90%, Q3T5 to 100% and all
others to 100%
(Q4)
Egress on the EDGE layer (towards the CE) it is also necessary to implement a QOS policy for an end to
end qos service, however it is not needed under unmanaged CE offering. Under managed CE scenario,
this will be the same policy in egress direction towards PE-CE as it is applied in egress direction from
CE-PE.
For 7600 series routers in city and remote PoP, if the traffic is coming in on a trusted interface (managed
CE) “mls qos trust [dscp | ip-precedence | cos]” can be used to derive relative qos values and match on
ip-precedence or exp value for egress queueing.
For 7600 Series router in the core PoPs , only Ethernet cards are used, as such line card specific policies
should be used as described in the previous section. When 7600 receives MPLS packets, the PFC3BXL
always trusts EXP in the received topmost label. None of following have any effect on MPLS packets:
•Interface trust state
•Port CoS value
•Policy-map trust command
Qos should be enabled on the system with “mls qos” command.
Once the core layer is reached (the GSR’s), Wateen Telecom will need to perform appropriate queuing
and scheduling in the core.
Below is a sample QOS configuration for the GSR for egress queueing for an STM-16 linecard.
Engine 3 & Engine 5 support the Cisco MQC for egress QOS.
policy-map core_oc48_policy
class REALTIME
priority
police 958400000 conform-action transmit exceed-action drop
class VIDEO
bandwidth percent 20
random-detect # potentially not used for Video
#traffic
random-detect precedence 4 6220 20733 1 # depends on application profile
class BUSINESS-DATA
bandwidth percent 10
random-detect
random-detect precedence 6 6220 20733 1
random-detect precedence 3 3220 10733 1 # Randomly populated to drop pre 3
#first
class CRITICAL
bandwidth percent 10
class INTERNET
bandwidth percent 10
random-detect
random-detect precedence 1 6220 20733 1
random-detect precedence 0 3220 10733 1 # Randomly populated to
#differentiate and drop prec 0 first
This shows a sample egress QOS policy for an STM-16 linecard, we can see the LLQ is policed at a
maximum of 958.4 Mb, this is approximately 40% of STM-16 linerate, the critical class uses 10% of the
bandwidth, WRED values used are values as requested by the customer, however these may still need to
be tuned by Wateen Telecom depending on the types of traffic etc in the network.
Finally there is a class called class-default which gets all other traffic types (unaccounted for by Qos
policies). Depending on any unaccounted traffic, it is possible to make bandwidth guarntees for class-
default as well. e.g. it is possible to use class class-default as a class to account for internet and any other
traffic. Given that in our case internet traffic is explicitly accounted for, class-default configuration is not
needed.
Please note that Cisco 12000 series routers by default perform qos based on Layer-3 payload
(i.e.excluding layer-2 overhead of underlying physical transport mechanism). While it is possible to have
this adjusted for some line cards and certain transport types, it is not possible to do the same for all line
cards and all interface types. The Layer-2 overhead adjustment is primarily targeted towards customer
facing ports to meet customer SLAs rather than core facing ports. This is discussed further below:
http://www.cisco.com/en/US/products/sw/iosswrel/ps1829/products_feature_guide09186a00805e8fbb.ht
ml
For Wateen Telecom it is suggested to leave qos features such as policing, shaping, and bandwidth as
layer-3 as the these are core links (not customer facing ports) and the feature is not available on all
linecards. What this implies is that while performing bandwidth allocation, overall Layer-3 capacity of
the interface should be considered rather than physical transport characteristics. E.g. STM-1 link which
has an overall capacity of 155Mbps cannot be used to make bandwidth computations, instead 149 Mbps
should be used for such computations. This number excludes SONET overhead of the traffic. Similarly
STM-4 will have overall capacity of 599Mbps and STM-16 will have overall capacity of 2396Mbps.
WRED threshold is defined as queue depth in terms of packet size. On a 12000 series router, current
version of IOS also allows defining the threshold in terms of units of times (which is subsequently
converted into packets by the system) as seen below. This is referred as “Time-Based Thresholds for
WRED” and was introduced in 12.0(28)S.
router(config-pmap-c)#random-detect precedence 3 30 ?
<1-262143> maximum threshold (in packet by default)
cells in cells
ms milliseconds
packets in packets
us microseconds
router(config-pmap-c)#random-detect precedence 3 30 ms 100 ms
STM-16 Link:
For an STM-16 (OC-48) link, minimum threshold of 6220 and maximum threshold of 20733 is suggested
per the customer requirement. It is understood that customer will like to use buffering thresholds of 30
and 100 milliseconds for MinTH and MaxTH respectively. These values are based on MTU size of 1500
bytes (worst case scenario). A better approximation can be had by using the average size of packet on
customer network. As indicated earlier, these values will need to be tuned over period of time.
Computation for different link sizes is shown below.
B= 2488Mbps/1500*8bits = 207333.3
For Min threshold: .03 * B = 6220 packets
For Max threshold: .1* B = 20733 packets
• Please note that values will be adjusted to be a number which is a power of 2 for min threshold based
on max threshold i.e. the (MaxTH – MinTH) needs to be a power of 2 as documented earlier.
1Gigabit Ethernet link:
B=1000Mbps/1500*8=83333.3
For Min threshold: .03*B = 2500 packets
STM-4 link:
B=622Mbps/1500*8= 51833
For Min threshold:.03*B= 1555 packets
For Max threshold: .1*B =5183 packets
RSVP signaling & keepalive mechanism keeps sure that tunnel state is always detected & they stay up as
long as they’re needed & then thry’re torn down via RSVP signalling.
Then RSVP sends a PATH message via the path obtained through CSPF. The PATH message travels
from the head-end to the tail-end. This PATH message checks with each router in the path if they have
the rsources available that were requested by the tunnel. The PATH message also has new objects for
MPLS TE, which are discussed in later sections. One important new object is the LABEL_REQUEST
object. This object has information about the Layer 3 protocol being used, & in case of ATM or Frame
Relay networks, it also has a range in which labels can be allocated. This object indicates that Label
binding for this path is requested
Once the PATH message reaches the tunnel destination (aka tail-end), the tail-end generates an RSVP
RESV message which goes back to the head-end router.
The RESV message travels from the tail-end router to the head-end, in the reverse direction of the PATH
message. Every router sending the RESV message sends a LABEL object to the upstream router, so that
the upstream router can bind this as an out-going LABEL for the in-coming LABEL (which the upstream
router will send as the LABEL object in the RESV message to its upstream router).
Routers actually lock the resources requested by tunnels, when they see the RESV message. This prevents
pre-mature locking of resources during the PATH message, say if the PATH message were to fail
somewhere downstream
So in short RSVP helps signal a tunnel through a network & makes sure that the resource requirement for
the tunnel can be met.
After a tunnel is established, PATH & RESV messages keep on going thru the tunnel path & the
receiving routers use this as a keepalive mechanism for this tunnel. We tear down the tunnel, if we miss 4
consecutive PATH or RESV messages. This is because RSVP is a soft-state protocol & this is the only
way RSVP has to keep track of state.
TE tunnels forward traffic using label switching. The advantage with this is that the tunnels can be routed
independent of IGP best-path routing & this makes Source Based Routing possible, where the head-end
decides that packets need to go thru a certain tunnel because of some policy.
Bandwidth
This attribute specifies the amount of bandwidth the traffic trunk requires.
Adaptability
This attribute indicates whether the traffic trunk should be re-optimised. The re-optimisation procedure is
discussed in a later section.
Resilience
This attribute specifies the desired behaviour under fault conditions, i.e., the path carrying the traffic
trunk no longer exists due to either network failures or preemption. TE restoration operation is discussed
in a later section.
Priority
Priority is the mechanism by which the operator controls access to resources when the resources are
under contention. It is a required function to place all traffic trunks. Another important application of the
priority mechanism is supporting multiple classes of services. We assign two types of priorities to each
traffic trunk: holding priority, and setup priority. Holding priority determines whether the traffic trunk
has the right to hold a resource reservation when other traffic trunks attempt to take away its existing
reservation. Setup priority determines whether the traffic trunk as the right to take over the resources
already reserved by other traffic trunks.
Resource Attributes
Resource attributes are used to describe the network links used for path calculations. There are three
resource attributes, each of which is described below.
Available Bandwidth
This attribute describes the amount of bandwidth available at each setup priority. Note that the available
bandwidth for the higher setup priority is always larger than that for the lower setup priority. This
attribute needs not necessarily reflect the actual available bandwidth. In some cases, the network operator
may oversubscribe a link by assigning a value which is larger than the actual bandwidth, e.g., 49.5 Mbps
for a DS-3 link.
Resource Class
This attribute indicates the resource class of a link. Recall that the trunk attribute, resource class affinity,
is used to allow the operator to administratively include or exclude links in path calculations. This
capability is achieved by matching the resource class attribute of links with resource class affinity of
traffic trunks. The resource class is a 32-bit value. The resource class affinity contains a 32-bit resource
affinity attribute and an associated 32-bit resource class mask. A link cannot be considered in the path
calculation for a traffic trunk unless the following equation holds:
Path Selection
Path selection for a traffic trunk takes place at the head-end routers of traffic trunks. Using extended IS-
IS/OSPF, the edge routers have knowledge of both network topology and link resources. For each traffic
trunk, the router starts from the destination of the trunk and attempts to find the shortest path toward the
source (i.e., using the shortest path first (SPF) algorithm). The SPF calculation does not consider the
links which are explicitly excluded by the resource class affinities of the trunk, as well as the links which
have insufficient bandwidth. The output of the path selection process is an explicit route consisting of a
sequence of label switching routers. This path is used as the input to the path setup procedure.
Path Setup
Path setup is initiated by the head-end routers. RSVP is the protocol which establishes the forwarding
state along the path computed in the path selection process. The head-end router sends a PATH message
for each traffic trunk it originates. The PATH message carries the explicit route computed for this traffic
trunk. As a result the PATH message always follows this explicit route. Each intermediate router along
the path performs trunk admission control after receiving the PATH message. Once the router at the end
of the path receives the PATH message, it sends a RESV message in the reverse direction towards the
head-end of the traffic trunk. As the RESV message flows toward the sender, each intermediate node
reserves bandwidth and allocates labels for the trunk. Thus when the RESV message reaches the sender,
the LSP is already established. The following Figure shows an example of the path setup procedure.
R8 R9
R3
R4
R2 Pop
R5
R1 32
49
17
R6 R7
22
Path Maintenance
Path maintenance refers to two operations: path re-optimization and restoration. Re-optimization
periodically checks to see if here is a better path for the LSP than has previously been calculated.
The following figure illustrates this process of Fast ReRoute. A primary tunnel has been created that
traverses {R1, R2, R4, R5} with R1 being the head-end (where the tunnel was configured) and R5 being
the tail end (the tunnel destination). In our example, the link between R2 and R4 is protected by Fast
ReRoute using a secondary tunnel referred to as the backup tunnel. This backup tunnel traverses {R2, R3,
R4) where R2 is the head-end for the backup tunnel and R4 is the tail-end. In the event of a R2/R4 link
failure, R2 will detect the link is not operational, and immediately reroute the primary TE tunnel for that
Backup R3
Primary Tunnel
Backup R3
It is important to understand the TE tunnels are unidirectional. Therefore there must be one in each
direction. In the case of the previous example, there would be a backup tunnel running between {R4, R3,
R2} and also a primary tunnel from {R5, R4, R2, R1}
Primary Tunnels
To provide link protection two types of tunnels exist:
• Primary tunnels
• Backup Tunnels
Under normal or steady state conditions (no failure present) traffic offered to the primary tunnel will
follow the label switch path (LSP) of the primary tunnel between the head-end and the tail-end routers. In
the event of a failure the primary tunnel traffic will be diverted along the backup tunnel LSP.
For traffic to be protected by a backup tunnel with FRR it must originally be carried in a primary tunnel.
In the event of a link failure, any traffic that is not part of a tunnel will not be able to take advantage of
the FRR of a backup tunnel but instead rely on the convergence of the IGP. The following figure
illustrates this concept. All traffic that enters the tunnel at the head-end router R1 will be encapsulated
with the label provided by RSVP for the primary tunnel. Upon failure of the protected link a second label
representing the backup tunnel will be pushed onto the stack for MPLS packets holding the primary
Non
Tunnel R3
Traffic
Link Failure
Head End Tail End
R1 R5
R2 R4
Tunnel
Primary Tunnel Backup Tunnel Primary Tunnel
Traffic
B Í
el R3 Pr
unn Î Pr ima
T A im ry
ry n el ar Tu
a n yT nn
Pr
im Tu un el
B
a ry ne
Í lÎ
im
Pr
R2 R4
R1 R5
For the case of link protection, primary and backup tunnels are created as zero bandwidth tunnels.
There is no bandwidth constraint placed upon them for LSP creation.
OSPF Autoroute
By default TE tunnels are not visible to the IGP. A feature called Autoroute allows TE tunnels to be made
available to OSPF. Autoroute causes OPSF to include the primary tunnel interfaces in its enhanced SPF
calculations thereby replacing the physical next hop interfaces with a tunnel interface.
For example, Lhe-12816-P has three physical interfaces connecting to Karachi, Islamabad and
Faisalabad. This means there will be three one hop primary tunnels corresponding to each physical
interface. OSPF will use each of these tunnels in its SPF calculations.
LDP LDP
Í
lB R3 Pr
n ne Î Pr im
ar
Tu lA im yT
ar
y
nne ar
yT un
ne
Pr
im Tu un lB
y ne
LDP Í ar lÎ LDP
im
Pr
R2 R4
R1 R5
If an LDP session is not established over the tunnel then the outgoing label will be set to “untagged” that
will either black hole the traffic or misinterpret the exposed label (usually the VPN label) at the next hop
in the TFIB.
Backup Tunnels
As previously mentioned every link in the Wateen Telecom core can be associated with a one hop
primary tunnel. In addition, every physical interface can be protected by an associated backup tunnel. The
backup tunnel is configured on the physical link and the primary tunnel is configured to be protected by
the backup tunnel through the use of the “fast reroute” option.
Backup tunnels are uni-directional, therefore as is the case with primary tunnels there will be two backup
tunnels for each link representing the A Æ B and B Æ A direction.
Even though each backup tunnel traverses a series links that are associated with one hop primary tunnels,
the backup tunnels themselves do not travel within a primary tunnel. They are distinct LSPs in their own
right that have been calculated independently. As with all TE tunnels, RSVP is used to distribute the
labels of a backup tunnel.
In event of a backup tunnel being brought into operation due to a link failure, the relevant outgoing label
associated with the backup will be pushed onto the label stack of the affected packets to reroute the traffic
around the fail link.
General Configuration
The following configuration example shows the general TE configuration for all Wateen Telecom core
routers. The first command mpls traffic-eng tunnels enables TE on a global basis. The second command
mpls traffic-eng reoptimize events enables automatic reoptimisation of TE tunnels when a link comes
into the UP state. The default reoptimisation timer is set to 3600 seconds.
The OSPF TE sub-commands associated the TE router-id with the loopback interface and also allow
flooding of TE information in to OSPF area 0.
mpls traffic-eng tunnels
mpls traffic-eng reoptimize events link-up
!
router ospf 1
mpls traffic-eng router-id Loopback0
mpls traffic-eng area 0
!
Link Configuration
The following configuration example shows the interface configuration for POS interfaces.
Sample Configuration
The sample configuration in the following example is shown as a template. Two tunnels would be
required for the primary and backup and associated explicit paths. The primary tunnel will need to be
configured with fast reroute and autoroute.
The example is using the details for the Lahore-Karachi link, from the LHE-12816-P perspective:
!
!
mpls traffic-eng auto-tunnel primary tunnel-num min 61000 max 61999
mpls traffic-eng auto-tunnel backup tunnel-num min 62000 max 62999
!
mpls traffic-eng auto-tunnel backup nhop-only
mpls traffic-eng auto-tunnel primary onehop
mpls traffic-eng auto-tunnel primary config mpls ip
mpls ldp discovery targeted-hello accept
!
Overview
Following section briefly discusses relevant network management and infrastructure related best practices
as they pertain to Wateen Telecom’s IP/MPLS network elements e.g. 7600 and 12000 series routers.
Complete network management low level design for overall Wateen Telecom’s infrastructure
including all applications and network elements is discussed in a separate low level design
document.
Similarly overall security and data center solution for Wateen Telecom is documented in a separate low
level design document.
There are many other traps, which can be enabled in IOS. Their use should be carefully considered as the
intention should not be to overwhelm the NOC staff with too much information; rather provide them with
useful information. Following is an example list of traps that can be enabled.
The following commands will be required to configure SNMP trap notification on Cisco IOS devices.
snmp-server trap-authentication
snmp-server host <nms-server ip address> XXXX #Send SNMP traps to NMS server
snmp-server trap-source loopback 0 #Source all SNMP trap messages from
loopback interface
snmp-server enable traps snmp # provide linkUp, linkDown, coldstart &
warmstart traps
snmp-server enable traps config
snmp-server enable traps envmon
snmp-server enable traps bgp
snmp-server enable traps mpls ldp
snmp-server enable traps mpls vpn
CDP is a valuable tool for use when troubleshooting network issues. CDP output can be used to quickly
construct a visual depiction of the network as CDP neighbors can be traced from one to the next through
the use of the “show CDP neighbor” command.
From network security perspective:
• Disable CDP Server where not used - Typically CDP is used on internal networks for network
management, but can pose a risk at the edge of the network.
• From a security perspective, CDP should be disabled where it is not being used, in order to avoid
unnecessarily advertising system specific information. The following guidelines should be followed
when implementing CDP:
o CDP can be disabled on a per interface basis. Run it only where absolutely necessary,
disable it on a per interface basis everywhere else.
o Confine CDP deployment to run between devices under your control that are directly
adjacent. Since CDP is a link-level protocol, it is not transient across a network. Limit it to
run only between trusted devices.
o Disable CDP everywhere else, especially on the interfaces at the edge of the network.
To disable CDP globally on an IOS device use the command:
“no cdp run”
To disable CDP on an IOS device interface use the following interface command:
“no cdp enable”
Most service providers choose to disable CDP to provide extra security, so for Wateen Telecom the
recommendation is to leave CDP enabled during the initial installation and then to disable it
subsequently.
Logging
Logging of events on a network is an important part of any network security policy. The logging of
events assists problem troubleshooting and security investigations.
• Message Logging - Cisco routers can log information regarding configuration changes, ACL
violations, interface status changes and many other types of events. Logging messages can be
directed to several different locations:
o Console – used when modifying or testing the device while connected to the console.
Messages sent to the console are not stored by the device, and are therefore not very
valuable as security events.
o Terminal lines – enabled EXEC sessions can be configured to receive log messages on any
terminal lines. Like console logging, this type of logging is not stored by the device and
therefore is only valuable to the user on that line at that particular time.
o Memory buffer – the device can be directed to store log messages in the local memory.
Buffered memory is more useful as a security tool than console and terminal messages, but
has the disadvantage that the events are cleared out if the device is reloaded.
o SNMP traps – certain router events may be processed by the router SNMP agent and
forwarded as SNMP traps to an external SNMP host. This is a viable security logging
facility, but requires the configuration and maintenance of an SNMP system.
o Syslog – Cisco devices can be configured to forward log messages to an external Syslog
service. It is highly recommended that networks implement a logging structure based on a
Syslog infrastructure. This is a scalable solution, which provides long-term storage
capabilities and a central location for all device messages.
N.B. performing analysis on a device’s logs and messages can be difficult if the device clocks are all
running on different times. It is recommended that the NTP (Network Time Protocol) be utilised to ensure
that all devices are operating with the same time information. See later section on securing NTP.
• Log Severity Levels - When logging, it is important to capture the necessary amount of information.
The granularity of detail in logging information can also be configured to one of eight levels, as
shown below:
Table 39 Security Levels
The lower the level number, the higher the security level. A good level of general logging for everyday
information capture is “informational”. Additional detail can be captured with the “debug” level, but
should only be used temporarily for specific situations. Debug logging levels can become extremely
processor intensive thereby impacting system performance.
• Securing Syslog Messages - There are several security issues related to Syslog messages as detailed
below:
o Syslog is sent as clear text between the managed device and the management host (UDP
port 514).
o Syslog has no packet-level integrity checking to ensure that the packet contents have not
been altered in transit.
o There is a potential for the Syslog data to be falsified by an attacker.
o An attacker could send large amounts of false Syslog data to a management server in order
to confuse the network administrator during an attack.
When possible, the following security practices should be implemented to mitigate these issues:
o If possible Syslog traffic should be encrypted within an IPSec tunnel in order to mitigate the
chance of it being altered in transit.
o If there are no managed devices outside of the network that send Syslog messages to the
Syslog server, then ACLs should be used at the edge of the network to block external Syslog
messages from entering the network.
o If there are managed devices outside of the network, ACLs should be implemented at the
edge of the network to allow only Syslog data from the managed devices themselves to
reach the Syslog server.
• Enable Buffered Logging - Cisco devices can store log messages in memory. The buffered data is
available only from an exec or enabled exec session, and it is cleared when the device reboots. This
form of logging is useful, even though it does not offer enough long-term protection for the logs.
Command:
“logging buffered”
• Disable Logging to the Console - Limit the severity level of messages sent to the console or disable
logging to the console. This form of logging is not persistent; the device does not store messages
printed to the console. Although useful for troubleshooting from the console port, it is possible that
excessive log messages on the console could make it impossible to manage the device, even from the
console. Console logging can be disabled using the following command:
“no logging console”
Console logging may later be enabled when required, such as during a debugging or trouble-shooting
session, or if detailed tracking is being implemented via hard logging from the console directly to a
printer.
• Set trap level – Determine the severity of messages that will generate a trap. The following
command sets the trap level:
“logging trap <level>”
• Set loopback as source of logging – If the loopback interface is not available, for example in Metro
Ethernet solutions, then a suitable interface should be used, such as the management VLAN.
“logging source-interface Loopback0”
The following example shows how to enable syslogging to a server:
no logging console
no logging monitor
!
!
service time log datetime localtime show-timezone msec
service time debug datetime localtime show-timezone msec
The configuration above sends each Syslog message to two different servers for analysis. It is preferable
to send all messages to a single server to save bandwidth, and then allow other NMS platforms access by
doing an NFS mount to the syslog directory.
Cisco recommends as a best practice to always configure no ip unreachable on an interface that has an
ACL applied, to ensure that someone sending traffic that is denied by an ACL does not get a notification
regarding this.
Before any of these services are disabled, they should be reviewed in the context of the network to ensure
that they will not effect the current operation of the network.
The following services should be disabled if they are not being used:
• Disable Small Servers - Small services, such as echo, chargen, daytime and discard are enabled by
default on a router in IOS version < 11.3, and disabled by default in IOS version > 11.3. These
services, especially their UDP versions, can be used to launch DoS and other attacks that would
otherwise be prevented using packet filtering. As these services are rarely used they should be
disabled, especially on routers acting as firewalls and those in security-critical parts of the network.
These services are disabled by default in IOS version 12.0 and later; in earlier software can be
disabled using the following global commands:
“no service tcp-small-servers”
“no service udp-small-servers”
• Disable Finger Service - Cisco routers provide an implementation of the “finger” service, which is
used to discover the users that are logged into a device. The information it provides could be useful
to an attacker. The service is enabled by default, and if not required should be disabled using the
following commands:
“no ip finger”
“no service finger”
• Disable Configuration Auto-loading Service - Routers can be configured to load their startup config
from a network server as well as their local memory. Loading router configurations across a network
can be dangerous and should only be used in fully trusted network situations. The service is disabled
by default, but if enabled can be disabled using following commands:
no boot network <remote-url>
“no service config”
• Disable Bootp Server - Bootp is a UDP based service that allows Cisco routers to obtain copies of
IOS from other routers also running the bootp service. In reality, the service is rarely used and could
be used by an attacker to download a copy of a router’s IOS. The service is enabled by default, and
can be disabled using the following command:
“no ip bootp server”
• Disable IP Source Routing - IP source routing enables the sender of a datagram to specify the route
the datagram will take on route to its destination, and generally the route that any reply will take
when returning to the originator. Although enabled by default, source routing is rarely used and
could be used by an attacker. If a network does not require source routing, it should be disabled using
the following command:
“no ip source-route”
This will cause the router to drop any IP packet that carries a source route option.
• Disable TFTP Server - It is recommended to have a central storage location for IOS images and
restrict access to only the loopback ip address of the routers and switches on the network. It is
therefore recommended to disable the local tftp servers running on the network and to put all the
images needed for the operation of the network on a dedicated server inside the NOC with strict
ACLs that only allow tftp traffic inside, if sourced from the loopback ip addresses of the routers or
switches on the network. The following command will disable the tftp-server:
“no tftp-server”
The following interface services should be disabled if they are not being used:
• Disable IP Directed Broadcasts - An IP directed broadcast is a datagram sent to the broadcast address
of a subnet to which the sending device is not directly attached. The directed broadcast is routed
through the network as a unicast packet until it arrives at the target subnet, where it is converted into
a data-link-layer broadcast.
Directed broadcasts are rarely used for legitimate purposes, and are utilised in the “SMURF” and
other related attacks. The service is enabled in IOS versions < 12.0 and disabled in IOS versions ≥
12.0. The following interface command will cause directed broadcasts to be dropped instead turned
into data-link-layer broadcasts:
“no ip directed-broadcast”
As only the last router in the chain of routers that the packet passes through can recognise that the
packet is a directed-broadcast, the above command must be configured on every interface that could
be a target. It is not sufficient to configure only the edge routers or firewalls.
• Disable IP Redirects - ICMP redirect messages are enabled by default, and instruct an end device to
use a specific router in its path to a destination. By design, a router will send redirects only to hosts
on its local subnet, no end device will ever send a redirect, and no redirect will be sent more than one
network hop away. However, an attacker may violate these rules to launch an attack on a network. If
not required, ICMP redirect messages can be disabled using the following interface command:
“no ip redirects”
It is also possible to filter out ICMP redirect messages using ACLs on routers located at the edge of
the network. This will prevent redirect attacks launched by remote attackers.
• Disable IP Unreachable Messages - Attackers use ICMP unreachable messages to try to map a
network. These messages are enabled by default, and should be disabled on all interfaces if not
required, especially those connected to untrusted networks: Use the following interface command to
disable unreachable messages on an interface:
“no ip unreachable”
• Disable IP Mask Replies - Mask replies are disabled by default, but if enabled, the router responds to
ICMP mask requests with ICMP mask replies, which could provide an attacker with important
network information. Mask replies should be disabled on all interfaces, especially those at the edge
of a network, using the following command:
“no ip mask-reply”
• Disable Unused Router Interfaces - If an attacker has physical access to a router, any un-used
interface could be used to gain access to the router and the network. All un-used interfaces on a
router should be disabled in the configuration using the shutdown command.
Cisco routers and switches support a number of services that can be enabled on a device to improve the
overall security of a network. Before any of these services are enabled, they should be reviewed in the
context of the network to ensure that they will not effect the current operation of the network.
• Enable TCP Keepalive Feature – Idle logged-in user sessions can be susceptible to unauthorized
access and hijacking attacks. By default, Cisco routers do not continually test whether a previously
connected TCP endpoint is still reachable. If one end of a TCP connection idles out or terminates
abnormally (for example, crashes or reloads), the opposite end of the connection may still believe the
session is available. These “orphaned” sessions use up valuable router resources. Attackers can take
advantage of this weakness to attack devices.
To mitigate this problem, Cisco routers can be configured to send periodic keepalive messages to
ensure that the remote end of a session is still available. If the remote device fails to respond to the
keepalive message, the sending router will clear the connection. This immediately frees router
resources for other more important tasks.
The following IOS commands enable the tcp-keepalive feature:
service tcp-keepalives-in
service tcp-keepalives-out
• Core Dumps – The core dump facility allows a copy of a device’s memory image to be stored and
uploaded to another device after a crash. When a router crashes, a copy of the core memory is kept.
Before the memory is erased on reboot, the router can be set up to copy the core dump to a UNIX
server. An account (FTP, TFTP, or RCP) and sufficient disk space (equal to the amount of memory
on the router per dump) needs to be set up and allocated. The following example shows how FTP is
configured to copy the dump to a server:
Note the use of the loopback interface as a source interface and use this address in any system-filter
lists. It is recommended that access to the ‘username’ account above be made as secure as possible.
For example, do not send core dumps to the same FTP server as the one used to provide generic
anonymous or user FTP accounts.
Note that RCP (Remote Copy Protocol) is inherently insecure and is not recommended for use over a
public network. Also TFTP core dumps (which are the default in IOS) only support system memory
sizes up to 16 Megabytes. Generally, it is recommended that FTP core dumps be configured whatever
the situation or router hardware configuration.
Core Dumps must only be configured while troubleshooting a problem and requested by Cisco
TAC.
Enable Port Security on switches - Use the port security feature to block input to an Ethernet port
when from any of the MAC addresses specified for that port.
IOS switch command:
“switchport port-security”
• Service password-encryption - With the exception of the enable secret password, all IOS device
passwords are by default stored in clear text form within the device configuration. These passwords
can be clearly seen by performing a show running-config command.
The service password-encryption command directs the IOS software to encrypt the passwords, CHAP
secrets, and similar data that are saved in its configuration file. This is useful for preventing casual
observers from reading passwords, for example, when they happen to look at the screen over an
administrator's shoulder.
The algorithm used by service password-encryption is considered relatively weak, and was not
designed to protect configuration files against serious analysis, and should not be used for this
purpose. Any Cisco configuration file that contains encrypted passwords should be treated with the
same care used for a clear text list of those same passwords. It is important to use a protected link,
when accessing the router config using the console, or when TFTPing config files. Command:
service password-encrypt
login con 0
login
password <password>
login aux 0
login
password <password>
Alternatively, the console lines can be protected using AAA authentication
login authentication <AAAlistName>
The console may also be secured physically, by placing the router within a lockable enclosure.
If the aux port is not being used, it can be disabled using the following command:
“no exec”
As well as authenticating users accessing the network devices using AAA, access to the virtual terminal
(VTY) should also be locked down to certain addresses only. The suggestion is to restrict this to NMS
platforms as shown by the following template.
functionality that is similar to a telnet connection except that the connection is encrypted, which
allows for secure communication over an insecure network. DES and 3DES encryption is supported.
The SSH server also supports user ID and password authentication using standard authentication
methods: local authentication, RADIUS, and TACACS+. SSH should be used in place of Telnet
wherever possible.
IOS commands:
line vty 0 4
transport input telnet ssh !permits both telnet and ssh, if possible ssh only
!should be used
ip domain-name
crypto key generate rsa
ip ssh time-out 60
ip ssh authentication-retries 3
It is understood that Wateen Telecom will like to use ssh as the only means of remote access into the
device in production (as long as ssh access is supported in IOS image). Hence above example will only
contain “transport input ssh”.
• Set Timeouts for Device Lines - By default, an administrative interface stays active (i.e. logged in)
for ten minutes after the last session activity. After that, the interface times out and logs out of the
session. It is recommended that these timers are fine-tuned to limit the amount of time that an
inactive session is left open. Command:
“exec-timeout x y”
Depending on Wateen Telecom network operations team, this value can be tuned to a lower value or
left at the default provided by Cisco IOS.
Login Banners
For both legal and administrative purposes, configuring a system-warning banner to display prior to login
is a convenient and effective way of reinforcing security and general usage policies. By clearly stating
the ownership, usage, access, and protection policies prior to a login, future potential prosecution
becomes a more solidly backed case.
Wateen Telecom is advised to configure a login banner on every device stating that access is restricted,
and with the NOC contact details. Of course, no one but authorized Wateen Telecom staff should be
accessing the devices but this is a common best practice.
banner login ^
Authorized access only
Unauthorized users will be prosecuted
HTTP Server
Cisco IOS has an inbuilt HTTP service to allow management from a GUI front-end. Most people manage
their routers & switches using the CLI, most service providers keep this service shutdown as part of
security practice. To confirm that service is indeed turned off use the following command in global config
mode.
“no ip http server”
NTP
Network Time Protocol (NTP) is used to synchronise the clocks of various devices across a network. The
local NTP client, which runs on the device, accepts time information from other remote time-servers and
adjusts its clock accordingly. Synchronisation of the clocks within a network is critical for the correct
interpretation of events within Syslog data. The following should be considered when deploying NTP:
• Define a trusted time source and configure all devices as part of an NTP hierarchy.
• Use proper authentication of NTP messages. Version 3 and above of NTP supports a cryptographic
authentication mechanism between peers.
• ACLs should be used to specify which network devices are allowed to synchronise with which other
network devices.
• Disable NTP messages on interfaces that are connected to untrusted interfaces (ntp disable
command). If NTP messages have been disabled on an interface of a router, the router could still
receive and process NTP messages on another interface. ACLs could also be used to drop NTP
messages.
IOS commands to enable and authenticate NTP:
Out-of-band Management
It is highly recommended that the management for all devices in the network should be performed via
console access to the devices and through a dedicated out–of-band management network. This separates
the management traffic plane from the data traffic plane on the network.
Having direct access to the console ports of the managed devices is a good practice. It allows the
operator to manage and maitain the device freely without the fear of being disconnected or loosing remote
access to the device when connecting via the in-band network.
At each remote PoP a router that acts as a terminal server should be deployed that is reachable via a
separate OOB network. For further protection, or if a permanent OOB network is not feasible, ISDN or
POTS line could be used to access the remote sites via an ISDN or a modem module in the terminal
server to provide a temporary OOB network.
It is understood that in current phase of deployment Wateen Telecom has decided not to deploy an out of
band management solution. All management of the network is expected to be inband.
ACL Log
On a 7600 series router if an access-list is configured with a log entry at the end of some lines, all traffic
matching those lines will be inspected and handled by the CPU instead of in hardware.
If there is a need to follow up what has been denied on an ACL in a border router, it is advisable not to
configure log on the ACL, but instead to poll Netflow from the router and look for entries with 0 byte.
mls rate-limiters
The 7600 architecture includes a number of rate-limiters for specific traffic that can impact the
performance of the router. These “mls rate-limiters” are processed in hardware.
Of all the rate-limiters, the following ones are recommended to enabled (in addition to the ones enabled
by default):
Guidelines
In CoPP deployment guidelines customer is recommended to explicitly identify every type of traffic.
There are three CoPP config strategies:
1. Explicitly identify all unwanted traffic, and let all others goes through the class-default (configured
with a policer to safeguard to CPU).
2. Explicitly identify all useful traffic, and harshly police the rest of traffic.
3. Combine 1 and 2, but still need class-default (configured with police) to catch the rest.
For service provider networks Cisco recommends strategy 3 since the combination of 1 & 2 is clearly the
better way to go. User already knows most/all of the useful traffic, so allow it. User already knows
previous attack traffic, so block it completely, and then allow a trickle for anything else, either
unidentified useful traffic, or un-identified attack traffic.
Having the different classes is useful to, as you may not want BGP and SNMP to have the same rates. By
identifying the useful traffic and adjusting the policed BW, you are creating a QoS profile for the control
traffic, where the really important stuff has more access to the RP.
Classes
Critical
Traffic that is crucial to the operation of the router and the network like routing protocols like BGP, ISIS,
LDP etc.
Important
Necessary and frequently used traffic that is required during day-to-day operation, like SSH, Telnet, NTP,
SNMP, igmp etc
Normal
Traffic that is expected but not essential to network operation, like ICMP echo, ping, etc
Bad
Explicitly identify bad or malicious traffic that should be dropped and denied access to the RP.
Match-all
All the remaining IP traffic destined to the RP that has not been identified.
Class-default
This is the built in class that will work as a last resort. The reason for having the Default class above as
“match all IP traffic, we will ensure that all “non IP protocols” ending up in this pre-defined class .
Rate
Rate limiting punted traffic is recommended, care must be taken to ensure that the required rate of traffic
are well understood. The proposed configuration will not implement any rate-limiting functionality that
should have a negative impact on the convergence of routing protocols.
! CoPP configuration
!
mls qos
!
class-map match-all CoPP-CRITICAL
match access-group 120
class-map match-all CoPP-IMPORTANT
match access-group 121
class-map match-all CoPP-NORMAL
match access-group 122
class-map match-all CoPP-BAD
match access-group 123
class-map match-all CoPP-Match-all
match access-group 124
!
access-list 120 remark *** ACL for CoPP-Critical ***
access-list 120 remark *** LDP ***
access-list 120 permit udp 10.0.100.0 0.0.0.255 eq 646 any
access-list 120 permit udp 10.0.100.0 0.0.0.255 any eq 646
access-list 120 permit tcp 10.0.100.0 0.0.0.255 eq 646 any
access-list 120 permit tcp 10.0.100.0 0.0.0.255 any eq 646
!
access-list 120 remark *** OSPF ***
access-list 120 permit ospf 10.0.100.0 0.0.0.255 any
policy-map CoPP
class CoPP-CRITICAL
police 1000000 31250 31250 conform-action transmit exceed-action
transmit
class CoPP-IMPORTANT
police 125000 3906 3906 conform-action transmit exceed-action drop
class CoPP-NORMAL
police 64000 2000 2000 conform-action transmit exceed-action drop
class CoPP-BAD
police 32000 1500 1500 conform-action drop exceed-action drop
class CoPP-Match-all
police 64000 2000 2000 conform-action transmit exceed-action drop
! class class-default
! police 1000000 31250 31250 conform-action transmit exceed-action
transmit
! Note until 12.2.18SXE it is not supported to configure class-default with a
! police action. In the show commands the class default will appear since
this
! class is always there.
!
control-plane
srvice-policy in CoPP
!
*Note the Above ACL’s need to be changed to reflect the addressing in the Wateen Telecom network.
Design of IP/MPLS network for Wateen Telecom as it pertains to L3 vpns, L2vpns, internet and ip
connectivity is discussed in this document. Possible addressing scheme, IGP and BGP configuration
along with possible service offering models, are discussed in detail.
This document is part of a larger low level design effort; other documents that discuss, metro Ethernet
architecture, network security, data center architecture and network management are compiled separately.
TBD To be delivered
Please refer to the CCO Internetworking Terms and Acronyms Guide at
http://www.cisco.com/univercd/cc/td/doc/cisintwk/ita/index.htm for additional terms.
Name Name
Title Title
Company Company
Signature Signature
Date Date
Name Name
Title Title
Company Company
Signature Signature
Date Date
Name Name
Title Title
Company Company
Signature Signature
Date Date
Cisco Systems has more than 200 offices in the following countries and regions. Addresses, phone numbers, and fax numbers are listed on the
Argentina • Australia • Austria • Belgium • Brazil • Bulgaria • Canada • Chile • China • Colombia • Costa Rica • Croatia • Czech Republic Denmark • Dubai, UAE
Finland • France • Germany • Greece • Hong Kong SAR • Hungary • India • Indonesia • Ireland • Israel • Italy • Japan • Korea • Luxembourg • Malaysia • Mexico
The Netherlands • New Zealand • Norway • Peru • Philippines • Poland • Portugal • Puerto Rico • Romania • Russia • Saudi Arabia • Singapore • Slovakia • Slovenia
South Africa • Spain • Sweden • Switzerland • Taiwan • Thailand • Turkey • Ukraine • United Kingdom • United States • Venezuela • Vietnam • Zimbabwe