on 09-26-2012 12:06 PM
This document decribes the ASR9000 netflow architecture.
It provides a basic configuration how to set up netflow and what the parameters are for scale and how netflow is implemented in the ASR9000/XR
The basic configuration for netflow consists of an
Flow monitor map
An exporter map
and a sampler map.
The Flow monitor MAP pulls in the Exporter map
On the interface you want to enable netflow on, you pull in the monitor map and the sampler map.
flow monitor-map FM
record ipv4
exporter FE
cache permanent
cache entries 10000
! cache timeouts define how frequently we export what, max of 1M per LC
cache timeout active 2
cache timeout inactive 2
!
flow exporter-map FE
version v9
options interface-table timeout 120
! these 2 define the exports of the sample map and interface table to theflow collector for sync'ing indexes
! to names etc.
options sampler-table timeout 120
!
transport udp 1963
destination 12.24.39.1
source <interfacename>
!
sampler-map FS
random 1 out-of 1
interface GigabitEthernet0/0/0/20
description Test PW to Adtech G4
ipv4 address 16.1.2.1 255.255.255.0
flow ipv4 monitor FM sampler FS ingress
!
Ø Trident: 100kpps/LC (total, that is in+out combined) Typhoon: 200kpps/LC Ø 1M records per LC (default cache size is 64k) Ø 50K Flows per sec export per LC Ø Sample intervals from 1:1 to 1:64k
Ø Up to 8 exporters per map, vrf aware
Netflow is not hardware accelerated in the ASR9000 or XR for that matter, but it is distributed.
What that means is that each linecard individually runs netflow by itself.
Resources are shared between the interfaces and NPU's on the linecard.
When you have 1 interface to one NPU on one linecard enabled for netflow, the full rate is available to that interface, which is 100k pps for trident and 200k for typhoon.
When you enable 2 interfaces on the same NPU on the same LC, then both interfaces share the 100k pps (trident) or 200k pps (typhoon)
When you enable 2 interfaces on 2 different NPU's, then both NPU's share the total rate of 100k/200k amongst them giving each NPU 50k or 100k depending on the LC type.
In IOS-XR platforms, it is the LC processor memory that holds the netflow cache.
NetFlow Cache is a Section of memory that stores flow entries before they are exported to external collector.
The ‘nfsvr’ process running on the linecard, manages the netflow cache.
Memory usage
The memory used can be monitored via this command:
show flow monitor FM cache internal location 0/0/CPU0
...
Memory used: 8127060
Total memory used can be verified by checking the process memory util of "NFSVR"
show processes memory location 0/0/CPU0 | inc nfsvr
257 139264 65536 73728 12812288 nfsvr
The memory used with the cache size of default 64k entries for ipv4 & MPLS is about 8MB & for ipv6 is about 11MB.
The memory used with the cache size of maximum 1M entries for ipv4 & MPLS is about 116 MB & for ipv6 is about 150MB.
The memory used with cache size of maximum 1M entries (default is 65535) is about 116 MB per ipv4 flow monitor .
If ‘n’ ipv4 flow monitors are used all with maximum 1M entries, the memory used would be n x 116 MB.
Configuration to set the cache entries to ten thousand looks as follows:
flow monitor-map FM
cache entries 10000
95% of configured cache size is the high watermark threshold. Once this threshold is reached, certain flows (longest idle ones etc) are aggressively
timed out. XR 4.1.1 attempts to expire 15% of the flows.
The show flow monitor FM cache internal location 0/0/cpu0 command will give you the data on that:
Cache summary for Flow Monitor :
Cache size: 65535
Current entries: 17
High Watermark: 62258
this syslog message means that we wanted to add more entries to the cache than what it could hold. There are a few different reasons and remediations for it:
- the cache size is too small, and by enlarging it we can hold more entries
- the inactive timeouts are too long, that is we hold entries too long in the cache not getting aged fast enough
- we have the right size cache, and we do export them adequately, but we are not getting the records out fast enough due to volume, in that case we can tune the rate limit of cache expiration entries via:
flow monitor <name> cache timeout rate-limit <time>
The permanent cache is very different from a normal cache and will be useful for accounting or security monitoring. The permanent cache will be a fixed size chosen by the user. After the permanent cache is full all new flows will be dropped but all flows in the cache will be continuously updated over time (i.e similar to interface counters).
Note that the permanent cache uses a different template when it comes to the bytes and packets.
When using this perm cache, we do not report fields 1 and 2, but instead use 85 and 86.
Fields 1 and 2 are “deltas” 85 and 86 are "running counters".
In your collector you need to "teach" it that 1 and 85, 2 and 86 are equivalent.
All packets subject to sampling, regardless or whethe they are forwarded or not are subject to netflow.
This includes packets dropped by ACL or QOS policing for instance!
A drop reason is reported to NF.. * ACL deny * unroutable * policer drop * WRED drop * Bad IP header checksum * TTL exceeded * Bad total length * uRPF drop
IPV4SrcAddr IPV4DstAddr L4SrcPort L4DestPort IPV4Prot IPV4TOS InputInterface ForwardStatus ByteCount PacketCount Dir 17.1.1.2 99.99.99.99 3357 3357 udp 0 Gi0/1/0/39 DropACLDeny 415396224 8654088 Ing
As described in the architecture section, the total sampling capability depends on the number of interfaces having netflow enabled.
It shaped up to be something like this table:
# of NPs Enabled for Netflow |
Policing Rate Per Trident NP (Unidirectional) |
Policing Rate Per Typhoon NP (Unidirectional) |
1 |
100kpps |
200kpps |
2 |
50kpps |
100kpps |
3 |
33kpps |
66kpps |
4 |
25kpps |
50kpps |
All packets that exceed this rate are dropped by the punt policer.
You can verify that by the controllers np counters command.
show controllers np counters all
Node: 0/0/CPU0:
----------------------------------------------------------------
Show global stats counters for NP0, revision v2
Read 67 non-zero NP counters:
Offset Counter FrameValue Rate (pps)
-------------------------------------------------------------------------------
....
934 PUNT_NETFLOW 18089731973 6247
935 PUNT_NETFLOW_EXCD 6245 0
...
The _EXCD depicts that the police rate had been exceeded.
This means that you likely have to increase your sampling interval.
sh flow monitor FM cache format table include layer4 tcp-flags ipv4 sour dest prot tos count pack byte location 0/0/CPU0
Mon Apr 19 09:31:19.589 EDT
Cache summary for Flow Monitor FM:
Cache size: 10000
Current entries: 1
High Watermark: 9500
Flows added: 1
Flows not added: 0
Ager Polls: 580
- Active timeout 0
- Inactive timeout 0
- TCP FIN flag 0
- Watermark aged 0
- Emergency aged 0
- Counter wrap aged 0
- Total 0
Periodic export:
- Counter wrap 0
- TCP FIN flag 0
Flows exported 0
IPV4SrcAddr IPV4DstAddr IPV4Prot IPV4TOS L4TCPFlags ByteCount PacketCount
16.1.2.2 16.1.1.2 tcp 0 S| 4282560 71 376
Matching entries: 1
Export occurs when data in the cache is removed which can occur in one of three ways.
The netflow exporter can be in a VRF, but can not be out of the Mgmt Interface.
Here’s why. The netflow runs off of the line card (LC interfaces and NP) and there is, by default, no forwarding between the LCs and the management Ethernet. This because the MGMT ether is designated out of band by LPTS (local packet transport services). More detail in the ASR9000 Local packet transport services document here on support forums).
Netflow records can be exported to any destination that may or may not be local to the LC where netflow is running. For example, LC in slot 1 & 2 are running netflow & the exporter may be connected to an interface reachble via LC in slot 3.
A total of 8 exporters per MAP is allowed.
.....
Ø DBNA
Ø Cisco netflow mib is not supported.
show flow exporter-map ..
show flow monitor-map ..
show sampler-map ..
show processes memory location <0/0/CPU0> | inc nfsvr
show flow monitor .. cache internal location <0/0/CPU0>
show flow exporter .. location <0/1/CPU0>
show flow platform producer statistics location <0/0/CPU0>
show flow platform nfea policer np <np_num> loc <node-id>
show controller np ports all location <0/Y/CPU0>
show controller np count np<number> loc <0/Y/CPU0>
flexcli could be usable here Alexandr?!
some new enahcenemnts in 53 can show the interfaces that currently have nf assigned.
that wont help you in terms of bulk config, but flex cli would be your friend here?!
cheers!
xander
Thank you for idea with group apply, but I think, I should search other points of netflow attachment. E.g. couple egress mpls interfaces from this PE.
Hi Xander,
Great tutorial as usual, thanks!
I have a question related to the meaning of aging fail (ring ful) counter. I see logs of cache entries exceeded in my CU network, first idea is to increase the cache entries, but high water mark counters are 0 .. on the other hand aging fail (ring ful) are high in the affected LC. Can you please help me to find out the meaning of that counter? (ASR9922 typhoon LC IOS XR 5.3.2)
RP/0/RP0/CPU0:i79zhb-005#show flow monitor IPv4-map cache internal location 0/1/CPU0
Wed Jan 20 12:07:38.384 MET
Cache summary for Flow Monitor :
Cache size: 500000
Current entries: 36925
Flows added: 9741866372
Flows not added: 2279161
Ager Polls: 7516085
- Active timeout 835450311
- Inactive timeout 8631982666
- TCP FIN flag 274396470
- Emergency aged 0
- Counter wrap aged 0
- Total 9741829447
Periodic export:
- Counter wrap 0
- TCP FIN flag 0
Flows exported 9741829447
Internal Information:
Cache Hits: 13390877387
Cache Misses: 9744145533
Cache Overflows: 2279161
Cache above hi water: 0
Flows dropped by ager: 0
Aging fail (ring full): 16679138----> what does it mean?
Periodic export fail: 0
Hash buckets: 0
Max hash depth: 0
Memory used: 68800796
Hash Statistics:
Bkt Depth:00000-00000 00001-00001 00002-00002 00003-00003 00004-00005 00006-00007 00008-00010 00011-00015 00016-00020 00021-00050 00051-00100 00101-00200 00201-00500 00501-01000 01001-65535
Current: 69132 25473 4762 566 57 0 0 0 0 0 0 0 0 0 0
Maximum: 1 0 0 0 0 0 43662 56204 123 0 0 0 0 0 0
RP/0/RP0/CPU0:i79zhb-005#
Thanks in advance!!!
an thanks :) oh that counter, I had to look up inside the process when this gets incremented and under what conditions/what it means.
when a flow is to be exported, this can be for aging reasons or that the cache is near full, the flow is Q'd against the exporter handler. The exporter has a ring buffer for the flow (records) and when that ring is full we see this counter bump.
It can mean we can't export fast enough (we do 5k per sec, so well enough I'd say) or it means that we have too agressive aging (we can tune the timers) OR.... yeah if the cache is small and we have too many flows, we have to age aggressively at 85% cache util,
and that latter I think you may have.
when the ring full is seen, we just retry later on, so you didn't lose anything yet.
cheers
xander
Hi Xander,
Thanks for your fast and clear reply.
After trying increasing the number of cache entries to maximum (1M) with no success.
But I forgot about the cache timeout rate-limit 7000 configuration we have, and I think that is causing the problem in the first place. If I am not wrong this is letting too many flows to be exported and therefore filling the buffer leading to logging errors.
If I understood your answer properly we are supporting 5K flows per second for active flows...so this config we have in place is not making a lot of sense :) is that right?
Thanks and best regards
Cristina
Xander
Does IOS XR support multiple flow export destinations with different sampler rates? I attempted to configure multiple flow monitors per interface and I get the message:
'nfea' detected the 'warning' condition 'Platform limit for requested operation exceeded'
that you can't do Evan. the sampler is set per interface, but you can have multiple export destinations, but those multiple desintations basically exert the same info from the cache
cheers
xander
Ah a bit different than IOS-XE in that regard. We will find a compromise.
Thanks for the quick reply!
Hey Xander how are you!
Releted to this topic for collecting netflow information of IPv6 on ASR9K with 6PE .. only needs this to be configured?
flow ipv6 monitor FM sampler FS ingress
Best regards my friend..
Cheers
Javier
Hi, Xander:
I was looking for ASR9k/CRS support for Netflow over SCTP, but as far as the configuration guide they state it's not supported. Can you please confirm either way and if yes can you point to a document so I can lookup restrictions, etc?
Thank you very much!
c.
hi carlos,
yeah we only export neflow over v9/udp, no sctp.
we have telemetry coming up soon that may be of interest?
cheers!
xander
thank you very much, xander!
Hi there
Think you have a minor type here:
The default size of the netflow cache is 64k entries. The maximum configurable size of the netflow cache is 1M entries.
Configuration for that is as follows:
flow monitor-map FM
cache entries 10000
I think you need to add some extra 0's to make it 1M :)
good eye ;) I fixed it ;)
xander
I happened to be looking for FNF support in XR and found this thread - thanks Xander.
Did FNF make it into 6.0? How about for the 9001, would that be 6.1.1?
As you mention it, I don't know much about it but have seen some basics around streaming telemetry feature, but can it be leveraged as a method for sending flow data?
Find answers to your questions by entering keywords or phrases in the Search bar above. New here? Use these resources to familiarize yourself with the community: