verio arrogance
David Diaz
dave at smoton.net
Fri Jul 19 17:05:02 UTC 2002
At 12:16 -0400 7/19/02, Richard A Steenbergen wrote:
>On Fri, Jul 19, 2002 at 11:00:38AM -0400, Daniel Golding wrote:
>>
>> I think we are at the point where the vast majority of backbone routers can
>> handle 200K+ routes, at least in terms of memory. The interesting point we
>> are getting to, is that the most popular router in the world for multihoming
>> can't handle the routing table. I'm referring to the Cisco 3640, which has
>> largely supplanted the venerable 2501 as the low-end multihomer's edge
>> router of choice.
>>
>> With a reasonable number of features turned on (i.e. SSH, netflow, CEF), the
>> 3640 can't handle two full views anymore, due to it's limitation of 128MB.
>> While this may be a good thing for Cisco's sales numbers, in this winter of
>> financial discontent, I wonder how this is effecting the average customer,
>> and what is generally being installed to replace the 3640s.
>
>If a 3640 customer can't handle multiple full views, why can't they
>filter some junk /24s themselves? This isn't really a good enough reason
>for backbone providers to do the filtering.
That was my thinking also. I would imagine a lot of customers what a
full route view, it's what they are paying for especially if they are
an ISP or multihomed large customer. They should have their own
policies then.
>
>As for the convergence time argument, the limiting factor is CPU time,
>not the number of routes or amount of data exchanged (though obviously
>more routes == more cpu). In the core, is there really that big a
>difference between 93k and 113k? On the borders, how much cpu time is
>saved vs how much cpu time is burned doing the filtering?
I would assume a flapping session with a large backbone would cause
much higher load time and stress on the router then simply a large
table. It's the reason why some backbones have Draconian route
dampening policies, and rightly so. I would love to see some
engineers from vendors weight in on this (did I just say that?).
Most brag that they can handle large tables without a problem. A
good question might be, if a large backbone started flapping 150,000
routes, what would that do to the peers. Perhaps a better issue
much be CPU usage of complex route filters on large tables, as a
limitation on performance.
>
>Which leaves us with the question of, are there still MSFC1's or other
>devices with 128mb memory limits in these networks which are hurting at
>113k? Is there actually a legitimate technical need to filter off 20k
>routes, or are the people doing it stuck in a mental time warp from the
>days when it was a necessity?
>
>Or, is it really just people trying to do the "correct" thing? If you see
>"almost" no change in connectivity after removing 20k of cruft, and the
>very few people who are broken are the ones who needed attention called
>to their poor route announcing skills anyways, maybe it's a good thing
>for reasons other than router performance?
Interesting thought is, there are probably a great many engineers on
this list that have /24s at their home, that dont enjoy being
filtered. Some of us just get tired of reIPing our servers.
dave
>
>--
>Richard A Steenbergen <ras at e-gerbil.net> http://www.e-gerbil.net/ras
>PGP Key ID: 0x138EA177 (67 29 D7 BC E8 18 3E DA B2 46 B3 D8 14 36 FE B6)
--
David Diaz
dave at smoton.net [Email]
pagedave at smoton.net [Pager]
Smotons (Smart Photons) trump dumb photons
More information about the NANOG
mailing list