"firewalls" at high speed -- resending

Hank Nussbacher hank at ibm.net.il
Sun Oct 3 06:34:45 UTC 1999


Hopefully this gets thru this time since the last time I sent it (Sept 30)
I received:
<nanog at merit.edu>: Command died with status 2: "/private/majordomo/wrapper
    resend -l nanog -h merit.edu -s nanog-outgoing". Command output: Can't
    locate getopts.pl in @INC (@INC contains:
    /usr/local/perl-5.004_04/lib/sun4-solaris/5.00404
    /usr/local/perl-5.004_04/lib
    /usr/local/perl-5.004_04/lib/site_perl/sun4-solaris
    /usr/local/perl-5.004_04/lib/site_perl .) at
    /private/majordomo-1.94.4/resend line 74.
and no one from Merit has responded yet to my email.

>Alex Rudnev observed,
>
>>Folks, why all you are saying about the Gigabit traffic for the firewall?
>>
>>Usially, firewall stand between intranet and internet, and it should
>>proceed your upstream traffic, not more... And than, it's important to
>>measure the throughput in packets/per_second, not in the gigabits...
>>
>>Everything other is true - I suggess no one good firewall can proceed
>>gigabit traffic at all, and only a few specially designed boxes can
>>proceed 100Mbit traffic. But just again - it's a rare case when you does
>>have 100Mbit upstream link.

Super Firewalls! 
http://www.data.com/issue/990521/firewalls.html

Almost all hit 72Mb/sec or more whether NAT was disabled or enabled.

To quote from the article:

                            Once again we started with a baseline test.
With no
                            firewall on the test bed, we achieved TCP
                            forwarding rates of 15.6 Mbyte/s over each
                            four-minute run of the test scripts; this works
out to
                            nearly 125 Mbit/s. So why didn’t we get the
                            200-Mbit/s theoretical maximum of a switched,
                            full-duplex test bed? First and foremost, our
                            measurements are taken at the application
layer. It’s
                            likely that packet headers and the continuous
                            opening and closing of SMTP connections took a
                            bite out of the effective data rate. Second,
clients
                            offered lots of traffic that the rule sets
expressly
                            denied—so at least some of the time the wire was
                            occupied carrying traffic that wouldn’t be
forwarded.
                            Third, there may have been a firepower
limitation in
                            the amount of traffic our clients and servers
could
                            offer. We can’t say for certain how much the
                            degradation is due to application overhead, and
how
                            much to test platform limitations. 

I would think that 200Mb/sec is achievable with not much effort and perhaps
the next testing round will prove it. 

-Hank





More information about the NANOG mailing list