Affiliation:
1. University of California, San Diego, La Jolla, CA
Abstract
Accurate network traffic measurement is required for accounting, bandwidth provisioning and detecting DoS attacks. These applications see the traffic as a collection of flows they need to measure. As link speeds and the number of flows increase, keeping a counter for each flow is too expensive (using SRAM) or slow (using DRAM). The current state-of-the-art methods (Cisco's sampled NetFlow), which count periodically sampled packets are slow, inaccurate and resource-intensive. Previous work showed that at different granularities a small number of "heavy hitters" accounts for a large share of traffic. Our paper introduces a paradigm shift by concentrating the measurement process on large flows only---those above some threshold such as 0.1% of the link capacity.We propose two novel and scalable algorithms for identifying the large flows:
sample and hold
and
multistage filters
, which take a constant number of memory references per packet and use a small amount of memory. If
M
is the available memory, we show analytically that the errors of our new algorithms are proportional to 1/
M
; by contrast, the error of an algorithm based on classical sampling is proportional to 1/√
M
, thus providing much less accuracy for the same amount of memory. We also describe optimizations such as
early removal
and
conservative update
that further improve the accuracy of our algorithms, as measured on real traffic traces, by an order of magnitude. Our schemes allow a new form of accounting called
threshold accounting
in which only flows above a threshold are charged by usage while the rest are charged a fixed fee. Threshold accounting generalizes usage-based and duration based pricing.
Publisher
Association for Computing Machinery (ACM)
Cited by
310 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献